* Team Name: Stack
* Team Members: Guanlin Li & Yudong Liu & Jiajun Chen
* Github Repository URL: https://github.com/upenn-embedded/final-project-stack
* Github Pages Website URL: https://upenn-embedded.github.io/final-project-stack/
* Description of hardware: esp32, ATmega328PB, Nano33 IoT
https://github.com/user-attachments/assets/94df1cdc-a7fc-4cc9-9b69-86574b1cf152
You could also find the video at ./videos/final_demo_lite.mp4
The team’s system design consists of two main components: a smart furniture module powered by an ATmega328PB microcontroller and a wand controller centered around the Nano 33 microcontroller, equipped with Wi-Fi signal transmission capabilities. The wand integrates a power supply system using three AA batteries to ensure stable power delivery to the Nano 33 and the WS2812b light strip. The Nano 33 leverages its onboard gyroscope and Inertial Measurement Unit (IMU) to capture real-time angular velocity and motion data, which form the basis for gesture recognition.
Figure 1: The Overview of Design Schematic
The motion data collected by the Nano 33 is aggregated and processed to train machine learning models aimed at accurately classifying three distinct wand gestures: circular, triangular, and linear motions. The resulting trained parameter matrix is then deployed directly onto the Nano 33, enabling on-device gesture recognition. When the user performs a gesture, the wand processes the motion data and encodes the recognized gesture into an 8-bit binary signal (e.g., 1 for a circle, 2 for a triangle, and 3 for a line). This signal is transmitted wirelessly via the Nano 33’s Wi-Fi module to an ESP32 device connected to the same local network.
Figure 2: The Flow Diagram Of the Adapted Program Design
The ESP32 is configured with a static IP address defined at the software level, ensuring reliable communication within the network. Upon receiving the gesture signal, the ESP32 forward it to the ATmega328PB microcontroller using the SPI communication protocol. The ATmega328PB then decodes the signal and executes the corresponding predefined actions, such as controlling lighting, activating a water pump, or driving a motor. This integrated system demonstrates a seamless interaction between the wand and the smart furniture, achieving the team’s vision of a gesture-based controller for smart home automation. Each subsystem is designed with modularity and scalability in mind, ensuring a robust and efficient implementation of the team’s concept.
We waved 100 times for different postures and records their recognized results. The graph below shows the simulation results:
Figure 10: The Simulation Results
The Summary Table:
Posture | Total Test | Accuracy | Misrecognize Gesture 1 | Misrecognize Gesture 2 | Misrecognize Gesture 3 |
---|---|---|---|---|---|
1 | 100 | 90% | 0 | 5 | 5 |
2 | 100 | 89% | 7 | 0 | 4 |
3 | 100 | 95% | 3 | 2 | 0 |
The Test result shows that we perfectly meet our requirement.
As the image shows, the average latency is below 200ms. Therefore, this target is achieved.
Battery performance: At voltage = 5V, the current is around 60ma when the light is on. 3 normal Alkaline battery has around 2850 mAh at 4.5V. Based on this data, a rough estimate of the continue use time could be calculated, which is around $\frac{2850}{60} \approx 47.5$ hours. This exceed the target requirements of 4 hours. Therefore, the battery performance meets its requirement.
In this project, we delved deeply into various aspects of MCU development, exploring both familiar and unfamiliar platforms. The experience extended from foundational circuit-level programming on the ATmega328, progressing through advanced SPI communication on the ESP32, to managing interrupts on the Nano 33. By directly manipulating registers and developing drivers, we gained a profound understanding of these MCUs, allowing us to implement core functionalities effectively and seamlessly integrate them into our project. A notable achievement was the development of hardware drivers. For example, the motor driver required generating a 50Hz PWM signal with a variable duty cycle, which performed as expected. However, writing the driver for the LED light strip proved significantly more challenging. While an initial approach leveraging the SPI protocol was considered, the SPI signal generated by the ATmega328PB introduced gaps every 8 bits, necessitating a switch to the bit-banging method. This adaptive approach showcased our problem-solving capabilities and strengthened our understanding of low-level communication protocols. Throughout the project, we encountered and overcame difficulties related to integrating multiple components and systems, such as synchronizing the gesture recognition algorithm with hardware control mechanisms. The algorithm, which performed flawlessly, highlighted the synergy between software and hardware in achieving the project’s goals. Looking forward, a potential enhancement would involve integrating machine-learning-based algorithms onto the MCU to detect intentional gestures more accurately, further advancing the system’s functionality. Our project also emphasized the importance of teamwork and proactive planning. From defining functional goals at the outset to achieving them through close collaboration, we were able to complete the project as scheduled. This accomplishment reflected not only individual contributions but also the cohesive efforts of the entire team. The breadth of this project was another significant aspect, encompassing topics such as digital signal processing, serial communication protocols, PWM, interrupts, and wireless transmission. Tackling these diverse challenges required iterative problem-solving, adaptability, and a willingness to engage with new concepts. Ultimately, this integrative experience provided a comprehensive understanding of MCU development and its practical applications, leaving us with a sense of pride in our accomplishments and gratitude for the collaborative effort.
The implementation for the ATmega328PB is on pure bare-metal C and without external libraries. The interrupt on the Nano33 is also implemented without the use of external libraries. These should conclude at least 3 required topics. For the rest of the part which more or less uses some external libraries, please check the appendix below.
We used math.h for data preprocessing. The function we used included an absolute value function (abs) for measuring the change between two adjacent sensory data, sin and cos function for performing time-series data filtering and normalization, and square root function (sqrt) for calculating the standard deviation of a time-series sequence data for normalization. The above functions could all be implemented based upon bitwise operations of AND, OR, NOT, XOR and SHIFT.
we use this function to establish connection between nano33 and esp32. The Wifi.h library for the Arduino Nano 33 series provides an abstraction layer for interacting with Wi-Fi hardware. At bare-metal C level, the chip uses the u-blox NINA-W102, a Wi-Fi and Bluetooth combo module that communicates with the microcontroller over a Serial Peripheral Interface (SPI). The wifi is established through the above procedures on a lower level:
The library function we used include:
This library uses the bit-banging method to implement the driver for the ws2812b light strip. The basic method is the same as how I’ve implemented it on the ATmega328PB using bare metal C code.
In this project, the ESP32 acts as the SPI master, while the ATmega functions as the slave. We implemented the SPI controller on the ATmega328PB side using bare-metal C code and without any external library. We use SPI.h on the STM32 side to help us finish the project faster.
The spi.h library facilitates the following operations for data transfer:
The spi.h library provides a high-level interface for SPI communication and consists of the following components: (1) Configuration Functions These functions initialize SPI communication parameters, including: · SPI Mode: Defines the clock polarity (CPOL) and phase (CPHA). Common configurations are Mode 0 or Mode 1. · Clock Frequency: Determines the data transfer speed. · Bit Order: Specifies whether data is transmitted Most Significant Bit (MSB) first or Least Significant Bit (LSB) first.
Example:
spi_bus_config_t bus_config = {
.mosi_io_num = PIN_NUM_MOSI,
.miso_io_num = PIN_NUM_MISO,
.sclk_io_num = PIN_NUM_CLK,
.quadwp_io_num = -1,
.quadhd_io_num = -1
};
spi_bus_initialize(HSPI_HOST, &bus_config, 1);
(2) Data Transfer Functions
The key functions for data transfer include: · spi_transfer(): Performs a complete transaction, including sending and receiving data simultaneously. · spi_write() and spi_read(): Perform write-only or read-only operations, respectively.
spi_transaction_t trans = {
.length = 8 * sizeof(data),
.tx_buffer = data_to_send,
.rx_buffer = data_received
};
spi_device_transmit(spi_handle, &trans);
(3) Interrupts and Event Handling · The SPI driver on the ESP32 supports event callbacks, such as triggering an interrupt when a transfer completes. · Interrupts can be configured to synchronize communication with the ATmega.
We only use the Serial class for debugging purposes, it could be removed without problems.
1.1 SPI: It is used for data transmission between the ESP32 S2 and the Atmega328. The ESP32 receives commands from the Nano33 via Wi-Fi and transmits them to the Atmega328, enabling it to control the furniture. Two versions of the communication protocol were implemented on the ESP32: one in bare C and the other using the SPI.h library. All code on the Atmega328 was written in bare C.
1.2 Wireless communication:: Wi-Fi was utilized to enable wireless connection and data transmission between the ESP32 and Nano33. Both devices employed the WiFi.h library, but they were manually configured to connect within a local network with assigned IP addresses to complete the data reception logic.
1.3 Interrupts: In the Nano33, we utilized interrupts to enable the MCU to periodically receive data from the IMU and store it in a sequence. This approach ensured the formation of a temporally uniform dataset, unaffected by variations in computation time during each recognition process. This was achieved by directly accessing the Nano33 registers to configure the interrupt and set its frequency to 10 Hz.
1.4 Digital Data Processing: In the posture recognition module, we designed and implemented an algorithm tailored to our application needs. This algorithm analyzes a complete data sequence to determine the number of inflection points, which are then compared to identify the gesture’s shape and corresponding command. For certain complex mathematical operations, the math library was utilized.
1.5 Timer(PWM): We designed a system to adjust the duty cycle of the PWM signal, enabling the motor to rotate from 0 to 90 degrees to achieve the door-opening and closing functionality.
The IMU & Gyroscope sensor: Responsible for collecting angular velocity and acceleration data during the motion of the wand to facilitate data processing. This data is subsequently input into the Nano 33.
Motor: Responsible for controlling the opening and closing of the door, which is regulated by a PWM signal.
LED light strip: They were applied to the smart home model and the wand separately. One was used to demonstrate the wand’s ability to control the status of lights, while the other was integrated into the wand to provide feedback for different commands, enhancing its visual appeal and interactivity.
Relay: Controlled by simple signals, it essentially functions as a switch to control the water pump.
The LED strip requires precise signals (0.4 micro seconds) in order to control the light. We tried to use SPI to control it at first, but the ATmega328PB outputs a gap (2 cycles long, pulling everything low) between two 8-bit SPI transmission, which is not what we want. Therefore, after some research, we went for the bit-banging method, which is to use assembly code to control the GPIO pin in order to reach the required frequency.
The development of the ESP32 and Nano33 was a critical aspect of the project. Notably, we worked on three different MCUs, each with unique and significant functionalities. For example, the ESP32 required configuring SPI in Master mode, while the Nano33 involved direct register access to set up a 10 Hz interrupt.
Both chips, being part of integrated modules, presented additional challenges. The ESP32 was part of the ESP FeatherS2 board, and the Nano33 included components like an IMU, resulting in highly complex GPIO mappings. The available documentation and resources typically relied on pre-built libraries, making bare-metal C development exceptionally challenging due to differing implementations across MCUs.
We conducted extensive research and referred to numerous resources and forums to understand how to achieve our desired functionality via direct register access. Ultimately, both parts were successfully implemented using bare-metal C, overcoming the inherent complexities and demonstrating a deep understanding of each MCU’s architecture.
Recognizing gestures is a complicated task as there are no existing datasets and ready-to-use algorithms that’ll fit our use-case and hardware sensories. Developing a recognition algorithm also requires software hardware codesign since we need to make sure it could run properly on a platform with limited computational power. The difficulties come from three aspects, data collecting, data filtering (preprocessing) and computation.
The data for gesture recognition is sampled from the IMU and Gyroscope from the Nano33 IMU & Gyroscope module using built-in packages. To ensure consistent data reading from the gyro and IMU, we added interrupts to the Nano33. This allows us to read data every 100ms, avoiding irregular and inconsistent readings caused by varying processing times of the data matrix.
One of the issues we encountered was the race condition of reading and updating the gyroscope data, and using it for computation. We solved this issue using a cyclic queue with semaphores.
Each time a new data is sampled from the IMU and Gyroscope, we insert the new data into a cyclic queue with a maximum capacity of N elements. The purpose of the queue is to allow concurrent O(1) insertion and deletion of sensory data. During computation, the main thread keeps popping elements from the queue to detect possible sharp turns and shifts in IMU and gyroscope data, while the time interrupt is triggered once per 100ms to collect data from the register that stores IMU readings. Due to the asynchronous execution between the two sets of operations, we use semaphores to lock the queue during insertion, popping, element counting and copying. This ensures there’s no race conditions in the queue.
Another issue we need to handle is the loss of packets when collecting time-series data for computation (the new data overwrites the old data in the queue before they’re processed and popped away). In order to ensure consistency of our training data format and record unintended data losses due to processing latencies, label each data with a unique id that increments for each sample. Before we perform computation, we measure the gap of id value between two consecutive IMU data samples to make sure they’re off by 1. Otherwise, we clear the queue to restart data collection.
Due to the intense workload of implementing networks from scratch on the MCU and the huge latency introduced by its inference, we choose to use a simpler thresholding method for gesture recognition on denoised time-series IMU data. Each time we pop a new element from the queue, we compare it with the previous element for Euclidean distance. If the two IMU data vectors are close enough (within a certain threshold), we ignore the data and throw it away. Otherwise, we enter a data-collection mode and store the data into a vector, and exit the data-collection mode if the data becomes stationary for a period of time (a series of N data points have Euclidean distance within a certain threshold). After we finish collecting the data, we perform normalization on the data-series using their max and min values. We also perform principal component extraction for denoising the data series by implementing matrix computation from scratch. We measure the number of sharp turns in the denoised data series by comparing consecutive data vectors through Euclidean distance calculation. By analyzing the number of sharp turns we could classify the gesture into different shapes such as a line, triangle, square and star.
When connecting the ESP32 to the ATmega328PB using SPI, some package was sent by the ESP32 but not received by the ATmega328PB. After some time of research, we figured out that the frequency of the SPI clock is too high for the ATmega328PB to stably receive data. We solved this problem by lowering the frequency of the SPI clock.
We also encountered packet drops and connection issue when trying to connect Nano33 to the ESP32 via WiFi. We set up the WiFi separately and used a python script to test if it’s working properly. However, when we tried to connect the two devices together, they cannot communicate with each other. We figured out that the initialization of the two parts need to be adjust before they can connect together.
Another integration problem is when we try to burn the algorithm that decides what shape the user just waved to the Nano33. The code works fine when we compile it locally, but when trying to move it onto the board it gives several errors that we were managed to solve in around a week’s time.
The team’s system design consists of two main components: a smart furniture module powered by an ATmega328PB microcontroller and a wand controller centered around the Nano 33 microcontroller, equipped with Wi-Fi signal transmission capabilities. The wand integrates a power supply system using three AA batteries to ensure stable power delivery to the Nano 33 and the WS2812b light strip. The Nano 33 leverages its onboard gyroscope and Inertial Measurement Unit (IMU) to capture real-time angular velocity and motion data, which form the basis for gesture recognition.
Figure 1: The Overview of Design Schematic
The motion data collected by the Nano 33 is aggregated and processed to train machine learning models aimed at accurately classifying three distinct wand gestures: circular, triangular, and linear motions. The resulting trained parameter matrix is then deployed directly onto the Nano 33, enabling on-device gesture recognition. When the user performs a gesture, the wand processes the motion data and encodes the recognized gesture into an 8-bit binary signal (e.g., 1 for a circle, 2 for a triangle, and 3 for a line). This signal is transmitted wirelessly via the Nano 33’s Wi-Fi module to an ESP32 device connected to the same local network.
Figure 2: The Flow Diagram Of the Adapted Program Design
The ESP32 is configured with a static IP address defined at the software level, ensuring reliable communication within the network. Upon receiving the gesture signal, the ESP32 forward it to the ATmega328PB microcontroller using the SPI communication protocol. The ATmega328PB then decodes the signal and executes the corresponding predefined actions, such as controlling lighting, activating a water pump, or driving a motor. This integrated system demonstrates a seamless interaction between the wand and the smart furniture, achieving the team’s vision of a gesture-based controller for smart home automation. Each subsystem is designed with modularity and scalability in mind, ensuring a robust and efficient implementation of the team’s concept.
The configuration of ATmega328PB is shown below:
Figure 3: The Wire Configuration of ATmega328PB
This block diagram illustrates the connection between the ATmega328PB microcontroller, the ESP32 module, and peripheral components. The ATmega328PB connects to peripherals as follows: PB0 handles LED, PB1 controls the motor (door), and PC0 interfaces with the pump, with Vdd providing 5v and GND providing the ground connection. The ATmega328PB communicates with the ESP32 using the SPI protocol, where PB2 connects to SS, PB3 to MOSI, PB4 to MISO and PB5 to SCK. The ESP32’s SPI pins are connected to SDI (GPIO 37), SDO (GPIO 35), SS (GPIO 33) and SCK (GPIO 36). Both the ATmega328PB and ESP32 share a common 5V power supply via Vdd and are grounded to the same GND pin. This configuration enables the ATmega328PB to control peripherals while communicating with the ESP32 for advanced processing or wireless functionality.
The walls of our smart furniture model are created using laser cutting. The core part is designed to accommodate a servo motor, which controls the door’s opening and closing. We modeled it to ensure the servo motor can be properly installed.
Figure 4: The 3D Module of The Gate
In Atmega328Pb, we should continuously check whether data sent from the ESP32 has been received. Upon reception, we use an if statement to evaluate its value and execute different operations accordingly. The core bare metal C are shown below:
while (1) {
if (data_received) {
if (data == 1) door_status ^= 1, motor_angle(door_status ? 90 : 0);
else if (data == 2) light_status ^= 1, ws2812b_send(light_status ? 255 : 0, 255, 255);
else if (data == 3) pump_status ^= 1, PUMP_PORT ^= (1 << PUMP_PIN); }}
We will excute different command based on the data transferred.
SPI (Serial Peripheral Interface) is a high-speed, full-duplex communication protocol commonly used for short-distance communication between microcontrollers and peripheral devices. It operates with a master-slave architecture, where the master device controls the clock signal (SCK), enabling synchronized data exchange. SPI uses four main lines: MOSI (Master Out Slave In), MISO (Master In Slave Out), SCK (Serial Clock), and SS (Slave Select). It is known for its simplicity, high data transfer rate, and flexibility, making it ideal for applications like sensor data acquisition, memory interfacing, and communication between processors.
In our project, it is used to transfer data from ESP32 to Atmega328PB to send instructions. To implement this communication protocol, we need to code Slave Mode in Atmega328pb and Master Mode in ESP32.
The core codes are shown below to initialize Slave Mode:
#include <stdlib.h>
#include <avr/interrupt.h>
#include <stdio.h>
#include <avr/io.h>
uint8_t SPI_SlaveReceive();
void SPI_Slave_initialize()
{
SPI_DDR |= (1 << SPI_MISO);
SPCR0 |= (1 << SPE);
}
uint8_t SPI_SlaveReceive()
{
while(!(SPSR0 & (1<<SPIF)));
return SPDR0;
}
Overall, we enabled SPI, configured it in Slave mode, and set the data transfer mode to the default Mode 0, which must match the configuration in the ESP.
The SPI configuration process on the ESP32 is straightforward and similar to previous setups. It involves accessing the relevant registers, enabling SPI, setting it to Master mode, and using Mode 0 for the transfer mode to ensure proper communication. By default, the ESP32 supports four SPI buses, but currently, only HSPI and VSPI are available for use. However, due to the unique GPIO mapping of the ESP32 Feather S2, the pin assignments differ, as shown in the diagram below:
Figure 6: The Pin Layout Of ESP32 Feather S2
The core codes to implement are shown:
#include "esp32/rom/gpio.h"
#include "soc/spi_struct.h"
#include "soc/gpio_sig_map.h"
#include "soc/io_mux_reg.h"
#include "driver/gpio.h"
These headers provide access to the ESP32’s hardware abstraction layer (HAL) and GPIO control, enabling direct manipulation of hardware registers.
-SPI Master Initialization Function
void spi_master_init() {
gpio_set_direction(VSPI_SCK, GPIO_MODE_OUTPUT); // Set SCK as output
gpio_set_direction(VSPI_MOSI, GPIO_MODE_OUTPUT); // Set MOSI as output
gpio_set_direction(VSPI_MISO, GPIO_MODE_INPUT); // Set MISO as input
gpio_set_direction(VSPI_CS, GPIO_MODE_OUTPUT); // Set CS as output
Configures the GPIO pins for SPI functionality. SCK, MOSI, and CS are outputs; MISO is input.
SPI3.user.val = 0;
SPI3.user.doutdin = 0; // Data is not bidirectional
SPI3.user.duplex = 0; // Full duplex mode disabled
SPI3.ctrl.val = 0;
SPI3.ctrl.wr_bit_order = 0; // Use MSB-first for writes
SPI3.ctrl.rd_bit_order = 0; // Use MSB-first for reads
Maps the GPIO pins to the corresponding SPI functions using the ESP32’s GPIO matrix. Then we Initialize the SPI3 user and control registers. Configures bit order and sets the transfer mode.
SPI3.clock.val = 0;
SPI3.clock.clkcnt_n = 7; // Clock divider setup
SPI3.clock.clkcnt_h = 3;
SPI3.clock.clkcnt_l = 7;
SPI3.cmd.usr = 0; // Ensure no ongoing SPI command
}
Configures the SPI clock, setting the frequency divider. Then we implement SPI Master Send Function:
void spi_master_send(uint8_t *data, int len) {
for (int i = 0; i < len; i++) {
SPI3.data_buf[i] = data[i]; // Load data into SPI buffer
}
Configures the data length for transmission, starts the SPI transaction, and waits for it to complete. The CS (chip select) line is toggled to signal the start and end of the transmission. Then Load the data to be transmitted into the SPI3 buffer.
SPI3.mosi_dlen.usr_mosi_dbitlen = len * 8 - 1; // Set data length (in bits)
gpio_set_level(VSPI_CS, 0); // Assert CS (active low)
SPI3.cmd.usr = 1; // Start SPI transaction
while (SPI3.cmd.usr); // Wait until transmission completes
gpio_set_level(VSPI_CS, 1); // Deassert CS
}
The LED Strip has three ports: Vdd, GND and Signal Pin.
Individually addressable LED strips operate using a digital data protocol. The Atmega328Pb sends a signal that represents 1s and 0s through the data pin. The LEDs interpret this signal to set their colors and brightness.
For WS2812 LEDs (Common Protocol): The microcontroller sends data in the form of pulse-width modulation (PWM). The duration of the pulse determines whether it’s a 1 or 0.
Figure 9: The Protocol In WS2812 LEDs
Protocol Timing: The WS2812 operates at a frequency of 800 kHz. Each LED receives a 24-bit data frame (8 bits each for Red, Green, and Blue channels).
In our Atmega328Pb, we write bare metal C to give signal data. The core function is ‘ws2812b_send_bit’, which is designed to send a single bit (0 or 1) to a WS2812B LED strip by generating a pulse-width-modulated (PWM) signal with precise timing. It uses inline assembly to directly manipulate the microcontroller’s hardware registers for accurate timing control.
if (bit != 0)
{
asm volatile (
"sbi %[port], %[pb] \n\t" // Set the LED pin high
"nop \n\t" // Delay to achieve ~0.8 μs
"nop \n\t"
"nop \n\t"
...
"cbi %[port], %[pb] \n\t" // Set the LED pin low
"nop \n\t" // Delay for the remaining ~0.45 μs
"nop \n\t"
...
::
[port] "I" (_SFR_IO_ADDR(LED_PORT)), // I/O port for LED
[pb] "I" (LED_PIN) // Pin connected to LED
);
}
else
{
asm volatile (
"sbi %[port], %[pb] \n\t" // Set the LED pin high
"nop \n\t" // Delay to achieve ~0.4 μs
"nop \n\t"
...
"cbi %[port], %[pb] \n\t" // Set the LED pin low
"nop \n\t" // Delay for the remaining ~0.85 μs
...
::
[port] "I" (_SFR_IO_ADDR(LED_PORT)), // I/O port for LED
[pb] "I" (LED_PIN) // Pin connected to LED
);
}
It is conntrolled by command: 2
else if (data == 2) // toggle light
{
light_status ^= 1;
if (light_status)
{
ws2812b_send(255, 255, 255);
ws2812b_send(255, 255, 255);
ws2812b_send(255, 255, 255);
}
else
{
ws2812b_send(0, 0, 0);
ws2812b_send(0, 0, 0);
ws2812b_send(0, 0, 0);
}
}
The water pump is controlled by a relay, powered by a 5-24V supply. The Atmega328 microcontroller uses the PC0 pin to control the relay, which in turn switches the pump on and off. The connection diagram is shown below:
Figure 10: The Configuration of Pump
The pump is controlled by command: 3
else if (data == 3) // toggle pump
{
pump_status ^= 1;
if (pump_status)
{
PUMP_PORT |= (1 << PUMP_PIN);
}
else
{
PUMP_PORT &= ~(1 << PUMP_PIN);
}
Generally, we use PWM to control the motor to enable it to rotate 90 degrees. There are three ports in the motor: PWM, Vdd and GND. The motor is powered by 5V. And the PWM singnal is controlled by the ATmega328Pb through PB1, its connection configuration is shown below:
Figure 11: The Configuration of the Motor
To controll the PWM singals, we coded the ATmega328Pb. There are two main functions:
. motor_initialize(int angle): Sets up Timer 1 in Fast PWM mode with a frequency of 50 Hz and an initial angle for the motor.
MOTOR_DDR |= (1 << MOTOR_PIN);
TCCR1A |= (1 << COM1A1) | (1 << WGM11);
TCCR1B |= (1 << WGM13) | (1 << CS11);
This function initializes the motor’s control settings using Timer 1 in Fast PWM mode and Configures the motor control pin (MOTOR_PIN) as an output by setting the corresponding bit in the data direction register (MOTOR_DDR):
ICR1 = 20000;
It sets the top value for the timer counter to 20000. This defines the PWM period. Assuming an 8 MHz clock and a prescaler of 8, this corresponds to a 50 Hz PWM frequency (commonly used for servo motors).
OCR1A = angle;
This part sets the initial duty cycle for the PWM signal on the OCR1A pin. The value corresponds to the desired initial angle of the motor.
.motor_angle(int angle): Converts a desired angle into a PWM pulse width and updates the timer’s compare register (OCR1A) to adjust the motor’s position.
This function adjusts the motor’s angle by setting the appropriate PWM duty cycle
int output = angle / 90 * 1000 + 500;
OCR1A = output;
This updates the OCR1A register to change the PWM duty cycle and adjust the motor angle accordingly.
It is controlled by command: 1
if (data == 1) // toggle door
{
door_status ^= 1;
if (door_status)
{
motor_angle(90);
}
else
{
motor_angle(0);
}
}
This diagram illustrates the integration of a Nano 33 microcontroller, an LED strip, and a battery pack in the wand basic circuit. The Nano 33 is powered by a 4.5–4.8V battery pack while sharing a common ground with the entire circuit. The LED strip, also powered by the same battery pack, receives control signals from the Nano 33 via the PB6 pin, enabling dynamic operation such as color and brightness adjustments. The setup emphasizes efficient power distribution and signal synchronization, with a shared ground ensuring proper circuit functionality.
Figure 7: The Pin Layout Of The Wand
To ensure consistent data reading from the gyro and IMU, we added interrupts to the Nano 33. This allows us to read data every 100ms, avoiding irregular and inconsistent readings caused by varying processing times of the data matrix. This improvement helps maintain the accuracy of gesture recognition and ensures the standardization of the data.
The walls of our smart furniture model are created using laser cutting. The core part is designed to accommodate a servo motor, which controls the door’s opening and closing. We modeled it to ensure the servo motor can be properly installed.
Figure 8: The 3D Module of The Wand
My wand requires a wide-voltage power supply using three batteries along with a buck converter to provide stable power for the Nano 33. The Nano 33 connects to an LED strip to provide feedback for different gestures, offering users a good interactive experience. Therefore, we designed a dedicated model for it. The model includes a centrally positioned battery compartment with a knob-style latch for easy battery installation and replacement, ensuring continuous usage. Additionally, we designed a groove for the LED strip to enhance the wand’s aesthetic appeal.
The LED Strip has three ports: Vdd, GND and Signal Pin. The connection configuration is shown before in Fig 7.
Individually addressable LED strips operate using a digital data protocol. A microcontroller (like Nano 33) sends a signal that represents 1s and 0s through the data pin. The LEDs interpret this signal to set their colors and brightness.
For a 1 bit: The signal stays high for approximately 0.8 μs. Then it goes low for approximately 0.45 μs.
For a 0 bit: The signal stays high for approximately 0.4 μs. Then it goes low for approximately 0.85 μs.
In our design, we write a function to create a “wave-like” animation effect on an LED strip controlled by the FastLED library. It will sequentially light up LEDs in a “wave” pattern along the strip. Only one LED is actively lit at any given time, while the previously lit LED is turned off. The animation stops when it completes one full pass along the length of the strip. The core codes are shown below:
if (!LED_flow_enable)
return;
if (LED_flow_counter >= LED_NUM)
leds[LED_NUM - 1] = CRGB::Black;
The last LED in the strip is turned off by setting its color to CRGB::Black.
FastLED.show();
The FastLED.show() function sends the updated LED data to the strip to reflect the changes.
LED_flow_enable = 0;
LED_flow_counter = 0;
else
If the wave animation is not complete (LED_flow_counter < LED_NUM), the function lights up the next LED in the sequence and turns off the previous one:
leds[max(LED_flow_counter - 1, 0)] = CRGB::Black;
The LED at the position LED_flow_counter - 1 (the previous LED) is turned off.
The max() function ensures the index does not go below 0 (to avoid accessing out-of-bounds memory).
leds[LED_flow_counter] = LED_flow_color;
The LED at the position LED_flow_counter (the current LED) is set to the specified wave color (LED_flow_color).
FastLED.show();
Sends the updated data to the LED strip to reflect the changes (turning off the previous LED and lighting up the current one).
++LED_flow_counter;
The counter is incremented to move to the next LED for the subsequent call of this function.
On every call to show_wave(), this function checks if the wave animation is enabled and either:
Basically, the whole reading process is shown below:
Figure 8: The Data Processing Flow Chart
To ensure consistent data reading from the gyro and IMU, we added interrupts to the Nano 33. This allows us to read data every 100ms, avoiding irregular and inconsistent readings caused by varying processing times of the data matrix. This improvement helps maintain the accuracy of gesture recognition and ensures the standardization of the data.
The interrupt function is:
extern "C" void TC3_Handler()
{
// Check if the match interrupt flag is set
if (TC3->COUNT16.INTFLAG.bit.MC0)
{
// Clear the interrupt flag
TC3->COUNT16.INTFLAG.reg = TC_INTFLAG_MC0;
show_wave();
}
}
These codes are to implement Timer3 in the Nano33 for trigerring interupt. We aim to read data every 100ms while not stop the data processing.
void configure_timer()
{
GCLK->CLKCTRL.reg = GCLK_CLKCTRL_ID_TCC2_TC3_Val; // select TC3 peripheral channel
GCLK->CLKCTRL.reg |= GCLK_CLKCTRL_GEN_GCLK0; // select source GCLK_GEN[0]
GCLK->CLKCTRL.bit.CLKEN = 1; // enable TC3 generic clock
// Configure synchronous bus clock
PM->APBCSEL.bit.APBCDIV = 0; // no prescaler
PM->APBCMASK.bit.TC3_ = 1; // enable TC3 interface
// Configure Count Mode (16-bit)
TC3->COUNT16.CTRLA.bit.MODE = 0x0;
// Configure Prescaler for divide by 2 (500kHz clock to COUNT)
TC3->COUNT16.CTRLA.reg |= TC_CTRLA_PRESCALER_DIV1024;;
// Configure TC3 Compare Mode for compare channel 0
TC3->COUNT16.CTRLA.bit.WAVEGEN = 0x1; // "Match Frequency" operation
// Initialize compare value for 100mS @ 500kHz
TC3->COUNT16.CC[0].reg = 4687;
// Enable TC3 compare mode interrupt generation
TC3->COUNT16.INTENSET.bit.MC0 = 0x1; // Enable match interrupts on compare channel 0
NVIC_EnableIRQ(TC3_IRQn); // Enable TC3 interrupt in the NVIC
// Enable TC3
TC3->COUNT16.CTRLA.bit.ENABLE = 1;
// Wait until TC3 is enabled
while(TC3->COUNT16.STATUS.bit.SYNCBUSY == 1);
NVIC_SetPriority(TC3_IRQn, 3);
NVIC_EnableIRQ(TC3_IRQn);
__enable_irq();
}
After reading the data accelerometer and gyroscope, we assign an index_counter to the data starting from 0 and accumulate for each instance of data by 1 to allow the calculation process to monitor the continuity of the input data, in the form of
[index, gyroX, gyroY, gyroZ, accX, accY, accZ]
We use ISR to push the data onto a queue to make sure the input data to form a continuous sequence of serial data list.
Below is the ISR handler logic we use to update data on the queue. We have a semaphore to avoid race conditions for checking the queue availability during insertion, popping and checking the queue sizes. When the queue modification is available and the queue is not Full, we simply push the data onto the queue. Otherwise if the queue is Full, we pop the first element from the queue. If we have a missing element (the gyroscope data is not continuous) then it indicates we have a missing element when collecting data, in this case we remove all elements from the queue to restart collecting data for calculation.
if (xSemaphoreTake(QUEUESemaphore, ENQUEUE_MAXWAIT) == pdTRUE) {
Serial.println("inserting sensory data %d", sensory_data.idx);
int queue_end_idx = GyroQueue.queue_end_idx;
if (isFull(GyroQueue)) {
dequeue(GyroQueue, &PriorSensoryData);
}
if (queue_end_idx + 1 != gyro_idx) {
clear_queue(GyroQueue);
}
enqueue(GyroQueue, &sensory_data);
Figure 9: Diagram and Mechanism of the cyclic queue
The sensory data is stored in a cyclic queue with a maximum capacity of N elements. The purpose of the queue is to allow O(1) insertion and deletion of sensory data. Moreover, a queue allows two concurrent threads to add and remove element from the queue without causing signiticant race conditions that’ll corrupt the memory. The queue is a datastructure made up of a buffer as a continuous chunk of memory, am index counter that indicate the index of the first element of the queue in the memory array, and the index counter that indicate the block after the last element of the queue in the array
typedef struct {
SensoryData_t data[BUFFERSIZE]; // Array to store data
int front; // Index of the front element
int rear; // Index of the rear element
int count; // Number of elements in the queue
int queue_start_idx;
int queue_end_idx;
} GyroQueue_t;
The storage queue is a chunk of memory of $N \times sizeof(SensoryData)$ . We use the start index and the end index to store the start and end element of the queue. Each time a new element is inserted, we update the end of the queue as $(endIndex + 1) \pmod N$. If the queue is Full, we pop the least recently inserted element from the queue by updating the start index to $(startIndex + 1) \pmod N$. Below are the implementation of insertion and deleting element (the 6 axis gyroscope data) into and from the queue after checking if the queue is Empty or Full
int enqueue(GyroQueue_t* queue, SensoryData_t* input_data) {
if (isFull(queue)) {
return 0; // Queue is full
}
queue -> data[queue -> rear] = *input_data; // Add element to the rear
queue -> rear = (queue -> rear + 1) % BUFFERSIZE; // Update rear (wrap around)
queue -> count++; // Increase element count
queue -> queue_end_idx = input_data -> idx;
return 1;
}
// Dequeue an element
int dequeue(GyroQueue_t* queue, SensoryData_t* output_data) {
if (isEmpty(queue)) {
return 0; // Queue is empty
}
SensoryData_t output_data_tmp = queue->data[queue -> front]; // Get the front element
output_data -> idx = output_data_tmp.idx;
output_data -> gyro_Angular_X = output_data_tmp.gyro_Angular_X;
output_data -> gyro_Angular_Y = output_data_tmp.gyro_Angular_Y;
output_data -> gyro_Angular_Z = output_data_tmp.gyro_Angular_Z;
output_data -> gyro_X = output_data_tmp.gyro_X;
output_data -> gyro_Y = output_data_tmp.gyro_Y;
output_data -> gyro_Z = output_data_tmp.gyro_Z;
queue -> front = (queue -> front + 1) % BUFFERSIZE; // Update front (wrap around)
queue -> count--; // Decrease element count
if (queue -> count > 0) {
queue -> queue_start_idx = queue->data[queue->front].idx;
}
return 1;
}
In the main calculation loop, we keep popping elements from the queue and compare with the previous element that we popped. We use a flag Recog
initialized to be 0 to indicate whether we’re collecting data for pose recognition. If $Recog == 1$, we place the new data into a buffer that we use for performing pose recognition. If the new data is not continuous in index
$(IndexNew \neq Indexold + 1) $ with the previous data, we clear the buffer and set the flag Recog
to be 0
If the new data has an index that’s continuous with the previous data, we consider the below 3 cases:
Recog=1, STABLETIMEGAP =0
(STABLETIMEGAP++)
(STABLETIMEGAP > Threshold)
, we think the user has finished waving the wand and set Recog = 0
After we enter and quit the Recog
by setting the flag to be 1 and then back to 0, we’ve stored the time sequence data into an array for pose recognition. We currently conduct pose recognition using a version of thresholding according to the sharpe-turns recognized in the series data. This method is shown to be quite effectively and more importantly, deployable to be performed in realtime on the edge compared to Deep Learning algorithms (which are not practical to be implemented on such a lightweight platform given the amount of effort we’ll need)
Specifically, we measures the number of sharp turns as fingerprints to classify the pose recognition into different types of commands. Consider the two neighboring data D1, D2
where D2.idx=D1.idx + 1
, we think of they have a sharp turn if
Based on the number of sharp turns found in the buffer, we could classify the waving action into drawing different shapes in the air. Below are samples of recognizing triangles and quadrilaterals
Figure 10: Example of Pose Recognition
The implementation of the above formula is illustrated below
static int compute_angular_Change(SensoryData_t* prior_data, SensoryData_t* current_data) {
int d_angular_x = current_data -> gyro_Angular_X - prior_data -> gyro_Angular_X;
int d_angular_y = current_data -> gyro_Angular_Y - prior_data -> gyro_Angular_Y;
int d_angular_z = current_data -> gyro_Angular_Z - prior_data -> gyro_Angular_Z;
int angular_shift = sqrt(d_angular_x * d_angular_x + d_angular_y * d_angular_y + d_angular_z * d_angular_z);
return angular_shift;
}
To further improve the throughput of calculation, we improved the procedure of taking elements from the queue and performing classification into one function, where we constantly keep track of the previously popped element from the queue and compare it with the new element for angular shift. If we’re currently in counting Recog==1
mode, we increment the number of sharp turns by 1.
static int execute_calculations(SensoryData_t* current_data, SensoryData_t* prior_data) {
if (!dequeue(GyroQueue, current_data)) {
// the queue is currently empty
return -1;
} else {
//the queue is not empty
int result = -1;
int prior_idx = prior_data -> idx;
int current_idx = current_idx -> idx;
if (current_idx != prior_idx + 1) {
// discontinuous cnt caused by package loss, replace the PriorSensory data with new data
copy_sensory_data(current_data, prior_data);
if (TRACKING) {
STABLETIMEGAP += current_idx - prior_idx;
}
} else {
// the data is purely continuous
int angular_shift = compute_angular_Change(prior_data, current_data);
if (angular_shift > ANGULAR_THREHSOLD) {
if (TRACKING) {
STABLETIMEGAP = 0;
SHARPTURNCOUNTER += 1;
} else {
TRACKING = 1;
STABLETIMEGAP = 0;
SHARPTURNCOUNTER = 0;
}
}
copy_sensory_data(current_data, prior_data);
}
if (STABLETIMEGAP > GAP_TILL_EXIT) {
result = SHARPTURNCOUNTER;
//reset the time gap to be zero to allow restarting
STABLETIMEGAP = 0;
return result;
}
}
}
This will ensure a high throughput of recognition to be much smaller than the gap of sampling data from IMU, which avoid data stagnation.
Simply put, we use an ESP32 and Nano 33, both connected to an ATmega, for wireless communication, as they are equipped with Wi-Fi modules. The Nano 33 recognizes hand gesture commands and transmits them to the ESP32 via Wi-Fi. The ESP32 then sends the commands to the ATmega328Pb through SPI.
We set up a private network, configuring the ESP32 and Nano 33 to be on the same local network. Each device is assigned its own IP address, which we define in the code for both transmitting and receiving functionalities. This setup ensures continuous communication between the devices.
const char* ssid = "R9000P";
const char* password = "12348765";
These variables store the network name (ssid) and its corresponding password (password) required to connect to the Wi-Fi.
const uint16_t port = 1234; // Port to listen on
WiFiServer server(port);
The WiFiServer object, server, will handle incoming client connections on the defined port.
void connectToWiFi()
{
WiFi.begin(ssid, password); // Start connecting to the Wi-Fi network
while (WiFi.status() != WL_CONNECTED) // Wait until the connection is established
{
delay(1000); // Retry every second until connected
}
}
WiFi.begin(ssid, password) initiates the connection to the Wi-Fi network using the provided credentials. The while loop ensures the program waits until the device is connected to the network by continuously checking the connection status (WiFi.status() == WL_CONNECTED).
server.begin();
WiFiClient client = server.available();
if (client)
{
// Wait for and process client data
}
If a client connects, the server enters a loop to listen for data sent by the client. Once the communication ends, the client connection is stopped using client.stop().
Similarly, we first need to declaring Wi-Fi Credentials as we did before. The code is same. The core parts are:
const char server_py[] = “192.168.137.218”; const uint16_t port_py = 1234;
server_esp32 and server_py represent the server addresses for communication with the ESP32 and a Python server, respectively.
port_esp32 and port_py specify the ports to be used for the connections.
2. Connecting to Servers
The connectToServer function handles the connection to a specified server:
```c
bool connectToServer(WiFiClient* client, const char * server, const uint16_t port)
{
if (DEBUG)
{
Serial.print("Connecting to server ");
Serial.print(server);
Serial.print(":");
Serial.println(port);
}
if ((*client).connect(server, port)) // Attempt connection
{
if (DEBUG)
Serial.println("Connected to server!");
return true; // Connection successful
}
else
{
if (DEBUG)
Serial.println("Connection to server failed.");
return false; // Connection failed
}
}
Uses a WiFiClient object to establish the connection. Checks and logs the connection status for debugging.
void sendMessage(WiFiClient* client, const char* message)
{
if ((*client).connected())
{
if (DEBUG)
{
Serial.print("Sending: ");
Serial.println(message);
}
(*client).println(message); // Send the message
}
else
{
if (DEBUG)
Serial.println("Client not connected. Cannot send message.");
}
}
Verifies that the client is connected before sending a message using println.
HRS01: The motor can successfully turn on and off the designed door Validation:
HRS02: The LED lights on the wand can respond to different commands with various effects, achieving flowing light effects instead of just functioning as a basic on/off LED. Validation:
We waved 100 times for different postures and records their recognized results. The graph below shows the simulation results:
Figure 10: The Simulation Results
The Summary Table:
Posture | Total Test | Accuracy | Misrecognize Gesture 1 | Misrecognize Gesture 2 | Misrecognize Gesture 3 |
---|---|---|---|---|---|
1 | 100 | 90% | 0 | 5 | 5 |
2 | 100 | 89% | 7 | 0 | 4 |
3 | 100 | 95% | 3 | 2 | 0 |
The Test result shows that we perfectly meet our exception.
SRS02: Achieve a response time under 200 ms, allowing users to feel immediate feedback from devices. Validation:
SRS03: Successfully achieve wireless communication Validation:
We successfully completed all our goals and designs on time. From wireless communication to gesture recognition and data processing, we met the requirements and objectives we initially set. However, there are still areas for improvement:
LED Strip Effects: The LED strip effects we implemented using bare C are quite limited. In the future, we can explore adding more effects and flashing patterns to more realistically simulate the spell-casting process.
Gesture Recognition: We opted against using ML-based methods due to limited on-device computational power. Instead, setting thresholds allowed us to achieve the desired performance. For more complex gestures, we could consider transmitting data to a computer via Wi-Fi for processing and then sending it back. However, this would increase both development time and response latency. As a result, the threshold-based method remains the most optimal solution for now.
Smart Furniture Models: We can enhance the furniture model by integrating more smart devices to expand functionality.
These improvements would make the system more robust and feature-rich.
The team has made substantial progress on several critical tasks this week, continuing to move closer to achieving our project goals. Key developments include successful communication with sensors, early stages of wireless module integration, and preparations for component integration testing. Our focus is on refining individual components, with a special emphasis on ensuring compatibility across the different hardware modules. We are on track to integrate these modules in the upcoming sprint
This week, the team has made excellent progress, achieving several important milestones. Notably, we have successfully tested and integrated key hardware communication methods, including wireless communication, sensor data reading, and LED control. As we prepare for component integration, the next focus will be on comprehensive testing and ensuring the modules work cohesively. We are on schedule to advance to full system integration in the next sprint.
Gyroscope Reading
This week, we completed the draft design for the home model, which was showcased successfully. However, due to the unavailability of actual components, we could not measure dimensions and therefore have not yet designed the wand model. We secured access to the RPL lab and are currently evaluating options to either 3D print the home model or use laser cutting to assemble the basic shell as soon as possible. Additionally, we studied the Nano 33 datasheet to understand Bluetooth communication and the IMU module, ensuring that future communication between wands can be implemented effectively. Current Status
The project is making steady progress with key foundational tasks completed and a clear roadmap for upcoming activities. The home model design has been finalized, and we are preparing for its physical prototyping through either 3D printing or laser cutting. Access to the RPL lab has been obtained, facilitating the next steps for hardware assembly. While the wand model design is pending due to the absence of necessary components for measurement, we have advanced in understanding the technical aspects of the Nano 33, including Bluetooth communication and the IMU module. Additionally, we tested the basic communication capabilities of the Nano 33 by successfully controlling an LED light via remote commands. This lays a solid groundwork for implementing inter-wand communication in the future. The focus now is on transitioning from design to prototyping and integrating hardware and communication systems.
We are still working on the posture recognition Test and reading the data sheet of the Nano 33. We are planning to fix basic posture recognition next week and make a demo trial test on Nano 33.
Test Existing implementation for gyroscope data reading
Have tested the functionality of the LED strip.
We finalized the selection of necessary components, including in-depth discussions and considerations for specific details such as the power supply, MCUand IMU&Gyroscope module.
Additionally, we updated our project implementation approach and identified new items to be procured. For the posture recognition module, we optimized the use of the gyroscope and decided to incorporate a new MCU, the Nano 33, which combines Bluetooth communication and a gyroscope, allowing us to reduce costs. This decision was reached after extensive discussions within the team and the TA’s advise. We also clarified the allocation of responsibilities among team members, completed the project contract, strengthened team cohesion, and established our final objectives.
Home automation is increasingly transforming modern households with smart technology, enabling users to control and monitor their home environment conveniently. The goal of the team’s project is to design and construct an innovative home automation centered around an interactive smart wand, leveraging the strengths of teamwork and technical expertise. The project’s design scheme includes an integrated electronic system that supports a range of functions, allowing users to control household devices such as lights, doors, and pet water dispensers through customized wand gestures and commands wirelessly. The project is carried out by the team, each contributing to various aspects of the design, with the flexibility to choose hardware components within a specified budget.
The rapid development of smart technology has revolutionized the way people interact with their living spaces, providing convenience, efficiency, and enhanced control over daily tasks. However, traditional interfaces like mobile apps or voice commands can sometimes feel impersonal or cumbersome, especially for users looking for a more intuitive and engaging experience. Inspired by the concept of bringing magic into everyday life, the team’s project aims to create a novel approach to home automation through a wand-based control system. This allows users to interact with household devices—such as lights, doors, and pet water dispensers—through simple, magical gestures. By combining the familiarity of Bluetooth and Wi-Fi technology with an innovative control interface, the team’s goal is to make smart home devices more accessible, enjoyable, and efficient, transforming ordinary tasks into experiences that feel both magical and practical.
Using gesture and pose recognition algorithms, the team aims to achieve convenient and smooth control of all home equipments through the wand. This would require capturing movements of users through sensors on the wand and matching them effectively with the “fingerprint” movements previously recorded for performing each action. To achieve this, the team need to integrate controlling units from both the wand and devices the team aims to control in a wireless setting.
To summarize, the primary objective of the project is to develop a wand-controlled smart home system capable of Bluetooth and Wi-Fi communication with compatible devices, bringing a seamless and engaging experience to users. The system comprises microcontrollers, communication modules, and sensors to achieve reliable interaction with various smart home components. The wand utilizes an ESP32 microcontroller and a Bluetooth module to manage Wi-Fi and Bluetooth communication, LED feedback, and gesture recognition, while the home devices - such as lights and doors - are equipped with Wi-Fi and Bluetooth modules to enable smooth connectivity.
Figure 1: Block Diagram of The Wand System
Figure 2: The Design Schetch
The smart home controller system allows users to control home appliances, specifically lights and doors, using pose recognition derived from gyroscope and IMU data when the user is waving the wand. The system processes sensor data to recognize gestures and sends commands over Wi-Fi and bluetooth to perform predefined actions on connected devices. The software part of the design include algorithm for pose recognition based on gyroscope and IMU data, communication between the wand and smart devices, and control mechanims of smart devices using the same MCU.
The team’s target audiences are smart home users who prefer voiceless remote control of all devices. Traditionally, smart home systems are managed through touch interfaces, voice commands, or smartphone applications, but these control methods may not be universally accessible or convenient. Therefore, the team introduces the wand as a universal agent for effortless smart home management.
A detailed description of software requirements of different features are provided below.
The system shall process gyroscope and IMU data to identify specific user gestures for controlling appliances. The system shall recognize at least three distinct poses for each action:
The team plans to use pattern similarity matching algorithms to perform pose recognition of waving the wand. The team starts recording the gyroscope and IMU data when witnessing a rapid change in captured acceleration by the sensor. The accuracy of recognition should exceed 80%. If traditional methodologies are not applicable, the team could also try Machine Learning based approaches to perform classification task on different posture commands of waving the wand from gyroscope and IMU data.
Communication is a critical part of the project since the wand needs to send commands to the smart devices using wireless communication protocols. The team decided to choose bluetooth and Wi-Fi due to its wide usage and the mature technology ecosystem supporting it.
In order to control devices effectively in the smart home setting, the system shall be able to:
Both of the above actions could be achieved through the use of PWM, I2C, UART or simple GPIO output on the MCU that controls the smart home devices. Additionally, the system shall send feedback to the user indicating successful pose extraction and action execution or error messages if the action fails.
In order to provide users with feedback on each command and action, the team also plan to integrate error handling mechanisms. The error handler should have the following functionalities:
The hardware design and implementation of the team’s wand-controlled home automation system consists of multiple integrated components to enable seamless control over smart home devices. The primary hardware structure includes modules for gesture recognition, wireless communication, and device-specific control mechanisms.
At the core of the team’s system is the ESP32 microcontroller, responsible for transmitting data from the wand to various peripheral devices. The ESP32 works alongside the ATmega328PB, which serves as the main controller for both the wand and the peripheral controllers. These microcontrollers coordinate the data flow and manage command execution.
For gesture recognition, the GY-521 IMU & Gyroscope module detects hand movements, allowing the wand to accurately interpret gestures, such as circles and triangles, for controlling different smart devices. To provide users with feedback during command execution, an LED strip is integrated into the wand, acting as a visual indicator.
Device-specific modules are included to enable interaction with household components: the MG995 servo motor for door control, a water pump for the pet water dispenser, and an HC-05 Bluetooth module for establishing communication between the wand and other Bluetooth-enabled devices.
To showcase these functions, a 3D-printed home model with embedded responsive devices was created, allowing for realistic testing and demonstration of the wand’s capabilities within a simulated environment. This setup enables the team to fine-tune hardware performance and ensure smooth, reliable operation in a practical setting.
The power supply for the home automation system is provided by three 1.5V AA alkaline batteries connected in series, offering a total source voltage of 4.5V. This power configuration was chosen due to restrictions on lithium battery usage, as well as the need for a reliable and safe energy source suitable for indoor applications. The 4.5V output is sufficient to power the low-voltage modules in the system, including the ATmega328PB microcontrollers, IMU and gyroscope module, and the Bluetooth module, which operate efficiently within this range. The team plans to build a voltage divider for the ESP32 board, which takes in a voltage of 3.0-3.6V.
Compared to single-cell batteries, the series connection provides a stable and sustained voltage, which enhances the consistency of control signals for the smart devices. Alkaline batteries were selected not only for their accessibility and ease of replacement but also for their low risk of overheating and high compatibility with consumer electronics. This choice supports both safety and user convenience, making it easy for users to replace batteries as needed.
A key consideration in this setup is the overall power management, especially given the continuous operation of modules such as the Bluetooth, Wi-Fi, and IMU components. To mitigate potential issues with voltage drops, capacitors are placed in parallel with critical components to smooth the power supply and stabilize voltage levels during sudden power demands, such as during Bluetooth transmission or LED strip activation.
The stability and longevity of the alkaline battery source have been tested to meet the expected usage cycle of the system, ensuring continuous operation without frequent replacements. Further analysis of the power source’s performance will be discussed in the Results & Discussion section.
The team plans to power the devices in the home model with a 120v AC to 5v DC converter. This suits the realistic home scenario and could provide the smart devices with a steady power source.
The ATmega328PB microcontroller is critical to the data processing and command execution within the team’s wand-controlled home automation system. On the wand, the ATmega328PB reads rotation and acceleration data from the GY-521 IMU & Gyroscope module, interpreting these inputs as gesture commands. It then sends the processed commands to the ESP32 or bluetooth module for wireless transmission to various devices.
On the peripheral control unit, a separate ATmega328PB serves as the controller for various home automation devices, including the light bulb, servo motor (for door control), and water pump (for the pet water dispenser). This microcontroller listens for incoming commands from the ESP32 and directly controls the connected devices to execute the desired actions.
The ESP32 board on the wand shall be able to transmit every command it reads from the ATmega328PB board on the wand to the other ESP32 board installed on the peripheral control unit.
The ESP32 board on the peripheral control unit shall be able to send every command it receives from the other ESP32 board to the ATmega328PB board on the peripheral control unit.
The Bluetooth communication in the team’s home automation system is facilitated by the HC-05 Bluetooth module, which enables wireless data transfer between the wand controller and various Bluetooth-enabled home devices. This module was selected for its ease of integration, reliable connectivity, and compatibility with microcontrollers such as the ATmega328PB and ESP32 used in the team’s setup.
The HC-05 operates within the standard Bluetooth 2.0 protocol, offering a stable connection range of up to 10 meters in indoor environments, which is ideal for home automation scenarios. It supports UART (Universal Asynchronous Receiver/Transmitter) communication, which allows seamless data transmission with microcontrollers at a configurable baud rate. In the team’s design, the baud rate is set to 9600 bps to ensure a balance between transmission speed and stability, critical for real-time control of smart devices.
Figure 3: The Bluetooth Module Connection
Powering the HC-05 requires a 3.3V - 6V input, which aligns with the output from the team’s voltage-regulated power supply. The module’s onboard voltage regulator and level shifter make it compatible with both 3.3V and 5V logic levels, simplifying integration with the ATmega328PB and ESP32 controllers. The HC-05’s low power consumption is essential for maintaining overall system efficiency, especially in a battery-powered setup, and contributes to the extended operational lifespan of the wand.
The HC-05 also provides flexibility for future expansion, allowing additional Bluetooth-enabled devices to be paired with the system.
The module’s pairing and communication status are indicated by an onboard LED, which blinks during pairing and remains solid when a connection is established. This visual feedback ensures that users can easily confirm the connection status, improving the reliability and user-friendliness of the system.
In summary, the HC-05 Bluetooth module is a critical component in enabling wireless communication across the smart home setup, providing reliable, low-power, and flexible connectivity for a smooth and responsive user experience.
The GY-521 IMU and Gyroscope Unit, featuring the MPU-6050 sensor, is responsible for detecting gestures and movements of the wand, enabling intuitive control over smart home devices. This module was selected for its high sensitivity and accuracy in motion detection, which is essential for recognizing specific gestures such as circles and triangles, which trigger commands for controlling devices like lights, doors, and pet water dispensers.
The MPU-6050 sensor integrates a 3-axis gyroscope and a 3-axis accelerometer, allowing it to detect both rotational motion and linear acceleration. This capability enables the wand to capture complex gestures with a high degree of accuracy. The sensor communicates with the ATmega328PB microcontroller on the wand through an I2C interface, which is configured at a standard data rate to ensure reliable data transmission with minimal latency.
Operating at a voltage range of 3.3V to 5V, the GY-521 is compatible with the power supply configuration of the wand, simplifying integration and ensuring stable operation. The module’s power consumption is optimized for battery-powered applications, which supports the overall energy efficiency of the system, prolonging battery life during continuous use.
Figure 4: The Schematic Diagram of MPU-6050
To achieve smooth and accurate gesture detection, the MPU-6050 is calibrated to filter out ambient vibrations and unintentional movements. The onboard Digital Motion Processor (DMP) of the MPU-6050 enables it to process complex motion algorithms directly on the sensor, reducing the computational load on the ATmega328PB and enhancing real-time response to user gestures.
The GY-521 module is mounted securely within the wand to minimize unwanted movement and ensure consistent data capture. Its sampling rate is set to provide real-time responsiveness to gestures, with minimal delay, ensuring an engaging user experience. The IMU data is processed by the ATmega328PB to interpret each gesture and send the corresponding control signal via the Bluetooth module to the designated home device.
By incorporating the GY-521 IMU and Gyroscope Unit, the wand is capable of accurately detecting and interpreting a variety of gestures, allowing users to control smart devices with intuitive motions. This setup not only enhances user convenience but also provides a novel and engaging method of interaction within the smart home environment.
The Status Indicator is a crucial component of the wand, providing real-time visual feedback to users on the command they just executed. This feedback ensures that users can easily confirm whether their gesture was accurately recognized and translated into the intended action. To meet these requirements, the team selected the WS2812B LED strip, a highly versatile LED solution with individually addressable LEDs that allow precise control over color and brightness.
The WS2812B LED strip operates by controlling each LED individually via addresses, enabling the system to display distinct colors and patterns corresponding to different commands. For example, a specific color may indicate that the wand has successfully sent a command to turn on a light, while another color may indicate that a command to open the door was recognized. This level of customization ensures that the user can intuitively understand the wand’s status and feel confident about their interactions.
Operating at 5V, the WS2812B LED strip is compatible with the wand’s power supply and can be directly controlled by the ATmega328PB microcontroller, which sends data to the LED strip to dynamically adjust colors based on recognized commands.
Figure 5: The Internal Circuit of LED Strip
In addition, the WS2812B supports a wide color range and fast response time, allowing it to provide smooth transitions and immediate feedback with minimal latency. This is particularly important in gesture-based control systems, where users expect instantaneous visual confirmation. The strip’s addressable nature also means that different sections of the LED strip can light up in different colors or patterns, offering further customization options to enhance the user experience.
Figure 6: The Color Control of LED Strip
By integrating the WS2812B LED strip, the status indicator effectively meets the requirement of providing clear, intuitive, and dynamic feedback, allowing users to seamlessly interact with their smart home devices and understand the wand’s operational state at a glance.
The MG995 servo motor serves as the core mechanism for controlling rotational movements required for opening and closing the door within the team’s smart home automation system. The motor operates on direct current and is widely used in applications requiring precise angular control, such as robotic arms, pan-tilt mechanisms, and automated door systems. The MG995 motor features high torque and a metal gear structure, which provides enhanced durability and performance stability under continuous load.
The motor includes three connection pins: power (VCC), ground (GND), and signal (SIG). The signal pin accepts PWM input signals to control the angular position of the servo arm, which can rotate between 0° and 180°. By adjusting the duty cycle of the PWM input, the motor’s position can be controlled with high precision, making it suitable for accurately managing the open and closed states of doors in a home setting.
Figure 7: The Configureation of The MG995 servo motor
The MG995 operates effectively within a voltage range of 4.8V to 7.2V, with an operating torque of 9.4 kg·cm at 6V, ensuring it has enough strength to manage typical household doors. The motor’s maximum stall current is 1.2A at 6V, which has been considered in the design of the power management system to prevent overloading. The motor’s position is updated every 20 ms based on the PWM signal, allowing smooth and reliable movement control.
Figure 8: The PWM Setting of Motor Control
To accommodate the MG995’s power and stability requirements, the motor is connected to a dedicated PWM output channel on the ATmega328PB peripheral controller, which regulates the door’s opening and closing actions. This setup enables a controlled and stable door movement, while the microcontroller ensures the appropriate signals are sent based on user commands received via the wand’s Bluetooth connection.
For enhanced stability and durability, the MG995 servo motor integrates with a mount on the smart home model, allowing it to securely handle repeated opening and closing actions without loss of precision. The setup provides a robust and efficient solution for automated door control within the home automation system.
Alongside the sophistication on the motor control circuit, the performance of motors has also been improved from another degree – the overall stability of the driving system. The practical approach to achieve better system stability is to add a decoupling capacitor to the circuit. To understand the functionality of decoupling capacitor, one should look into the internal structure of motors. The motors contain magnetic brush that might generate electromagnetic pulses. The pulses are likely to impair the circuit’s components, such as the microcontroller and motor control module. As the frequency of PWM power sources increases, the noise generated by the operating motors intensifies, and this is detrimental to the system. The decoupling capacitor aims to provide AC current flows near the load component to mitigate the detrimental effect of the noise. This capacitor is no simple capacitor: it contains equivalent se- ries inductance (ESL) and equivalent series resistance (ESR), and the equivalent circuit arrangement can be depicted as shown in following figures. In this case, I1 is AC current source, and its frequency equals the PWM source frequency. If the overall capacitance of the circuit is sufficiently large to be com- mensurate to a voltage source, the ESL and ESR are expected to be as small as possible to stabilize the circuit voltage.
Figure 9: Technical details regarding decoupling capacitors
The material of which the capacitor is made of is also one important consideration. In Figure, several capacitors that share the same capacitance but differ in materials are compared in terms of their impedance with changing frequency. In the figure, the lowest point on each curve corresponds to the resonance frequency of the capacitor, and at this point the impedance is the ESR. Conclusion can be drawn that the higher the resonance frequency, the lower the ESR; the lower the impedance, the lower the ESR. From these views, it can be indicated from the figure that ceramic capacitors outperform other capacitors in terms of ESR and ESL, so the author chooses a ceramic capacitor to decouple the circuit.
The light bulb is only a demo, so the team plans to use part of the LED strip from the status indicator to represent the light bulb.
The Water Pump is designed to provide fresh water to pets as part of the home automation system. To meet this need, the team selected a pet water pump available on Amazon, chosen for its compatibility with remote control operations and reliability in consistent water delivery. When the power is turned on via a command from the wand, the pump is activated to dispense water for pets, ensuring an efficient and controlled flow.
Selected for its quiet operation and compact design, the pump ensures minimal noise disturbance and easy integration within the 3D-printed home model, allowing realistic testing and demonstration of the water dispensing function. The pump’s simple on/off functionality aligns with the pet care automation goals, providing fresh water with minimal supervision and user effort.
Additionally, the water pump’s motor is protected by a back-emf diode to prevent voltage spikes upon deactivation, thus preserving the longevity of both the pump and the control circuitry. This protection is especially critical, as the wand-controlled system may issue commands in quick succession, requiring the pump to respond without risk of hardware failure.
The inclusion of the water pump in this automation system supports a convenient and reliable method for pet hydration, demonstrating how smart home technology can enhance pet care routines as part of a connected home environment.
ATmega328PB
Purpose: Serves as the core processing unit for both the wand and peripheral controllers, managing sensor data and controlling device-specific actions (e.g., light, door, and water pump). The ATmega328PB on the wand processes gesture data from the IMU and communicates with the ESP32 for Bluetooth commands.
ESP32
Purpose: Handles Bluetooth communication between the wand and home devices. The ESP32 allows for wireless control over multiple devices, enabling the wand to send commands to various wireless components in the home setup.
IMU and Gyroscope
GY-521 IMU (MPU-6050) Purpose: This module provides 3-axis gyroscope and accelerometer data to detect the wand’s movements and gestures, such as circles or triangles, enabling intuitive command control. The IMU and gyroscope data is processed by the ATmega328PB to interpret each gesture and send the corresponding command via Bluetooth.
Bluetooth Module
HC-05 Bluetooth Module Purpose: Allows the wand and home devices to communicate wirelessly. This module is used on both the wand and the home devices for seamless data exchange, which is crucial for controlling devices such as lights, doors, and water pumps.
Status Indicator LED Strip
WS2812B LED Strip Purpose: Provides real-time feedback to the user by displaying different colors and patterns based on the gesture command being executed. The individually addressable LEDs allow dynamic color changes that correspond to specific wand commands.
Servo Motor
MG995 Servo Motor Purpose: Controls the opening and closing of doors as part of the home automation setup. The MG995’s high torque allows it to operate reliably for controlling household doors, activated by commands from the wand.
Water Pump
Pet Water Dispenser Pump Purpose: Pumps water for pet hydration when activated. Selected for its compatibility with small-scale, controlled water dispensing and quiet operation, making it suitable for home environments.
Lighting
12V Smart LED Bulb Purpose: Provides smart lighting control in the home model, activated by commands sent from the wand. This bulb is compatible with Bluetooth control and enables various lighting effects, controlled remotely by the wand’s commands.
Power Supply
AA Alkaline Batteries (for Wand)
Purpose: Provides reliable power to the wand without the risk associated with lithium batteries. AA batteries are easily replaceable and compatible with the low-voltage requirements of the wand’s microcontroller and sensor modules.
DC Power Supply (for Home Model)
Purpose: Supplies stable power to components in the home model, such as the servo motor, water pump, and LED strip. This ensures consistent operation during testing and demonstration.
Component | Description | Link | Datasheet | Number |
---|---|---|---|---|
ESP32 | controller for sending data from the wand to the peripheral controller | datasheet | 2 | |
ATmega328PB | main controller for both the wand and the peripheral controller | datasheet | 2 | |
GY-521 IMU&Gyroscope | IMU and gyroscope module for the wand to detect user’s wavement | Amazon link | datasheet | 1 |
LED strip | status indicator for the wand & demo for light bulb | Amazon link | datasheet | 1 |
MG995 servo motor | server motor for open/close the door | Amazon link | datasheet | 1 |
Water pump | pump for water dispenser | Amazon link | 1 | |
Bluetooth module | HC-05 bluetooth for connecting other bluetooth devices | Amazon link | datasheet datasheet |
1 |
Show that the wand can successfully control household devices, such as lights, doors, and the pet water dispenser, through gestures and wireless connectivity.
Enable the wand to recognize different hand gestures - such as drawing circles, triangles, etc. - to execute corresponding commands. The wand will accurately interpret these gestures to control specific smart furniture, showcasing precise command recognition and responsiveness.
Provide a visually engaging LED feedback on the wand when commands are executed. When a gesture or “spell” is performed, the wand will emit dynamic LED patterns, adding a “magical” and visually impressive element to the user experience.
Present a 3D-printed model of a home environment, embedded with the team’s custom-designed responsive smart furniture. This model will allow the team to visually and functionally demonstrate the wand’s ability to interact with various home automation components in a realistic setup.
Ensure smooth communication between the wand and multiple wireless devices within the model, demonstrating responsiveness and reliability in real-time scenarios.
Showcase the wand’s capabilities, including gesture recognition, LED feedback, and the ability to interact with smart furniture within the model, providing a user-friendly interface for controlling home automation.
The team’s project needs to address the following challenges from both the domain of both software and hardware design:
A stepwise procedure of the team’s project is shown below
Figure 10: The stepwise procedure
Due to the difficulty of labeling gyroscope based training datasets and the limited computational resources of the wand. It’s not feasible to perform posture recognition based on gyroscope time-serial data on the edge. On the other hand, performing cloud based computation would introduce a huge delay and a high requirement for network stability, which is against the team’s original intention of building a convenient universal controller for smart home setting. In order to perform accurate similarity-based matching of postures and deploy it on the want, the team will experiment with the following approaches:
Predefined Shape Recognition using Thresholding:
One of the easiest and most applicable approaches for simple pose recognition would be to limit recorded postures into a few predefined postures such as drawing circle, square or a straight line. The team then set threshold of detecting criticle components of different shapes to decide which shape the user is currently drawing. For example, the team would expect a smooth and periodic change of signal for drawing a circle, a constant gyroscope read shift for drawing a straight line and different peak counts in gyroscope read changes when drawing rectangles or triangles. The advantage of such method is that it’s easy to implement as a prototype. However, the simple thresholding would limit the team’s choices of postures to only a few predefined shapes, inhibiting the user of customizing their posture at will. Moreover, each shape would only be applicable to one instruction since the predefined thresholding could not distinguish drawing a Equilateral Triangle from drawing an Obtuse Triangle
Fourier-based Analysis:
A more advanced method for posture recognition of time-serial data such as high-throughput gyrocope sequence data would be to use Fourier transformation. For example, weighted fourier combiner filters have been shown to be effective in accurately reconstructing the 3D orientation of IMU located on the lower trunk of a subject during treadmill walking (Bonnet,V. 2019). Similar method has also been effective in predicting head posture detection and analysis using Fourier Transforms (Severin,I. 2020). In brief, first preprocess the signal by splitting the time series data over the past few seconds in fixed-width sliding windows, and then apply fourier transform was applied to each window, which will provide us with a sum of sine and cosine waves of different frequencies and phases:
\[S_{RPY} (t) = \frac}}{2} \sum_{n=- \infty}^{ \infty} ( \cos(2 \pi nt) + {b_{n}} \sin (2 \pi nt)})\]Which provide us with a fingerprint of different frequencies and their relevant intensities and phase differences. The team could then split the normalized frequency distribution into a vector, where each entry represent the relevant intensity in a frequency domain. For example,
\[V_{i} = \sum_{a_{i} < f < a_{i + 1}} \frac{A_f}{\sum_{f}A_f}\]The team could then apply cosine similarity score to measure the similarity values of the posture frequency vector X with existing postures vector Y
\[Similarity(X, Y) = \frac{\sum_{i=0}^{n} X_{i} * Y_{i}}{Norm(X) * Norm(Y)}\]KNN Classification on Reduced Rank Data Another methodology of performing posture matching and classification is through non-parametric learning approaches such as K-Nearest Neighbors, which introduces much less computational load than parametric machine learning models. K-Nearest Neighbors have already shown to be effective in performing complex human activity recognition from 3-axis IMU and gyroscope data (Zmitri, M. 2019) Similar to the previous approach, the data needs to be preprocessed and normalized by sampling at fixed frequency, then padded into a fix length vector to perform normalization of intensities.
\[v_{i}^{t} = \frac{v_{i}}{\sum_{t=0}^{T} v_{i}^{(t)}}\]The team then collect 200 to 500 different postures (each lasts less than 5 seconds) and format the processed padded data into a table. Performing Principal Component Analysis (PCA) could produce a transitional matrix $\Sigma$ to map IMU and gyroscope time-series data onto a low-rank feature space. The team first calculate covariance matrix on the training data and then solve the equation of transitional matrix $\Sigma$ to map data vectors onto low rank space based on the large eigenvalues of covariance matrix.
\[Cov = \frac{1}{n-1} X^{T}X\] \[\Sigma v = \lambda v\]The team store the pre-obtained transitional matrix $\Sigma$ as a part of the preprocessing step on MCU. For the recored postures, a good choice would be to perform the transformation to map them as a vector on low-rank metric space. The team ask the user to record each posture at least 10 times upon assigning it to an instruction. During posture recognition, the team perform K-Nearest-Neighbor algorithm to classify the new posture p with the recorded postures based on their L2 distance in the metric space, given by
\[d(x, x') = \sqrt{\sum_{i} (x_{i} - x'_{i})^2}\]The team would also try methods such as Finite State Machine (FSM) or Markov Models if the team have enough time.
The team has developed and refined the smart wand’s design, testing it across multiple smart home scenarios. The wand successfully completed all intended tasks, demonstrating reliable performance in controlling home devices and interacting with Bluetooth-enabled appliances. Throughout the project, the team conducted rigorous testing and improvements, ensuring that the wand met both functional and usability standards.
Add your slides to the Final Project Proposal slide deck in the Google Drive.
I. -C. Severin and D. -M. Dobrea, “6DOF Inertial IMU Head Gesture Detection: Performance Analysis Using Fourier Transform and Jerk-Based Feature Extraction,” 2020 IEEE Microwave Theory and Techniques in Wireless Communications (MTTW), Riga, Latvia, 2020, pp. 118-123, doi: 10.1109/MTTW51045.2020.9245072.
Zmitri, Makia, Hassen Fourati, and Nicolas Vuillerme. “Human activities and postures recognition: From inertial measurements to quaternion-based approaches.” Sensors 19.19 (2019): 4058.
Bonnet, V., Mazzà, C., McCamley, J. et al. Use of weighted Fourier linear combiner filters to estimate lower trunk 3D orientation from gyroscope sensors data. J NeuroEngineering Rehabil 10, 29 (2013). https://doi.org/10.1186/1743-0003-10-29
Datasheet for bluetooth module
Datasheet for bluetooth module
You can remove this section if you don’t need these references.