Inspiration
Robotics and artificial intelligence are two of the most powerful technologies today, but building systems that combine them is often extremely difficult. Developers frequently spend more time integrating sensors, hardware, and software pipelines than actually building intelligent behaviors.
In Canada, this challenge is even more significant because of our large geographic size, remote communities, and growing need for automation in industries like healthcare, logistics, agriculture, and accessibility technologies. Many of these applications require intelligent machines that can operate autonomously or assist people where human resources are limited.
However, developing these systems typically requires deep expertise in machine learning, robotics, and embedded systems, which creates a barrier for students, researchers, and small teams who want to build solutions for real-world problems.
We were inspired by the idea that if we could dramatically lower the barrier to building AI-powered robotics, more people could develop solutions for challenges facing Canada — from assistive technologies to remote automation.
This led us to build AutoniMake, a platform that allows users to train computer vision models and connect them to hardware actions without needing to write complex machine learning code.
We were inspired by the question:
What if building an autonomous robot was as easy as training an image classifier?
Our goal was to reduce the barrier to entry for robotics and AI by creating a platform where users could teach machines new behaviors without writing complex machine learning code.
What it does
AutoniMake is a code-free AI robotics platform that allows users to train computer vision models and instantly connect those models to hardware actions.
Using our interface, a user can:
Capture training examples from a camera
Train a computer vision model in seconds
Map predictions to hardware actions
Control real-world devices like robots, displays, and sensors
For example, a user could train the system to recognize hand gestures:
👍 → move robot forward
✋ → stop robot
✌️ → turn robot left
Once trained, the system performs real-time inference and sends commands to modular hardware peripherals.
This allows anyone to prototype autonomous systems without building the entire robotics pipeline from scratch.
How we built it
AutoniMake combines computer vision, machine learning, and modular hardware control into a single framework.
Computer Vision Pipeline
We built a vision pipeline using OpenCV for image preprocessing and data capture. Camera frames are processed and passed to a lightweight convolutional neural network (CNN) for classification.
A CNN works by learning spatial patterns within images through convolutional filters. Each layer extracts increasingly complex features:
Feature Extraction→Pattern Recognition→Classification
This allows the system to recognize gestures, objects, or signals from camera input.
AI Training System
Users can collect datasets directly from the camera feed. These images are used to train a custom classification model that predicts labels and outputs confidence scores.
Example prediction output:
Gesture: thumbs_up Confidence: 94%
The system performs real-time inference, allowing visual input to instantly trigger hardware responses.
Hardware Architecture
AutoniMake uses a hub-and-peripheral architecture.
A Raspberry Pi acts as the central hub
The hub runs the AI pipeline and web interface
Commands are sent to hardware modules via serial communication
Peripheral devices are powered by ESP32 microcontrollers, which control hardware components like:
Displays
Motors
Sensors
Commands follow a simple protocol:
DISPLAY:Hello ROBOT:F ROBOT:STOP
This modular design allows new hardware components to be added easily without changing the AI system.
Interface
We built a simple web-based interface that allows users to:
capture training images
train AI models
map predictions to hardware actions
monitor predictions in real time
This makes the platform accessible even for users with no machine learning experience.
Challenges we ran into
One of the biggest challenges was integrating multiple systems in real time.
Our project involved:
computer vision
machine learning
web interfaces
microcontroller communication
robotics hardware
Ensuring all these components worked together with low latency required careful system design.
Another challenge was building a reliable communication pipeline between the Raspberry Pi hub and ESP32 peripherals. We had to design a simple command protocol and ensure commands were parsed correctly across devices.
Finally, training models quickly enough for a live demo required optimizing our dataset size and inference pipeline.
Accomplishments that we're proud of
Building a complete end-to-end AI robotics system in a short time. We successfully integrated computer vision, machine learning, web software, and embedded hardware into a single platform that allows AI predictions to directly control real-world devices.
Creating a code-free workflow for training AI models. Instead of requiring users to write machine learning code, AutoniMake allows users to capture training data, train a vision model, and deploy it to hardware through a simple interface.
Developing a modular hardware architecture. Our Raspberry Pi hub communicates with ESP32 peripherals using a custom command protocol, allowing multiple hardware modules like displays or robots to be connected and controlled by AI.
Achieving real-time AI inference controlling physical devices. Our system processes camera input, classifies it using a CNN-based model, and sends commands to hardware modules with minimal latency.
Designing an accessible platform for robotics experimentation. By simplifying both the AI and hardware setup, we created a system that lowers the barrier for students, makers, and researchers to build autonomous systems.
Successfully demonstrating live training and deployment. One of our biggest accomplishments was being able to train a model and immediately see its predictions trigger hardware responses in real time.
What we learned
Through this project we learned how to integrate AI systems with real-world hardware, which requires far more than just training models.
We gained experience in:
real-time computer vision pipelines
CNN-based image classification
embedded systems communication
modular robotics architecture
designing user-friendly interfaces for complex technologies
Perhaps most importantly, we learned how powerful AI becomes when it can directly interact with the physical world.
What's next for AutoniMake
Our vision for AutoniMake is to expand it into a full AI robotics development platform.
Future improvements could include:
support for more sensors and hardware modules
more advanced AI models
a drag-and-drop behavior builder
cloud-based model training
expanded robotics applications
Ultimately, we want to make building intelligent machines accessible to anyone, not just robotics experts.

Log in or sign up for Devpost to join the conversation.