UGV Rover ROS 2 Open-source 6 Wheels 4WD AI Robot, Dual controllers, All-metal Body, Flexible and Expandable, Suitable for Raspberry Pi 4B / Raspberry Pi 5 UGV Rover PI4B ROS2 Kit Raspberry Pi 4 NOT Included
UGV Rover ROS 2 Open-source 6 Wheels 4WD AI Robot, Dual controllers, Computer Vision, Comes With Interactive Tutorials, All-metal Body, Flexible and Expandable, Suitable for Raspberry Pi 4B / Raspberry Pi 5
Features at a glance
The UGV Rover ROS2 Kit is an AI robot designed for exploration and creation with excellent expansion potential, based on ROS 2 and equipped with Lidar and depth camera, seamlessly connecting your imagination with reality. Suitable for tech enthusiasts, makers, or beginners in programming, it is your ideal choice for exploring the world of intelligent technology.
Equipped with the high-performance Raspberry Pi computer to meet the challenges of complex strategies and functions, and inspire your creativity. Adopts dual-controller design, combines the high-level AI functions of the host controller with the high-frequency basic operations of the sub controller, making every operation accurate and smooth.
Easy to be controlled remotely via UGV Rover Web Application without downloading any software, just open your browser and start your journey. You can use the basic ROS 2 functions of the robot without installing a virtual machine on the PC. Supports high-frame rate real-time video transmission and multiple AI Computer Vision functions, the UGV Rover is an ideal platform to realize your ideas and creativity!
Based on Raspberry Pi
Supports Raspberry Pi 5 / Raspberry Pi 4B, with powerful computing performance to handle more complex tasks, offering more possibilities
Connecting with Raspberry Pi 4B
Connecting with Raspberry Pi 5
Dual-Controller design, provides efficient collaboration and upgraded performance
The host controller adopts Raspberry Pi for AI vision and strategy planning, and the sub controller uses ESP32 for motion control and sensor data processing
Raspberry Pi OS + ROS2 Docker
Ensures advanced decision-making performance of robot and system compatibility at the same time. Supports all AI functions of the previous AI Kit series products
360° Flexible Omnidirectional Pan-Tilt
equipped with 5MP 160° wide-angle camera for capturing every detail
The Pan-Tilt adopts high-torque bus servos with excellent expansion potential, providing a better control experience as FPS games
open source for all ROS 2 Development Resources
open source for all demos of host controller and sub controller, including robot description file (URDF model), sensor data processing node of sub controller, kinematic control algorithms, and various remote control nodes
Integrates various ROS 2 mapping methods
Meet the needs of mapping in different scenarios
Gmapping 2D mapping
Cartographer 2D mapping
RTAB-Map 3D mapping
multiple cost-effective sensors
Adopts multiple sensors with high cost-effectiveness and practicality
Auto exploration and mapping
Using SLAM Toolbox to implement mapping and navigation functions simultaneously in unknown environments, simplifying the task execution process. The UGV robot can autonomously explore unknown areas and complete the mapping, suitable for unmanned applications
Supports natural language interaction
Adopts Large language model (LLM) technology, users can give commands to the robot by natural language, enabling it to perform tasks such as moving, mapping, and navigation
Provides Web console tool
You can use the basic ROS 2 functions on the Web without installing a virtual machine on the PC, supports cross-platform operation on Android or iOS tablets. Users can simply open a browser and control the robot for moving, mapping, navigation, and other operations
ROS2 node command interaction
Users can send control commands to the robot by a script for performing operations such as moving, obtaining the current location, and navigating to a specific point, etc. which is more convenient for secondary development
Gazebo simulation debugging
Provides Gazebo robot mode and complete functionality library for simulation debugging, helping you verify and test the system during the early stages of development
Continuing the adventure as night falls
High-brightness LED light for Ensuring Clear Images In Low-Light Conditions
Suitable for tactical extension
Comes with 21mm wide rail and 30KG.CM high precision & high-torque bus servo for tactical extension
for reference only, the accessories in the picture above are NOT included
Standard Aluminum rail
Comes with 2 × 1020 European standard profile rails, and supports installing additional peripherals via the boat nuts to meet different needs, easily expanding the special operation scenarios
Note: Only the rail, boat nuts, and M4 screws are included, other accessories should be purchased separately.
Supports driving in complex terrain
6 wheels × 4WD design, using 6 wheels can provide a more stable platform and larger contact area, while 4WD can provide stronger power and traction to deal with various terrains and obstacles
Easy To Control Via
Cross-Platform Web Application
No App installation required, allows users to connect and control the robot via mobile phones, tablets and computers via browser Web App. Supports shortcut key control such as WASD and the mouse via a PC with keyboard
WebRTC real-time video transmission
Adopts Flask lightweight web application, based on WebRTC ultra-low latency real-time transmission, using Python language and easy to extend, working seamlessly with OpenCV
Recognition, Tracking, and targeting
based on OpenCV to achieve color recognition and automatic Targeting. supports One-key Pan-Tilt control and automatic LED lighting, allows expansion for more functions
Face Detection:
Automatic picture or video capturing
based on OpenCV to achieve Face recognition, supports automatic photo taking or video recording once a face is recognized
Intelligent Object Recognition
supports recognizing for many common objects with the default model
Gesture Recognition:
AI Interaction with body language
Combines OpenCV and MediaPipe to realize gesture control of Pan-Tilt and LED
Gesture control for photo taking
LED ON/OFF and blacklight control
More MediaPipe Demos for Easily Creating
Complex Video Processing Tasks
MediaPipe is an open-source framework developed by Google for building cross-platform multimedia processing pipelines, provides a set of pre-built components and tools, its high-performance processing capability enables the robot to respond to and process complex multimedia inputs such as real-time video analytics.
Face Recognition
Attitude Detection
40PIN GPIO Extended Header
The robot only occupies the URAT interface of the Raspberry Pi GPIO for communication, adapting outer side 40PIN header of the driver board for expanding more peripherals and functions
Obtains real-time information feedback
Real-time monitoring the operating status of the robot
Web page command line tool
Multiple Functions for Easier Expansion
Quick to set up, easy to expand
Easily customize and add new functions without Modifying front-end code
ESP-NOW Wireless communication
between robots
Based on ESP-NOW communication protocol, multiple robots can communicate with each other without IP or MAC address, achieving multi-device collaboration with 100-microsecond low-latency communication
Gamepad Control for better operation Experience
Comes with a wireless gamepad, making robot control more flexible. You can connect the USB receiver to your PC and control the robot remotely via the Internet. Provides open source demo for customizing your own interaction method
Supports 4G/5G Expansion
Supports installing 4G/5G module* for the application scenarios without WiFi
* You may need to use Tunneling Service such as Ngrok, Cpolar, or LocalTunnel to expose the local network service of the robot (Flask application) to the Internet so that you can control the robot from anywhere.
realizes remote control across the internet
- Our web application demos are based on WebRTC for real-time video transmission.
- WebRTC (Web Real-Time Communications) is a technology that enables web applications and sites to establish peer-to-peer connection and capture optionally stream audio and/or video media, as well as to exchange arbitrary data between browsers without requiring an intermediary.
- We provide comprehensive Ngrok tutorials* to help you get started quickly and realize robot control across the internet.
* Provides the usage tutorials of Ngrok only, we do not provide any Ngrok accounts or Servers. You can follow our tutorial to open your own Ngrok service, or choose other tunneling services according to your needs.
Supports installing smartphone holder
If you have a spare phone, you can install it on the robot via holder as below, using the phone to create a hotspot for the robot and achieving remote control across the internet at a lower cost
* Comes with a smartphone holder with 1/4″ screw in the package
Cross-platform interactive tutorial
Develop while you learn
supports accessing Jupyter Lab via devices such as mobile phones and tablets to read the tutorials and edit the code on the web page, making development easier
Open-source all demos
Full dual-controller technology stack
Product Show
Outline dimensions
Resources & Services