Optimize visual navigation and obstacle avoidance for mobile robots
Already deployed in tens of thousands of service robots produced by World’s Top 500 companies.
MYNT EYE S Color is designed to optimize visual navigation and obstacle avoidance for mobile robots. With multiple core parameters upgraded, it can provide reliable 3D perceptive information about the environment and more precise and stable image sources to support the development of positioning and navigation solutions for mobile robots.
Upgraded color image sensor
Upgraded color image sensor with a depth resolution of up to 2560x800. It also provides ISP image processing with richer and more stable image sources for colored object recognition, navigation and obstacle avoidance both indoors and outdoors.
Improved Depth Resolution
MYNT EYE S Color not only has the color image sensor upgraded but also has the output depth resolution increased to provide more accurate depth image information to support the research and development of applications in areas like 3D image reconstruction, visual navigation, and obstacle avoidance.
Better Imaging Quality
MYNT EYE S Color has a built-in ISP (Image Signal Processor) image processing chip that can restore scene details with higher fidelity under different indoor and outdoor lighting conditions, providing better image sources for the researches in visual processing field.
- Auto Exposure Control
- Auto White Balance
- Linear Correction
- Noise Removal
Overall Improvement in IMU Performance
Upgraded BOSCH BMI088 six-axis IMU sensor provides high anti-jamming and stability performance. The IMU synchronization accuracy of up to 0.01 ms provides more precise six-axis complementary data for visual SLAM researches in a complex environment.
IMU Timestamp and Image Synchronization in Microseconds
The IMU synchronization accuracy of up to 0.01 ms provides more stable image synchronization and six-axis information for the technological development in Visual Inertial Odometry (VIO).
Comparison of IMU Synchronization Accuracy
Built-in high-precision I2C time synchronization chip
Built-in high-precision I2C time synchronization chip and time synchronization interface provide hardware-level time synchronization for scenarios that integrate multiple sensors such as laser, ultrasound and millimeter waves, and the synchronization accuracy can be precise up to the millisecond level.
Connects to multiple sensors like lasers,
ultrasound and millimeter waves.
provide hardware-level time synchronization,
and the synchronization accuracy can be precise
up to the millisecond level
I2C time synchronization chip
It provides time synchronization up to the millisecond level for the researches on and applications of the integration of multiple sensors in complex scenarios, such as positioning and navigation, 3D perception and 3D reconstruction.
It is equipped with 146° ultra-wide-angle lens, 8 cm baseline, and provides image sources with distortion corrections and synchronized binoculars, effectively reducing recognition blind spots and providing more precise depth image data.
0.26-3m+ working distance
binocular frame synchronization
Image sources with distortion corrections and synchronized binoculars
Better integration performance
Rapid Integration and Performance Implementation
It provides a detachable external casing for better integration performance in terms of structural adaptation to robots.
Recommended Application Scenarios
MYNT EYE S Color is designed to optimize visual navigation and obstacle avoidance for mobile robots. We recommend that you use it in the following application scenarios. At the same time, you can also refer to the product specifications and parameter table to explore more application possibilities.
It can provide more precise and safer navigation, positioning and obstacle avoidance for unmanned deliveries, improve the performance of SLAM algorithms, and build an accurate semantic map.
It empowers the robot with stereoscopic navigation and recognition capabilities, allowing it to move, walk and avoid obstacles autonomously in complex spaces both indoors and outdoors.
Rapid acquisition of 3D environment information and establishment of 3D models of a specified space allow it to be widely used in commercial, medical, industrial and many other fields.
It detects lane obstacles and pedestrians and performs other assistant driving functions by acquiring images of the road ahead under different weather and lighting scenarios.