CASE STUDY
Optimizing the autonomous perception of vehicles through video annotation

+30%
Precision in the detection of pedestrians and mobile objects
÷ 1.5
ADAS algorithm calibration time
+8 hours
of annotated data ready for training per day
In the automotive industry, the development of autonomous vehicles relies on the ability of embedded systems to understand their environment in real time. This requires very high quality video datasets, annotated with extreme precision.
The mission
Objective: create a training dataset for the detection and classification of road objects (vehicles, pedestrians, signs, traffic lights) from video streams. To achieve this objective, Innovatiana has deployed a dedicated approach including:
- Frame-by-frame annotation in bounding boxes and polygons to capture the dynamics of road scenes;
- Systematic quality control to ensure temporal and spatial consistency in annotations.
The results
- Extensively annotated video sequences to feed the perception algorithms;
- A significant improvement in obstacle detection and autonomous navigation in various conditions;
- An acceleration of the testing and validation phases of ADAS and autonomous systems.
👉 Read our article on ADAS annotation : Learn how accurate video annotation enhances the intelligence of autonomous vehicles.