Jayesh Nair

Sensor data fusion: Camera & Radar

This was a team project done under supervision of Technische Hochschule University, Ingolstadt. The goal was to perform late-fusion on radar & camerea detections.

This was really fun to work on - got to learn about nitty-gritties of model training, evaluation and spatial fusion of detections. I would like to highlight a few sections here from the project.

Github Repo

Image Augmentation:

Primary problem when it comes to working on road user data is high levels of data imbalance among different classes, we were working with 6 road user classes with the following distribution.

Road user data distribution
Road user class distribution showing data imbalance

Tried a bunch of different techniques for image augmentation, one of the most interesting was Image Inpainting.

Image Inpainting augmentation technique
Image augmentation results

Model Fine-Tuning:

Used Ultralytics library along with YOLOv8-large for fine tuning on the INFRA-3DRC dataset

Sensor Fusion:

Performed spatial fusion of detection objects from two different sensors: Radar & Camera. As seen below, the box is detection result from the camera (YOLO v8) and the red dot represents the centroid of cluster (detection object from radar).

Sensor fusion technique
Sensor fusion results