Revolutionizing Autonomous Driving Software For Safer Object Detection

Welcome to the project website! This project focuses on improving object detection and localization using radar and image data in complex environments. Here, you can explore the various aspects of our end-to-end pipeline, including coordinate transformation, YOLO object detection, DBSCAN clustering, and merging radar clusters with image labels.

Overview

Coordinate Transformation

Converting radar data from its native coordinate system to a coordinate system aligned with camera data.

YOLO Object Detection

A real-time object detection algorithm that simultaneously predicts object classes and bounding boxes.

DBSCAN Clustering

A density-based clustering algorithm used for unsupervised machine learning, grouping data points based on distance.

Merging Radar Clusters & Image Labels

Combining radar point data clusters with object labels generated by the YOLO algorithm for the image data.

For more details about each step, please see the sections below.

Motivation & Benifits of Automatic Radar Labeling

Pipeline & Results

1. Coordinate Transformation

Coordinate transformation involves converting radar data from its native coordinate system (radar space) to a coordinate system that is aligned with the camera data (camera space). This process allows for the integration of information from multiple sensors to improve the accuracy of object detection and localization in complex environments.

Coordinate Transformation Result

Radar in native 3D coordinate

Coordinate Transformation Result

Projection to 3D vehicle coordinate

Coordinate Transformation Result

Projection onto 2D left image

Coordinate Transformation Result

Projection onto 2D middle image

Coordinate Transformation Result

Projection onto 2D right image

2. YOLOv3 Object Detection

YOLO (You Only Look Once) is a real-time object detection algorithm. It leverages a single neural network to simultaneously predict object classes and bounding boxes for those objects in input images. The algorithm is trained using a large dataset of labeled images and backpropagation to update the weights of the neural network.

YOLO Detection Result 1

Left Camera YOLO Results

YOLO Detection Result 2

Middle Camera YOLO Results

YOLO Detection Result 3

Right Camera YOLO Results

3. DBSCAN Clustering

DBSCAN (Density-Based Spatial Clustering of Applications with Noise) is a clustering algorithm used for unsupervised machine learning. It groups together data points that are close to each other in feature space and separates data points that are far away. The algorithm identifies clusters based on their distance from each other and recursively adds nearby points that meet the epsilon and minimum points criteria until no more points can be added to the cluster.

4. Merging Radar Clusters and Image Labels

The final step of the pipeline is to combine the results generated by the clustering and labeling algorithm. Radar point data clusters obtained from the DBSCAN algorithm are merged with object labels generated by the YOLO algorithm for the image data. This combination is accomplished by utilizing the overlapping bounding boxes between the radar points and image labels as a means of matching the results.

Before Merge

After Merge

YOLO Detection Result 1

Left Camera Clustering Results

YOLO Detection Result 1

Left Camera Labeling Results

YOLO Detection Result 2

Middle Camera Clustering Results

YOLO Detection Result 2

Middle Camera Labeling Results

YOLO Detection Result 3

Right Camera Clustering Results

YOLO Detection Result 3

Right Camera Labeling Results

5. Final Birds Eye View with Labeled Data

In the end, once we have radar data labeled in 3D, we are able to show the labeled point cloud relative to the vehicle's position.

Below is an example of the visualization.

YOLO Detection Result 2

Point Cloud Labeled Bird Eye View

6. Software & Demo Video

Limitations

While our end-to-end pipeline has shown promising results in object detection and localization, it is important to acknowledge some limitations and potential areas for improvement:

In future iterations of the project, we plan to address these limitations by refining our algorithms, incorporating more advanced sensor technologies, and optimizing the pipeline for better computational efficiency. The preliminary results demonstrate the potential of our approach in improving object detection and localization, and we believe further development can lead to more accurate and robust systems for autonomous driving applications.