For autonomous vehicles to navigate safely and efficiently in complex and ever-changing road environments, they are akin to an experienced driver, requiring constant and precise perception of their surroundings. This includes the speed, direction, and distance of other vehicles; the position and movement of pedestrians; the indications of traffic signs and lights; and a wealth of information such as road gradients, curvature, and the presence of obstacles. Sensors are the core components that endow autonomous vehicles with this "super-perceptive ability," with different types of sensors acting like the car's various "sensors," each playing a unique role.
Cameras, much like the "eyes" of autonomous vehicles, capture rich visual information about the vehicle's surroundings. Common automotive cameras can be categorized into several types, including front-view, rear-view, and surround-view cameras. Front-view cameras help cars "see" the road ahead, clearly identifying traffic signs and lane markings, and keenly detecting vehicles and pedestrians in front. Rear-view cameras excel in reversing scenarios, providing the driver with a rear view to assist in safe reversing. Surround-view cameras, through the collaborative work of multiple miniature cameras distributed around the vehicle, create a 360° bird's-eye view, giving the driver a clear understanding of the surrounding environment. Taking the ON Semiconductor AR0820AT, an advanced automotive CMOS digital image sensor, as an example, it features a 3848H x 2168V active pixel array. It not only accurately captures images within a linear or high dynamic range but is also optimized for performance in low-light and harsh high dynamic range scenarios. With its 2.1µm DR-Pix BSI pixels and on-chip 140dB HDR capture capability, it makes a significant contribution to obtaining clear images of the vehicle's surroundings. However, cameras also have their limitations. Their performance is greatly reduced in low-visibility environments such as nighttime, dim lighting, rain, and fog, just like how people can hardly see distant objects in thick fog.
Millimeter-wave radar, which can be seen as the "hearing" of autonomous vehicles, senses its surroundings by emitting and receiving millimeter waves. It excels in detecting the speed and distance of objects and possesses strong anti-interference capabilities, remaining unaffected by adverse weather and lighting conditions, and operating stably in rain, snow, and fog. In ADAS (Advanced Driver Assistance Systems) functions, millimeter-wave radar plays a crucial role. Compared to cameras, it can "see" objects more accurately, offering superior resolution, performance, and directional accuracy. For example, TI's TIDA-020047 dual-device millimeter-wave cascaded radar reference design perfectly meets the needs of automotive 4D imaging radar. By cleverly combining two 76GHz to 81GHz radar transceivers, a radar processor, two CAN-FD PHYs, an Ethernet PHY, and a low-noise power supply, it successfully overcomes the "seeing" challenge in ADAS functions. However, millimeter-wave radar is not without its limitations; it faces certain difficulties in identifying non-metallic objects.
LiDAR, acting as an extension of the "tactile sense" of autonomous vehicles, uses light pulses to measure the distance between the vehicle and surrounding objects, thereby constructing a high-precision 3D environmental map. It boasts significant advantages such as extremely high distance, angle, and velocity resolution, and strong anti-interference capabilities. It can acquire a wealth of data and information crucial for autonomous driving, including the distance, angle, velocity, and reflection intensity of objects, thus generating multi-dimensional images of those objects. Time-of-flight (ToF) LiDAR excels in short-range automotive use cases, acquiring rich details without scanning. In the automotive environment, high-resolution ToF cameras use 3D sensing technology to scan the area around the vehicle and the ground, accurately detecting curbs, walls, or other obstacles regardless of lighting conditions, providing strong support for functions such as self-parking. Infineon's IRS2877A, a leader in the REAL3 ToF LiDAR series for automotive applications, utilizes a 9 x 9 mm² plastic BGA package. With a tiny 4 mm photosensitive area, it achieves a VGA system resolution of 640 x 480 pixels. A single ToF camera can be used to create a powerful 3D facial recognition driver monitoring system. However, LiDAR is relatively expensive, and its performance can be affected by extreme weather conditions such as heavy rain and snow.
Since each sensor has its unique advantages and limitations, how can we fully leverage their strengths to achieve a synergistic effect where 1 + 1 > 2? This is where sensor fusion technology comes in. Simply put, sensor fusion cleverly combines data sources from multiple sensors to create more accurate and reliable information than when a single sensor works alone. In the field of autonomous driving, sensor fusion technology mainly employs three implementation methods: data-layer fusion, feature-layer fusion, and decision-layer fusion.
Data layer fusion is the process of fusing raw data from various sensors at the lowest level. By employing advanced data processing technologies, it integrates raw data collected by sensors such as cameras, radar, and lidar to obtain the most comprehensive and accurate information about the environment. This fusion method processes a massive amount of data but provides extremely high accuracy, and is often used in scenarios such as real-time environmental modeling, for example, to build a precise 3D model of the vehicle's surroundings by fusing camera and lidar data.
Feature layer fusion first extracts features from data from various sensors, such as visual features like shape and color from camera data, and features like speed and distance from radar data. These features are then organically integrated and analyzed. Feature layer fusion is commonly used in complex autonomous driving systems, effectively reducing redundant data and significantly improving data processing efficiency. For example, combining object shape features from cameras with speed features from millimeter-wave radar can generate more accurate object contours, helping to more accurately identify and track target objects.
Decision-level fusion integrates the recognition results obtained from independent processing by different sensors. Each sensor makes a judgment based on its own collected data, and the system then synthesizes these judgments to make a final decision. This fusion method is highly abstract and intelligent, effectively addressing conflicts between sensor data and conducting comprehensive risk assessments based on multiple pieces of information, significantly improving the system's overall response capability. Decision-level fusion plays a crucial role in high-risk scenarios such as emergency braking and obstacle avoidance. For example, when a camera detects an obstacle ahead, and millimeter-wave radar confirms the obstacle's presence and distance, the system integrates both to quickly issue an emergency braking or avoidance command.
Through sensor fusion technology, autonomous vehicles can organically combine the rich visual information from cameras, the precise ranging and speed measurement capabilities of millimeter-wave radar, and the high-precision 3D modeling advantages of lidar to achieve comprehensive and high-precision perception of the surrounding environment. This not only greatly improves the safety and reliability of autonomous driving systems but also provides a solid data foundation for vehicles to make accurate decisions in complex traffic scenarios. On urban roads, facing frequent pedestrians, vehicles, and complex traffic signs and signals, the fused sensor system can quickly and accurately identify and analyze various information, helping vehicles drive safely and smoothly. On highways, sensor fusion technology helps vehicles monitor the speed and distance of vehicles ahead in real time, enabling advanced driver assistance functions such as adaptive cruise control, effectively improving driving comfort and safety.
While sensor fusion technology offers numerous significant advantages for autonomous driving, it still faces a series of serious challenges in practical applications. First, different sensors have varying data acquisition frequencies, timestamps, and data formats, necessitating precise data synchronization and efficient data fusion processing to ensure the final fused data accurately reflects the real-time state of the vehicle's surroundings. Second, the increasing number of sensors and the explosive growth in data volume place extremely high demands on computing power, requiring robust onboard computing capabilities to support real-time and efficient data processing. Third, the cost of sensor fusion systems is a significant issue; reducing the cost of sensors and fusion systems while maintaining high performance is a critical challenge that urgently needs to be addressed to promote the large-scale commercial application of autonomous driving technology. Furthermore, cross-platform standardization of sensor fusion technology is still incomplete; the lack of unified standards between sensors and fusion algorithms from different manufacturers creates numerous obstacles to system interoperability and compatibility.
Looking ahead, with the continuous advancement of technology, sensor fusion technology is expected to achieve significant breakthroughs. On the one hand, sensor performance will continue to improve while costs will gradually decrease. For example, LiDAR is expected to maintain high precision while significantly reducing costs, enabling its wider application in various types of autonomous vehicles. On the other hand, cutting-edge technologies such as artificial intelligence and big data will be deeply integrated with sensor fusion technology, further improving the intelligence level and processing efficiency of fusion algorithms. Furthermore, with the widespread adoption of high-speed communication technologies such as 5G, vehicle-to-vehicle (V2V) and vehicle-to-infrastructure (V2I) communication will become smoother, providing sensor fusion technology with richer data sources, expanding the perception range of vehicles, and enhancing the ability of autonomous driving systems to cope with complex traffic environments.
Sensor fusion technology, as a core technology in the field of autonomous driving, provides strong support for autonomous vehicles to "see" their surroundings. Despite current challenges, with continuous innovation and development, it will undoubtedly shine even brighter in the future of intelligent transportation, leading autonomous driving technology to new heights and bringing people a safer, more convenient, and more efficient travel experience.