1) The document proposes a method for moving object detection and tracking using multiple sensors for improved perception. It performs fusion at the detection level by combining data from proximity sensors and cameras.
2) At the detection level, object position is obtained from proximity sensors while visual features provide preliminary classification information. The outputs are then fused to generate a combined object representation with position, shape, and a probability distribution over object classes.
3) The method was tested on a vehicle environment for detecting objects like pedestrians, bikes, cars and trucks in real-time. Fusing data from multiple sensors at the detection level improved the perceived model of the environment for applications like advanced driver assistance systems.