2018 Ce VK
2018 Ce VK
2018 Ce VK
net/publication/327253343
CITATIONS READS
211 21,785
4 authors, including:
Thomas H Bradley
Colorado State University
139 PUBLICATIONS 4,553 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Vipin Kumar Kukkala on 14 March 2019.
By Vipin Kumar Kukkala, Jordan Tunnell, Sudeep Pasricha, and Thomas Bradley
A
dvanced driver-assistance systems (ADASs) vehicles, passive safety systems that have been under contin-
have become a salient feature for safety in modern uous development for many decades have been augmented by
vehicles. They are also a key underlying technolo- active safety systems, which seek to prevent a crash from
gy in emerging autonomous
vehicles. State-of-the-art
ADASs are primarily vision based,
but light detection and ranging (lidar),
radio detection and ranging (radar),
and other advanced-sensing technolo-
gies are also becoming popular. In this
article, we present a survey of differ-
ent hardware and software ADAS
technologies and their capabilities and
limitations. We discuss approaches
used for vision-based recognition and
sensor fusion in ADAS solutions. We
also highlight challenges for the next
generation of ADASs.
OVERVIEW OF AUTOMOTIVE
SYSTEM SAFETY
Safety in automotive systems has been
a major concern since the early days
of on-road vehicles. Several original
equipment manufacturers (OEMs)
have attempted to address this issue
by developing various safety systems
to protect occupants within a vehicle
as well as prevent injuries to people
outside the vehicle. These systems are
mainly classified into two types: 1)
passive (or reactive) and 2) active (or
proactive). Passive safety systems pro-
tect vehicle occupants from injuries
after a crash, e.g., seat belts, air bags,
and padded dashboards. Due to a con-
sistent consumer demand for safer
MONOCULAR CAMERAS
These camera systems have only one lens. As these systems
have only one image output at any point of time, they have
low image-processing requirements compared to those of
other camera types. These cameras can be used for multiple
applications, such as the detection of obstacles, pedestrians,
lanes, and traffic signs [5]. They can also be used for moni-
toring the driver inside a vehicle, e.g., for face- and eye-
detection and head-pose analysis [23]. But monocular camera
images lack depth information and are, therefore, not reliable
sensors for distance estimation. Some techniques [5] allow
approximating distance by identifying key features in the
captured image frame and tracking their position when the
©iStockphoto.com/metamorworks
camera is in motion.
STEREO CAMERAS
These systems consist of two or more lenses, each with image
sensors, separated by a certain distance (known as stereo
base). Stereo cameras are useful in extracting three-dimensional
Blind-Spot
Traffic Sign Detection
Recognition
Park Assist
Park Assist
Adaptive Automatic Braking
Surround
Cruise Pedestrian Detection
View
Control Collision Avoidance
Lane-
Keeping
Blind-Spot
System
Detection
Surround View
ADAS
Short-/
Monocular Stereo Infrared Long-Range
Medium-Range PMD IMUs GPSs
Camera Camera Camera Radar
Radar
FIGURE 2. The taxonomy of an ADAS. PMD: photonic mixer device; IMUs: inertial measurement units; GPSs: global positioning systems.
(3-D) information from two or more two-dimensional images capture the reflected light. Passive IR cameras use an IR sen-
by matching stereo pairs (images from left and right sensors) sor, where every pixel on the IR sensor can be considered as a
and using a disparity map to estimate the relative depth of a temperature sensor that can capture the thermal radiation emit-
scene. These cameras can be used for a variety of applications, ted by any material. Unlike active IR cameras, passive IR cam-
such as traffic sign recognition, lane, pedestrian, and obstacle eras do not require any special illumination of the scene. Still,
detection as well as distance estimation, with much greater popular night-vision solutions mainly use active IR cameras to
accuracy compared to monocular cameras. assist the driver by displaying video data on a screen during
Stereo systems can be relied upon for accurate distance low light conditions.
(depth) estimation over short distances, up to 30 m. In most
production vehicles with stereo cameras, the cameras are locat- LIDAR
ed inside the vehicle, behind the rear-view mirror, angled slight- Lidar works by firing a laser beam at an object and then mea-
ly downward, and facing the road. suring the time taken for the light to bounce back to the sen-
sor, to calculate the distance of an object. These systems can
IR CAMERAS achieve high-resolution 3-D images and operate at longer
There are two main types of IR cameras. Active IR cameras ranges than camera systems. Some of the lidar scanners sup-
use a near-IR light source (with wavelengths from 750 to 1,400 nm) port surround-view sensors (that fire laser beams continuously
built in the vehicle to illuminate the scene (which cannot be in all directions), which can generate a 360° 3-D image of the
seen by the human eye) and a standard digital camera sensor to surroundings with extremely accurate depth information (as
ULTRASONIC SENSORS
Ultrasonic sensors use sound waves to measure the distance
to an object. These sensors are mainly used for detecting
objects very close to the vehicle. Some example applications
include automatic parking and parallel parking assist. These
sensors are mainly located under the front and rear bumper of
the vehicle.
OTHERS
A few other sensors are used to complement and improve the
functionalities of those discussed earlier. For instance, pho-
tonic mixer device (PMD) cameras consist of an array of FIGURE 3. The Google self-driving-car-generated 3-D image of its
smart sensors that enable fast optical sensing and demodula- surroundings using lidar [9].
tion of incoherent light signals simultaneously [12]. PMDs
can support parallel target pixel-wise distance measurement
without scanning, thus resulting in faster imaging, high later- Image Preprocessing Segmentation
al resolution, and depth information. IMUs and GPSs are Acquisition
examples of systems that help improve the distance measure-
ments with lidar and radar.
Object Detection Depth
System Control
VISION-BASED ADASs and Tracking Estimation
Vision-based ADASs rely on images from cameras and use
computer vision principles to extract useful information. FIGURE 4. The vision data flow for the ADAS used.
VEHICLE DETECTION
Confidence = 0.907121 String = STOP Vehicle detection is a major focus of
object detection in ADASs. The fact
that many vehicles share common fea-
tures, such as having tires, brake lights,
and license plates, allows the detection
Confidence = 0.889147 String = STOP of these objects to indicate the pres-
ence of a car. These features are all
used to distinguish the vehicle from
other objects, such as signs, roads, and
other miscellaneous objects. In Figure 6,
(a) (b) an example of vehicle detection is shown,
using a CNN framework (Darknet)
FIGURE 5. The stop sign detection: (a) a binary stop sign and (b) stop sign classification and a real-time detection system, You
using optical character recognition. Only Look Once [19]. The orientation
NEXT-GENERATION ADASs
Next-generation ADAS solutions are beginning to use sensor
fusion and other advanced communication systems, such as
vehicle-to-everything (V2X). FIGURE 6. The object (lane, vehicle, and sign) detection.
Zero No automation
GEOSPATIAL CONSTRAINTS
Many of the modern ADAS solutions being developed are
One Driver assistance Human driver tested within a geographic location or a group of locations
Two Partial automation where they are sold. This limits the ADAS to one or a certain
group of geographical locations. This is because not all coun-
Three Conditional automation
tries (or some states in a country) adhere to the same sign and
Four High automation System road conventions uniformly, which makes ADAS algorithms
that are often trained under one location hard to work
Five Full automation
efficiently in other locations. There is a need to improve