recognize both static and dynamic obstacles
simultaneously, although the data acquired may
sometimes experience relatively higher latency. The
second technology is LiDAR, first developed in the
1960s, which estimates the distance between the
reflected objects and the sensor by emitting infrared
or laser pulses. Its advantages include the ability to
conduct detection and estimation in three dimensions,
as well as providing intensity information about the
reflected objects. Radar operates on principles similar
to those of LiDAR, emitting electromagnetic waves
toward a target area and determining the target’s
relative speed and position by receiving the reflected
waves. It’s easy to see that combining the functions
of these three types of sensors creates a solid
framework for detecting obstacles and gathering
related information. By integrating sensor data with
advanced software and computing systems,
autonomous driving technology can reduce reliance
on human input while enhancing overall efficiency
and safety. As a prerequisite and foundation for all
subsequent calculations and processes, data
collection, along with robust and reliable
environmental evaluation and detection, warrants
significant attention. After all, one of the highest
potential risk factors to others on the road is the
presence of other vehicles. The ability to accurately
and promptly perceive and identify the surrounding
environment is critical to ensuring the safety of
passengers, pedestrians, and the vehicles themselves.
Only by predicting potential danger can the system
respond accordingly based on given instructions.
Equipping an autonomous vehicle with a
comprehensive data-capturing system can greatly
reduce safety risks, earn greater public trust, and
ultimately expand its marketing potential.
3.2 Target Detection
After discussing the data collection techniques in
autonomous vehicles, another critical aspect is object
detection. Vision-based object detection can
generally be categorized into three types: traditional
techniques, machine learning, and deep learning. This
paper focuses on the roles of machine learning and
deep learning in autonomous driving. As a highly
popular topic in artificial intelligence and computer
science in recent years, machine learning primarily
operates by using data and algorithms to split data
into training and testing sets, thereby emulating
human learning processes and gradually reaching
minimal error through continual learning and
training. When applied to autonomous vehicles,
machine learning typically involves two key steps:
first, inputting and processing images to obtain a
region of interest (ROI), and then transforming and
encoding the extracted images, converting high-
dimensional image-space data into lower-
dimensional data, all while continuously refining this
process through training and optimization.
First, during the feature extraction step, various
methods, such as Histogram of Oriented Gradients
(HOG), Haar-like descriptors, Local Binary Patterns
(LBP), Gabor filters, and Speeded-Up Robust
Features (SURF), are commonly used to capture
recognizable vehicle traits. These approaches help
identify consistent patterns that remain reliable when
the vehicle’s orientation or model changes. Some
techniques, like the Deformable Part Model (DPM),
further refine HOG features to deal with more
complex shapes. In the classification stage,
algorithms like AdaBoost, K-Nearest Neighbors
(KNN), Naive Bayes, Support Vector Machines
(SVM), and Decision Trees are often employed. Each
classifier must balance how well it fits its training
data with how effectively it adapts to new inputs.
Ensemble learning, which merges multiple
classifiers, can enhance overall detection
performance. A key challenge is the high
computational cost of searching the entire image for
potential vehicles. To address this, many systems
focus on areas of interest, such as identifying shadows
or other cues, before running feature extraction and
classification. This targeted approach can
significantly reduce processing time while
maintaining reliable detection.
Another object detection algorithm is deep
learning. Because machine learning's reliance on
feature extractors and classifiers can limit a model's
capabilities in certain scenarios, the rise of
convolutional neural networks brought significant
attention to deep learning algorithms, making them
more suitable for object detection methods in the field
of computer vision. Typically, deep learning-based
object detection methods are divided into object
detection-based approaches and segmentation-based
approaches. First, the object detection-based
approach can be categorized into three types: anchor-
based detectors, anchor-free detectors, and end-to-
end detectors. The key difference between anchor-
based and anchor-free detectors is that anchor-based
detectors predefine bounding boxes to detect objects
by partitioning and categorizing vehicles in proposed
regions, then predicting the vehicle's center and
bounding box. Anchor-free detectors directly predict
the object's center point and then cluster them into a
single entity to obtain the bounding box. Some
models derived from YOLO play a crucial role in