3.2 Emotion Detection Based on
Multimodal Data
In the study of road rage detection mechanisms, the
method of integrating facial expressions with driving
behavior has demonstrated significant advantages.
This method uses in-vehicle devices to capture the
driver's facial expressions in real time, combining
them with driving behavior data such as steering
wheel angle change rate, acceleration, and braking
pedal operation to construct a Fisher linear
discriminant model for comprehensive judgment
(Huang et al., 2022). Experiments have shown that
compared to using facial expressions or driving
behavior features alone, the fusion method
significantly improves recognition accuracy,
providing an effective means for online monitoring of
road rage. Key technological research on intelligent
vehicle occupant monitoring systems (OMS) has also
provided new insights into road rage detection. The
passenger-side child monitoring and warning scheme
achieved through technologies such as OpenPose
demonstrates the potential of multimodal information
fusion in enhancing vehicle safety monitoring (Fu,
2022).
Emotion perception technology plays a critical
role in driving safety within intelligent cockpits,
especially in high-risk emotional scenarios such as
road rage. A driver monitoring system (DMS) based
on a 940 nm wavelength infrared light source
combined with facial micro-expression analysis can
real-time capture physiological features of anger in
drivers (such as tense brow muscles and drooping
corners of the mouth) (Shaobi, 2021). When road rage
symptoms are detected, the system can collaborate
with the cockpit environment control module for
proactive intervention: for example, dynamically
adjusting the lighting to a cooler tone via
environmental light sensors to reduce visual
stimulation; simultaneously activating the fragrance
system to release calming scents to alleviate
emotional escalation (Shaobi, 2021). This real-time
response mechanism effectively suppresses
aggressive driving behavior and reduces the risk of
traffic accidents caused by emotional outbursts.
However, precise perception of road rage still faces
significant challenges. While existing algorithms
perform well in laboratory settings, they are easily
influenced by individual differences in real-world
driving scenarios (e.g., variations in facial
expressions due to cultural backgrounds or dialectal
intonation changes) (Guo et al., 2023). Additionally,
intervention strategies for road rage are currently
limited to environmental regulation functions and
have not yet achieved deep synergy with vehicle
control layers (e.g., adjusting adaptive cruise control
following distance or steering sensitivity). Future
research should establish a dedicated road rage
dataset covering multiple driving scenarios and
design a layered response logic: primary
interventions use environmental adjustments to
alleviate emotions, and if emotions continue to
deteriorate, the level of autonomous driving
intervention is automatically increased, with
cognitive guidance provided by a voice assistant
when necessary (Guo et al., 2023). This tiered
response system can achieve a closed-loop safety
protection mechanism from emotion recognition to
behavior correction.
Existing multimodal emotion recognition systems
have achieved a certain level of accuracy, collecting
data from multiple sensors to identify and analyses
the driver's state and provide emotion regulation
measures (Zhang & Chang, 2025). Building on this,
the relationship between emotional changes and
driving state is a critical step that requires
experimental validation. Monitoring driving
behaviour styles under different emotional states,
testing and analysing factual data, and constructing a
reasonable logical structure will lay the foundation
for subsequent integrated mechanisms.
Research indicates that a driver's emotional state
(primarily negative emotions) can spontaneously
influence driving behaviour without cognitive control
(Eherenfreund Hager et al., 2017). Additionally,
drivers cannot clearly articulate the specific triggers
for these emotions (Hu et al., 2013). Based on this,
previous studies have shown that experimental tests
using driving simulators with emotionally charged
vocabulary can induce emotional responses in drivers
(Steinhauser et al., 2018). Some tests have used video
presentations to induce fluctuations in drivers'
emotions (Gu, 2021), while others have employed
emotionally charged music and evocative
recollections to induce emotional responses (Kwallek
et al., 1988).
In summary, it can be seen that drivers' emotions
can be influenced by external factors, and different
methods can cause varying degrees of emotional
changes. Therefore, based on the identified emotional
states, we need to design an integrated mechanism for
the overall cabin environment to achieve positive
intervention and interaction with drivers, thereby
regulating users' psychological states, enhancing
driving safety and experience, and achieving