examines EEG signals under various noise levels
and semantic conditions, employing time,
frequency, and time-frequency analyses. For this
classification, we used ensemble classifiers such as
AdaBoost (ADA) (Hoseini, S.S 20204). XGBoost
(XGB) (Wang, F., et al., 2022) and Random Forest
(RF) (Edla, D.R., et al., 2018). XGB performs better
than the other two classifiers, particularly in
frequency domain analysis.
2 RELATED WORKS
This section discusses a variety of Auditory attention
detection task-related works. A novel neural-inspired
architecture for EEG-based auditory attention
detection (AAD) beat both linear and CNN models
on the KUL and DTU databases, with average
accuracies of 91.2% and 61.5% within 5-second
windows, respectively. It demonstrated continuous
improvements throughout 1 to 5-second decision
windows while being much more computationally
efficient (less than 1% of SSF-CNN's cost), making
it appropriate for neuro-steered hearing aids.
However, its accuracy lags behind the cutting-edge
SSF-CNN model (Cai, S., et al., 2022) STAnet, a
model for auditory spatial attention detection
(ASAD), uses spatial and temporal EEG data to
achieve excellent accuracy on the KUL and DTU
databases (92.6% and 76.1% within 5-second
windows, respectively). It works well with as few as
16 EEG channels and incorporates spatial-temporal
aspects for greater accuracy. However, it may
encounter overfitting or computational complexity
issues (Su, E., et al., 2022) EEG-Graph Net, an EEG-
graph convolutional network with a neural attention
mechanism, decodes auditory attention by modeling
EEG channels as nodes and their interactions as
edges, so representing the brain's spatial patterns as a
graph. It obtained outstanding 1-second window
accuracies of 96.1% and 78.7% on the KUL and
DTU databases, respectively (Cai, S., T. Schultz, and
H. Li 2023). A neuronal attention mechanism has
been proposed to dynamically assign weights to EEG
sub-bands and channels, collecting discriminative
representations for auditory attention detection
(AAD). When combined with an AAD system, it
achieved 1-second and 2- second average accuracies
of 79.3% and 82.96% on the KUL and DTU
databases, respectively. While effective, its precision
falls short of previous works (Cai, S., et al., 2021). A
unique Auditory Attention Decoding (AAD)
mechanism was presented, which combines CNN and
ConvLSTM to extract spectro-spatial-temporal
information from 3D descriptors using
topographical activity maps of multi-frequency
bands. Experiments on KUL, DTU, and PKU
databases with 0.1s, 1s, 2s, and 5s decision windows
outperformed SSF-CNN and cutting-edge models.
Even without auditory stimuli, the model improved
AAD accuracy, with different trends observed across
databases and decision windows. Multi-band
frequency analysis and ConvLSTM-based temporal
analysis made major contributions to the accuracy
gains (Jiang et al., 2022).
As previously stated, the majority of prior
classification work has used the KUL and DTU
datasets;
however, no
work
has
used
the
PhyAAt dataset for this form of categorization.
Nonetheless, some research has focused on attention
score (Bajaj et al., 2022) detection tasks on the
PhyAAt dataset, as described below. Study (Ahuja,
C. and D. Setia 2022) enhances selective auditory
attention research by combining 14-channel EEG
data from the PhyAAt dataset with speech-to-text
annotations. Using EEGNet and traditional machine
learning, it reduced test MAE from 29.65 to 22.47
for a single individual, resulting in an overall
MAE of 31.47. Incorporating speech-to-text data
improves attention monitoring, but the model is
prone to overfitting and lacks evaluation at various
noise and semantic levels. Study (Kim, D.-Y., et al
2022) used Multivariate Multiscale Entropy
(MMSE) to examine entropy variations in EEG data
in relation to auditory input and attention levels. The
MMSE, a measure of information in random signals,
demonstrated that entropy rises during sentence
practice, non-semantic sentence presentation, and
when noise is combined with input sentences. These
findings present a quantitative technique to assessing
cognitive models. However, the emphasis on specific
conditions (rehearsal, non-semantic words, noise)
limits its applicability to other situations, and
depending entirely on MMSE and entropy-based
analysis may neglect other relevant EEG signal
variables.
3 PROPOSED METHODOLOGY
To evaluate and detect task classification in the
PhyAAt dataset, we used a variety of domain
analyses along with ensemble classifier's. Our
proposed mechanism involves three stages of work:
preprocessing, feature extraction, and classification.
Figure 2. represent the general block diagram of
proposed work.