we used sentence BERT technique for efficient
classification of YouTube comments.
2 LITERATURE SURVEY
Text classification, named entity recognition (NER),
and sentiment analysis benefit from the ability to
assess contextual meanings in language using
bidirectional context. In contrast, the autoregressive
structure of the GPT model enables it to generate
more coherent content and achieve effective results in
tasks such as chatbots and creative writing. However,
applying these models to agglutinative and
morphologically complex languages, like Turkish,
presents significant challenges. (M. Salıcı, 2024)
An Aspect-based sentiment Analysis method
based on transformer-based deep learning models. It
uses the Hugging Face Transformers library, which
gives access to cutting-edge pre-trained models. For
sentiment categorization, the approach used the
BERT model, a potent transformer architecture. This
study adds to the expanding field of sentiment
analysis by offering a scalable and reliable
transformer-based learning method for ABSA. This
approach is a useful tool for applications in product
reviews, and customer feedback analysis. (G. P. M S,
2024)
RoBERT sentiment analysis generates negative,
positive and neutral sentiment in the analysed text.
These scores are described independently, it is very
difficult to complete understanding about the
opinions in the comments. To propose a more reliable
analysis of mixed emotions found in unstructured
data, a composite sentiment summarizer combines
positive, negative, and neutral scores into one. This
composite score gives a more accuracy and reliable
representation of the sentiment elaborated in
comments. (Z. -Y. Lai, 2023)
For multimodal tasks, the Adapted Multimodal
BERT (AMB) is a BERT-based model that integrates
adapter modules with intermediate fusion layers.
These fusion layers perform task-specific, layer-by-
layer integration of audio-visual data and textual
BERT representations. Meanwhile, the adapter
adjusts the pre trained language model to suit the
specific task at hand. This approach enables fast and
parameter-efficient training by keeping the
parameters of the pre trained language model frozen
during the adaptation process. Research shows that
this method yields effective models that are resilient
to input noise and can outperform their fine-tuned
counterparts. (O. S. Chlapanis, 2023)
The architecture comprising a temporal
convolutional network (TCN), a convolutional layer,
a bidirectional long short-term memory (BiLSTM),
and robustly optimized bidirectional encoder
representations from transformers pre-training
approach (RoBERTa) is proposed to address issues.
Dual branch feature coding network based on
RoBERTa (DBN-Ro) is the name of the suggested
architecture. The stitched vectors undergo
dimensionality compression via the convolutional
layer. (F. Wang, 2021)
To assess the performance of lexicon-based and
sentence-BERT sentiment analysis models used as
code-mixed, low-resource texts as input, it
summarizes the results of experiments. Some code-
mixed texts in Javanese and Bahasa Indonesia are
utilized as a sample of low-resource code-mixed
languages in this study. Google Machine Translation
is used first to translate the raw dataset into English.
The input text is translated into English and then
classified using a pre-trained Sentence-BERT model.
The dataset used in this study is divided into positive
and negative categories. The experimentation found
that the combined Google machine translator and
Sentence-BERT model achieved 83 % average
accuracy, 90 % average precision, 76 % average
recall, and 83 % average F1 Score. (C. Tho, 2021)
Classifying sentiment is a crucial step in figuring
out how individuals feel about a good, service, or
subject. Sentiment classification and numerous
models for natural language processing have been put
forth. But most of them have focused on categorizing
sentiment into two or three groups. The model tackles
the fine-grained sentiment categorization task using a
promising deep-learning model named BERT.
Without a complex design, experiments demonstrate
that the model performs better than other well-known
models for this task. (M. Munikar, 2019)
Sentiment classification used as an Indonesian
dataset, was explored with the problem utilizing a
two-step procedure: sentiment classification and
aspect detection. The bag-of-words vector, which is
handled by a fully connected layer, and the word
embedding vector, which is handled by a gated
recurrent unit (GRU), are two deep neural network
models with different input vectors and topologies for
aspect detection that are compared. They also contrast
two deep neural network methods for sentiment
classification. Word embedding, sentiment lexicon,
and POS tags are input vectors in the first method,
which has a bi-GRU-based architecture. In the
second, the word embedding vector is rescaled using
an aspect matrix as the input vector, and the topology