SlideShare a Scribd company logo
IAES International Journal of Artificial Intelligence (IJ-AI)
Vol. 13, No. 2, June 2024, pp. 1668~1678
ISSN: 2252-8938, DOI: 10.11591/ijai.v13.i2.pp1668-1678  1668
Journal homepage: http://ijai.iaescore.com
Aspect based sentiment analysis using a novel ensemble deep
network
Abraham Rajan, Manohar Manur
Department of Computer Science and Engineering, Christ University, Bangalore, India
Article Info ABSTRACT
Article history:
Received Apr 5, 2023
Revised Nov 11, 2023
Accepted Dec 1, 2023
Aspect-based sentiment analysis (ABSA) is a fine-grained task in natural
language processing, which aims to predict the sentiment polarity of several
parts of a sentence or document. The essential aspect of sentiment polarity and
global context have deep relationships that have not received enough
attention. This research work design and develops a novel ensemble deep
network (EDN) which comprises the various network and integrated to
enhance the model performance. In the proposed work the words of the input
sentence are converted into word vectors using the optimised bidirectional
encoder representations from transformers (BERT) model and an optimised
BERT-graph neural networks (GNN) model with convolutions is built that
analyses the ABSA of the input sentence. The optimised GNN model with
convolutions for context-based word representations is developed for the
word-vector embedding. We propose a novel EDN for an ABSA model for
optimised BERT over GNN with convolutions. The proposed ensemble deep
network proposed system (EDN-PS) is evaluated with various existing
techniques and results are plotted in terms of metrics for accuracy and F1-
score, concluding that the proposed EDN-PS ensures better performance in
comparison with the existing model.
Keywords:
Aspect-based sentiment
analysis
Ensemble deep network
Graph neural network
Sentiment analysis
This is an open access article under the CC BY-SA license.
Corresponding Author:
Abraham Rajan
Department of Computer Science and Engineering, Christ University
Bangalore, India
Email: abraham_ranjan@rediffmail.com
1. INTRODUCTION
Through blogs, discussion forums, social networking sites, and e-commerce websites, users have
expressed their opinions on a wide range of topics, including products, individuals, and organizations, during
the past several years. This swapping of ideas is made possible by the fast development of web applications
and the broad availability of inexpensive internet connections, both of which generate massive amounts of
online data. Sentiment analysis uses this large quantity of data that is used to analyses major challenges. Before
the widespread use of the internet as a resource for decision-making, due to the widespread use of the internet
and the world wide web, it is now possible to read the opinions and perspectives of anonymous individuals
from a range of cultures and places. Many users' shopping opinions are impacted by internet reviews by various
other shoppers [1].
Systems that analyses user sentiment will automatically summarize user reviews, which may help a
customer, make an opinion. Major advantages of sentiment analysis systems include scalability, the ability to
summarize massive volumes of text, real-time analysis, and the capability to deliver fast assessments, and
consistent criteria, which is the result of automation and lacks bias compared to user judgment. Organizations,
whether public or private, require a sentiment analysis system; it is feasible to supplement standard
recommendation systems using sentiment analysis. It assists the owners of the organization in understanding
Int J Artif Intell ISSN: 2252-8938 
Aspect based sentiment analysis using a novel ensemble deep network (Abraham Rajan)
1669
the sentiments of customers towards their items, however, this is suitable for market research and competition
analysis, politics, public policy development, and legal research are further uses [2]. Although sentiment
analysis may be undertaken at various levels, aspect-level sentiment analysis is the most useful. Researchers
have conducted an extensive study on the identification of explicit sentiments that result in the generation of
many alternatives.
Due to its complexity, embedded characteristic recognition receives less attention; this is
simultaneously a substantial portion of a sentence, which has various parameters, to detect sentiment analysis
in understanding the concept. Sentiment analysis research focuses on making predictions at the phrase or
document level by assessing the overall sentiment towards the entire phrase or content [2]–[4]. The opinion is
based on the perhaps inaccurate assumption that the available data determines a single opinion towards a
specific issue. In consequence, the need for aspect-based sentiment analysis (ABSA) is a method for
differentiating more complex aspect-level thoughts and feelings, which has surged over the past decade [5],
[6]. The target achieved for opinion mining on ABSA shifts its focus from being a phrase or document on a
whole to being an entity or a particular characteristic of an entity. For example, an entity may be an
online-sold product, with its features including its price and size. Refer to the following example review to
understand the words used in aspect-level sentiment analysis: "the display quality of the Samsung TV is
outstanding." The term outstanding in this context denotes a favorable opinion of the Samsung TV's display,
which is the element of the Samsung TV that is being discussed. As illustrated in Figure 1, the three methods
that may be employed to perform aspect-level sentiment analysis are aspect identification, sentiment evaluation
concerning the specific aspect, and sentiment aggregation.
Figure 1. Aspect level sentiment analysis
Context-based methods and syntax-based methods are the two main areas of focus in the existing
research for ABSA based on deep neural networks [5], [6]. To gather aspect word representations in a phrase,
context-based approaches often employ convolutional neural networks (CNN) or recurrent neural networks in
conjunction with attention processes. Graph convolutional networks (GCN) are frequently used to represent
phrases in syntax-based models. Recent research on ABSA has relied widely on deep neural models due to
their capability to capture the semantic features of specific aspects [1]–[4]. In the existing research on ABSA
that uses deep neural networks, context-based techniques [5], [6] and syntax-based methods [7]–[9] are
distinguished into two groups. Context-based algorithms frequently combine CNN or recurrent neural networks
with attention mechanisms to collect aspect word representations in a sentence, in syntax-based models, GCN
are typically employed to represent sentences.
Finding several aspect-level sentiment components, including sentiment polarities, opinion terms,
aspect words, and aspect categories, is the focus of this research work. Present ABSA models that have been
developed to achieve better performance in a variety of tasks, which focuses on training and testing data, which
is developed from the same distribution (e.g., the same domain or the same language). When the data
distribution changes, it is usually important to retrain the ABSA model to maintain its performance. Huge
quantities of labelled data are often too expensive or even impossible to collect for ABSA applications that
require aspect-level annotations. Incorporating various other methods for building ABSA-based models, which
generalize across various domains and languages, is cross-domain transfer or cross-lingual transfer, which
requires adapting the trained model to unknown domains or unknown languages.
− Convert the words of the input sentence into word-vectors using the optimized bidirectional encoder
representations from transformers (BERT) model.
− To extract the context-based word representation using the optimized graph neural network (GNN) model
with convolutions over the word-vector embedding.
− To develop a BERT-GNN model that analyses the ABSA of the input sentence.
 ISSN: 2252-8938
Int J Artif Intell, Vol. 13, No. 2, June 2024: 1668-1678
1670
− To develop a proposed novel ensemble deep network (EDN) for an ABSA model for optimized BERT over
GNN with convolution.
This particular research is organized as follows: first section of the research starts with background
work and significance of sentiment analysis in natural language processing (NLP); section concludes with
research motivation and contribution. Second section of the research discuss various existing deep learning
protocol along with its shortcoming; third section presents mathematical modelling of EDN. Thereafter, EDN
is evaluated in fourth section.
2. RELATED WORK
ABSA has been one of the key area of research in sentiment analysis, thus there has been plenty of
research in past decade. Traditional approach lacks the deep understanding of the common sense among the
sentences. This section discuss several deep learning mechanism for sentiment analysis. Huang and Carley [10]
have developed a novel, reliable CNN for aspect-level sentiment analysis, and integrating this aspect data into
a CNN with the aspect-level features results in developing a high-accuracy model. Long short-term memory
(LSTM) networks are difficult to analyses and have low temporal efficiency, whereas CNN can only allocate
local semantic data. The sparse attention-based convolutional neural network (SA-SDCCN) proposed by [11]
deals with semantic and sentiment embedding that is added to an embedding tensor at the multichannel
embedding layer, which leads to accurate representations of the input sequence.
According to Wang et al. [12], a novel model known as unified position-aware convolutional neural
network (UP-CNN), which initially proposed an aspect detection network with prior knowledge to
accommodate the absence of aspect positions in ACSA, and has proposed an aspect mask to construct
aspect-aware contextual representations to accommodate CNN in ABSA. Consequently, several NLP
applications have successfully integrated GCNs, by using grammatical links in aspect-level sentiment analysis
tasks, Zhang et al. [7] have developed the first aspect-specific graph convolutional network (ASGCN). Initially,
the context information is collected using a network with bidirectional long short-term memory (Bi-LSTM),
then aspect-based word information is collected using the masked approach. The ABSA approach developed
by [13] is based on the convolution of dependency trees.
Zhang and Qian [14] introduced a bidirectional GCN model consisting of a hierarchical syntactic
network and a lexical graph for each phrase based on syntactic links and word co-occurrence data. Based on a
dependency network, Tang et al. [15] suggested a modified double-transformer structure for aspect sentiment
categorization that focuses on essential sentence components. Using relational graph-based attention networks,
Wang et al. [16] have modified the dependent trees. The previously published GCN-based models perform
well on the ABSA; however, they lacked sentiment data across context-related factors. In addition to this the
interactive sentiment correlation between aspects recovered by GCNs on the dependency tree with external
sentiment-based knowledge and integrate with MHSA to generate semantic information between context and
aspects.
According to Zhang et al. [17], transformer-based semantic-primary knowledge transferring (TSPKT)
network for aspect-term sentiment analysis across domains makes use of semantic-primary information to
facilitate the transfer of knowledge across multiple domains. Before semantic-primary information is collected
from the S-Graph, it is generated using external semantic lexicons. The second postulate is that Aoa graphormer
learns syntactically meaningful aspect-term phrases. Thirdly, the standard by LSTM classifier is developed by
adding a knowledge-aware memory unit (KAMU) that holds the whole corpus of semantic primary input.
According to Hu et al. [18], this approach focuses on the above-mentioned two issues by combining aspect and
task-level regularizations. Aspect-level regularizations limit the attention-based weights to reduce noise. Sparse
regularization is more suited for sentences with a single aspect than orthogonal regularization for sentences
with various aspects. Here they have proposed a task-level regularization by adding aspect-level category
identification as an orthogonal auxiliary task to generate sentiment-dominant features.
3. PROPOSED METHODOLOGY
The proposed novel EDN for ABSA model for optimized BERT over GNN with convolutions.
Figure 2 shows the proposed architecture. The pooling layer here is responsible to extract features of similar
size by selecting the maximum number from each feature vector. Further, this connected layer fine-tunes the
model considering the sentiments taken from the previous layer. Finally determining the sentiments from the
softmax layer.
− Transfer the words of the input sentence into words using the optimized BERT model.
− To extract the context-based word representation using the optimized GNN model over the word-vector
embedding.
Int J Artif Intell ISSN: 2252-8938 
Aspect based sentiment analysis using a novel ensemble deep network (Abraham Rajan)
1671
− Build a BERT-GNN model with convolutions that analyses the ABSA of the input sentence.
Consider a given set of sentiments D = {d1 , d2, … . . , dn} for l ∈ L, let
Sen = {sen1 , sen2, … . . , senn} be a set of sentences for k sentiments. For sen ∈ Sen be the set of words
sen = {l1 , l2, … . . , ln} for one sentence denoted assen. In the proposed architecture for a novel EDN for
aspect-based sentiment, analysis the input is given in the form of sentiments, which is then converted into a
sentence–word-vector. This is further fed into the optimized BERT model where it is converted into word
embedding by the optimized BERT model. The optimized LSTM model here extract the context-word
representations, the context-based words are fed into the GNN with convolution, which converts the node
matrices into the Attention mechanism wherein the further bias is reduced while, processing the GNN model
with convolutions.
Figure 2. Proposed architecture
3.1. Preliminaries and problem definition
A fixed set of c sentiments D = {d1 , d2, … . . , dn} in context with the target entity, for a target
sentiment dk. Let fxy be the y − th aspect for the given entity, gxy is the essential feature relevant to the aspect
fxy and Sxy determines the sentiment of the aspect fxy. The main aim here is to develop an optimized GNN
model for aspect-level sentiment analysis. The main objective here is subjected to the mapping function as
mentioned in (1). The proposed model incorporates the following steps, to transform the words of the input
provided into word vectors along with the optimized-BERT model.
G: (gxy) = {
1, if Sxy is positive
0, if Sxy is neutral
−1, if Sxy is negative
(1)
3.1.1. Word-vector generation
The word vector model automatically inserts a cls at the start of the sentence as well as two separate
symbols in the initial phase and ending phase of the term to determine the sentiment. The word sequence here
is determined by converting a specific sentence into a separate form v = [cls] + v + [diff] + fx + [diff]. This
sequence of words is then fed into this model to obtain the vector for each word Oi as shown in (2). Here the
Opt − BERT model (Oi) is the word vector extracted via the pre-trained model, as well as dimo is the dimension
for the word vector. The input sequence is determined from the input sequence through c, the sequence obtained
through the G = {g1, g2, … , gk}.
 ISSN: 2252-8938
Int J Artif Intell, Vol. 13, No. 2, June 2024: 1668-1678
1672
gx = Opt − BERT(Oi) ∈ Hdimo∗1
(2)
3.1.2. Word–sentence–vector converter
The word representations are depicted as each word converted into an aggregated form through the
entire vector that is condensed into the entire input vector sequence. This study deploys an optimized LSTM
model over the word-vector embedding for word-vector context basis word representation. Optimised LSTM
by learning the context information through understanding the meaning of words by reading the input given in
two different ways. Context-based representations are made according to the following steps.
− Step 1: first step is dealing with input layer; the word-vector sequence is represented as G = {g1, g2, … , gk}
given as input to the optimised LSTM model.
− Step 2: second step is dealing with optimized LSTM this layer is responsible to co-ordinate the context-
based information by the remaining words in the direction of the sentence in two different ways for the
specific-target word. The optimized LSTM model incorporates a dual-LSTM model, left-to-right model
(→ Opt − LSTM ) model, and right-to-left model (← Opt − LSTM), which is responsible to encode the
sentence from left-to-right as (→ jx(x = [1, k])) and right-to-left (← jx(x = [1, k])) as shown in (3).
→ jx = (→ Opt − LSTM(Lg→ j gx + g→ j → j → j x−1
+ e→ j ) ∈ Npj (3)
← jx = (→ Opt − LSTM(Lg← j gx + g← j ← j → j x−1
+ e← j ) ∈ Npj (4)
jx = (L→ j j → jx + L→ j j ← jx + ej ) (5)
Here L is the weight matrix, by considering the Lg→ j amidst the input given and the forward hidden vectors, jx
for representing the context. jx+1 is the hidden vector represented for jx and jk+1 = 0. jx−1 for the previously
hidden vector for jx and j0 = 0.
− Step 3: the context-based word-vector matrix denoted as G = {g1, g2, … , gk}∈ Nk∗pj, here pj depicts the
size of the vector in the matrix G.
3.2. Building a graph neural network with convolutions
− Step 1: build a sentence GNN
A sentence-based GNN with convolution denoted as SGNN = ( N, V, AM) including a set of nodes in
correspondence with the k words in a given sentence, a set E of edges denotes the dependencies of the adjacent
node pairs for syntactic dependency as well as an adjacency matrix AM∈Z|N|∗|N|
denoting the node relations
denoted as shown in (6):
Fxy = {
1, if Nx, Ny ∈ N, and vxy ∈ V
1, if Nx = Ny,
0, else
(6)
Additionally, this graph SGNN has a feature matrix embedded with nodes denoted as Y = [G] ∈ N|N|∗pj, where
each row is denoted as Gx which is responsible for the context-based representation of word node Nx ∈ V.
− Step 2: generation of node-embedding
Representing the nodes: the matrices are generated as P ∈ N|N|∗pj and F ∈ N|N|∗pj are given as input into a
GNN network with convolution shown to generate the nodes represented as (7):
H = G(t)
= ϑ(PG(b−1)
L(b)
+ q(b)
(7)
Here b depicts the number of layers in the GNN network with convolutions layer, P ∈ N|N|∗ph; G(0)
= Y, here
ϑ is a non-linear activation function as ReLU. L(b)
∈ N|N |∗ph for the transformation layer in the b − th layer.
q(b)
is the bias associated with GNN layers associated with convolution, respectively:
P = K−1/2
TAM−1/2
(8)
P is the symmetric matrix of AM where T is the degree of the matrix AM whereas;
Kxx = ∑ Fxy
y (9)
Int J Artif Intell ISSN: 2252-8938 
Aspect based sentiment analysis using a novel ensemble deep network (Abraham Rajan)
1673
− Step 3: GNN with convolution transformation
In this step, the noise is reduced and the bias is associated with processing the GNN network. A position
attention mechanism, which captures the necessary parts in the sentence irrespective of the sentiment as shown
in (10):
Ix = ∑ ϑx
|N|
x=1 Hx (10)
Here,
ϑx =
exponential(CFtanh (LF[Hx⊕wx]+qF
∑
|N|
h=1
exponential(CFtanh (LF[Hh⊕wh]+qF
(11)
Here LF
∈ N(
pj + pv) ∗ (pj + pv), qF
∈ N(
pj + pv), and CF
∈ N pj + pv for learning matrix. ⊕ is the
concatenation operator. wx ∈ Npv is the weight associated with the x − th word formulated as in (12):
wx = {
|x−initiate|
k
, x ≤ initiate
0, initiate ≤ x ≤ terminate
|x−terminate|
k
, x > terminate
(12)
Here initiate and terminate determines the starting and ending phase of the sentiment in the sentence.
3.3. Constructing the classifier
The dimension of the node embedding is reduced by the convolution layer by generating the feature
vectors ux, which determines the sliding of the filter R ∈ N r∗ph with length r from x to x + r − 1 to extract
the necessary information shown as (13):
ux = ReLU(RꭈMx:x+|N|−1 + q) (13)
Here x = [1, |N|] determines the order of the node representation M; ꭈ is the convolution operator, the
activation function is denoted by ReLU and q is the bias term. The feature vectors are constructed by node
representations as shown in (14):
u = [u1,u2, … … , u|N|] (14)
− Step 1: max-pooling layer
The pooling layer builds the feature vectors for the same size by picking the maximum number of each vector
uk, the main aim here is to determine the size of feature vectors uk, which depends on the size of each matrix
M and R. The size will differ if the length of the sentence and the filter sizes are different, new features such as
u, are defined as in (15):
u = [u1,u2, … … , u|N|] (15)
− Step 2: fine-tuning of the layers
This layer is responsible for fine-tuning the characteristics of the previous layer, which necessarily determines
the sentiments as shown in (16):
Y = SM(HM
. u + q) (16)
Here HM
∈ Lb∗|N|
and q ∈ Hb
is the weight matrix and bias of the layer. b is the number of sentiment phases.
Training the optimized GNN model with convolution to reduce the entropy loss values.
3.4. Optimizing loss of ensemble deep network
To train the CNN model over optimized LSTM model integrated with GNN network with
convolutions to reduce the loss function denoted as shown in (17)
Loss = − ∑ yk logyk^ + δ||γ||2
q
k (17)
 ISSN: 2252-8938
Int J Artif Intell, Vol. 13, No. 2, June 2024: 1668-1678
1674
Here q denotes the sentiment classes for the training model, yk denotes the distribution of true sentiment for
the k − th class and yk^ denotes the distribution of the sentiment to be predicted for the k − th class, δ denotes
the regulation co-efficient, and γ is the parameter set from the previous layer.
4. PERFORMANCE EVALUATION
The task of solving ABSA is more complex than conventional text-level sentiment analysis. It focuses
on recognizing the characteristics or features of an item described in a text and the perspective has taken on
each characteristic. It has been proven that applying optimized BERT to ABSA as a sentence-pair classification
task and developing an auxiliary sentence yields better results than the existing state-of-art models and baseline
models. Further, we have evaluated our proposed ensemble deep network proposed system (EDN-PS) model
with existing state-of-art methods and baseline models in terms of accuracy and F1-score.
4.1. Dataset details
To evaluate the efficiency of the proposed model to test the performance of the association for
computational linguistics (ACL) twitter social dataset [19]. The sentiment polarity is determined by dividing
the dataset into three categories: positive, negative and neutral. Table 1 shows the dataset details. The details
of the various methodologies used for comparison with the proposed EDN-PS model.
Table 1. Dataset details
Dataset Positive Negative Neutral Total
Training Testing Training Testing Training Testing Training Testing
Twitter 1,561 173 1,560 173 3,127 346 3,127 346
4.2. Compared models
To comprehensively evaluate our proposed model, we compared it to certain baseline models and
state-of-the-art (SOTA) models from different times. The eight baseline models are evaluated are mentioned
here. The seventeen models listed below, including the existing ones, have always performed at SOTA level.
4.2.1. Baseline models:
− IAN [4]: constructs attentive representations of the target context and engages in interactive attentive
learning of the target and context, respectively.
− MemNet [20]: determines the importance of various context words for a single feature and utilizes this
information as the semantic representation of the sentence to distinguish the most essential context
information of a particular trait.
− ATSM-S [21]: classifies the sentiment of extremely complex and coarsely written language by specifically
modelling the desired qualities at three distinct granularity levels, followed by categorizing the sentiment.
− AOA [22]: aspects and sentences are represented in a decentralized manner, with explicit capture of the
relationship between context phrases and aspects. The AOA module simultaneously teaches the model
aspects and sentence representation, with an emphasis on the main parts of the phrase.
− ASGCN-DG [7]: possesses a structure of undirected dependency. The adjacency matrix is produced using
DSPT words to avoid the limitations of attention processes and CNN-based models and depends on an
ordered dependency tree. Two networks are distinguished by their adjacency matrices.
− MAN [23]: a multi-attention network (MAN) was developed to alleviate the problem of information loss
when a coarse-level attention mechanism is employed to simulate aspects.
− UP-CNN-BERT [12]: this model proposes a novel unified location-aware CNN that constructs an
aspect-aware contextual representation for ABSA using an aspect mask to match the CNN, to compensate
for the absence of aspect locations in ACSA.
4.2.2. State-of-art models at different times
− ATAE-LSTM [3]: by using LSTM in conjunction with the attention mechanism to do sentence-level
semantic modelling to address the issue of aspect-level sentiment analysis. This is accomplished by
evaluating the significance of diverse contextual information to a certain attribute.
− GANN [24]: the most important local components of the overall emotional signals are recovered using
convolution and pooling algorithms. This is done while monitoring the sentiment signal semantic linkages,
the distance between each context word and aspect target, and the sequence information.
Int J Artif Intell ISSN: 2252-8938 
Aspect based sentiment analysis using a novel ensemble deep network (Abraham Rajan)
1675
− RAM [25]: it increases resistance to irrelevant data by using a method called multiple attention to identify
sentiment features from a distance.
− MGAN [26]: employs a combination of coarse- and fine-grained attention to capture the interactions
between aspects and contexts at the word level and an aspect alignment loss to explain the aspect-level
interactions between aspects that share a shared context.
− CDT [13]: the BiLSTM-learned embedding are amplified using GCN after learning the feature
representation of sentences with BiLSTM. GCN directly adjusts a phrase's dependency tree.
− DMMN-SDCM [27]: integrates information from semantic analysis into the memory network and assigns
an auxiliary task to learn the sentiment distribution of the entire phrase to offer the necessary background
knowledge for sentiment classification of the target aspect.
− TD-BERT [28]: a max-pooling layer and an FCL are added to the BERT model to create the necessary
sentiment classification model.
− AEN-BERT [29]: uses attention-based encoding to model the link between context and target, examines
the issue of inaccurate labels, and suggests label smoothing regularization. The idea also applies to Bert's
prior condition.
− BERT-PT [30]: after training BERT, a technique is introduced to compensate for this lack of domain and
task-related knowledge and to improve the training's results by utilizing the remaining quantity of data.
− BERT-BASE [31]: in all stages, attempts are made to simultaneously condition on both left and right
contexts to pre-train deep bidirectional representations from the unlabeled text. The pre-trained BERT
model may be enhanced to produce existing models for a variety of purposes with just one additional output
layer. We employ ABSA for multilingual learning.
− RepWalk [32]: finds the context words that have a great impact on the ability of aspect words to predict the
sentiment by randomly searching the grammatical graph. To determine the importance of each word in a
phrase using grammatical knowledge.
− SK-GCN (BERT) [33]: combines two modelling strategies for both GCN's syntactic dependency trees and
common sense information to specifically enhance phrase representation.
− SAGAT-BERT [34]: the BERT language model's external pre-training data and understanding of syntax
perception are applied to the dependent tree structure using the graph attention network, which aids in
replicating the interaction between the context and aspects.
− ASGCN-BERT [7]: uses the syntactic dependent structure of the phrase to solve the long-standing multi-
word reliability issue of ABSA.
− R-GAT+BERT [16]: by using this methodology and pruning the traditional dependency parse tree, a unified
aspect-oriented dependency tree structure depending on the desired aspect is produced. It is also
recommended to encode a novel tree structure for sentiment prediction using a relational graph attention
network (R-GAT).
− LCFS-BERT-CDM [1]: this method recommends employing syntactic relative distance to reduce the
impact of unrelated words with unstable syntactic ties to aspects to improve the aspect-based sentiment
classifier’s accuracy. To improve the efficiency of the aspect extractor, it also incorporates part of speech
embedding, syntactic relation embedding, and context embedding (e.g., BERT and Roberta).
4.3. Experimental analyses
The proposed EDN-ABSA is compared with various state-of-art models and baseline models. Graph
is plotted for accuracy and F1-score for ACL Twitter Dataset. the Baseline models used for comparison here
are IAN, MemNet, ATMS-S, AOA, ASGCN-DG, ASGCN-DT, MAN, and UP-CNN-BERT and state-of-art
models at different times are ATAE-LSTM, GANN, RAM, MGAN, CDT, DMMN-SDCM, TD-BERT,
AEN-BERT, BERT-PT, BERT-BASE, RepWalk, SK-GCN(BERT), SAGAT-BERT, ASGCN-BERT,
R-GAT-BERT, LCFS-BERT-CDM, and LCFS-BERT-CDW.
4.3.1 Accuracy
The accuracy is evaluated for various existing state-of-art models and baseline models and a graph is
plotted against the proposed model we can conclude that the proposed EDN-ABSA model performs better in
comparison with the existing system. Figure 3 shows the accuracy comparison graphically. Table 2 shows the
accuracy comparison. The accuracy value for the baseline method for ASGCN-DT is 71.53 and whereas for
IAN is 72.5 and the maximum value among the baseline methods is MAN methodology, which gives an
accuracy value of 76.56. Whereas the accuracy value for the state-of-art-model for a methodology for
ATAE-LSTM is 68.64 and for the RepWalk methodology, it is 74.4 and the maximum value for the
state-of-art-model at different times is TD-BERT value is 76.7, whereas the proposed EDN-PS model
outperforms other existing state-of-art and baseline models with an accuracy value of 78.5.
 ISSN: 2252-8938
Int J Artif Intell, Vol. 13, No. 2, June 2024: 1668-1678
1676
4.4. F1-score
The F1-score is evaluated for various existing state-of-art models and baseline models and a graph is
plotted against the proposed model we can conclude that the proposed EDN-PS model performs better in
comparison with the existing models. Figure 4 shows the F1-score comparison graphically. Table 3 shows the
F1-score comparison. The F1-score value for the baseline method for ASGCN-DT is 69.68, which is the least,
and whereas for MemNet is 69.9 and the maximum value among the baseline methods is R-GAT-BERT
methodology, which gives an accuracy value of 74.88. Whereas the F1-score value for the state-of-art-model
at different times for methodology for MemNet is 69.9, and for SAGAT-BERT methodology it is 74.17 and
the maximum value for the state-of-art-model at different times is R-GAT-BERT value is 74.88, whereas the
proposed EDN-PS model outperforms other existing state-of-art and baseline models with an F1-score value
of 76.6.
Figure 3. Accuracy comparison Figure 4. F1-score comparison
Table 2. Accuracy comparison
Methodology Accuracy
ATAE-LSTM 68.64
RAM 69.36
UP-CNN-BERT 70.95
MemNet 71.48
ASGCN-DT 71.53
ASGCN-DG 72.15
AOA 72.3
IAN 72.5
MGAN 72.54
ASGCN-BERT 73.19
GANN 74.25
RepWalk 74.4
CDT 74.66
AEN-BERT 74.71
CFS-BERT-CDM 74.82
SK-GCN-BERT 75
SAGAT-BERT 75.4
LGCF-CDM 75.69
GCF-CDM-CDW 75.69
CFS-BERT-CDW 75.83
R-GAT-BERT 76.15
LGCF-CDW 76.27
TD-BERT 76.7
MAN 76.56
EDN-ABSA 78.56
Table 3. F1-score comparison
Methodology F1-score
ATAE-LSTM 66.6
RAM 67.3
ASGCN-DT 69.68
MemNet 69.9
AOA 70.2
ASGCN-DG 70.4
MGAN 70.81
IAN 70.81
GANN 72.01
MAN 72.19
ASGCN-BERT 72.54
RepWalk 72.6
SK-GCN-BERT 73.01
AEN-BERT 73.13
CFS-BERT-CDM 73.5
CDT 73.66
SAGAT-BERT 74.17
LGCF-CDM 74.31
CFS-BERT-CDW 74.4
LGCF-CDW 74.53
GCF-CDM-CDW 74.77
R-GAT-BERT 74.88
EDN-ABSA 76.6
5. CONCLUSION
ABSA, a method of text analysis, divides data into groups and evaluates the sentiment connected to
each group. ABSA may assess customer feedback by linking different sentiments with specific qualities of an
item or service. The existing research is focused on the correlation between the sentiment polarity of various
aspects and the local context. The essential deep correlations between the global context and aspects are not
focused on attention. The proposed EDN-ABSA is compared with various state-of-art models and baseline
models and a graph is plotted for accuracy and F1-score for ACL Twitter Dataset whereas the proposed
Int J Artif Intell ISSN: 2252-8938 
Aspect based sentiment analysis using a novel ensemble deep network (Abraham Rajan)
1677
EDN-PS model outperforms other existing state-of-art and baseline models with an accuracy value of 78.5 and
F1-score value of 76.6.
REFERENCES
[1] M. H. Phan and P. O. Ogunbona, “Modelling context and syntactical features for aspect-based sentiment analysis,” in Proceedings
of the 58th Annual Meeting of the Association for Computational Linguistics, 2020, pp. 3211–3220, doi: 10.18653/v1/2020.acl-
main.293.
[2] N. Majumder, S. Poria, A. Gelbukh, M. S. Akhtar, E. Cambria, and A. Ekbal, “IARM: inter-aspect relation modeling with memory
networks in aspect-based sentiment analysis,” in Proceedings of the 2018 Conference on Empirical Methods in Natural Language
Processing, 2018, pp. 3402–3411, doi: 10.18653/v1/D18-1377.
[3] Y. Wang, M. Huang, X. Zhu, and L. Zhao, “Attention-based LSTM for aspect level sentiment classification,” in Proceedings of the
2016 Conference on Empirical Methods in Natural Language Processing, 2016, pp. 606–615, doi: 10.18653/v1/D16-1058.
[4] D. Ma, S. Li, X. Zhang, and H. Wang, “Interactive attention networks for aspect-level sentiment classification,” in Proceedings of
the Twenty-Sixth International Joint Conference on Artificial Intelligence, Aug. 2017, pp. 4068–4074, doi: 10.24963/ijcai.2017/568.
[5] B. Xu, X. Wang, B. Yang, and Z. Kang, “Target embedding and position attention with LSTM for aspect based sentiment analysis,”
in Proceedings of the 2020 5th International Conference on Mathematics and Artificial Intelligence, Apr. 2020, pp. 93–97, doi:
10.1145/3395260.3395280.
[6] L. Xu, L. Bing, W. Lu, and F. Huang, “Aspect sentiment classification with aspect-specific opinion spans,” in Proceedings of the
2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020, pp. 3561–3567, doi:
10.18653/v1/2020.emnlp-main.288.
[7] C. Zhang, Q. Li, and D. Song, “Aspect-based sentiment classification with aspect-specific graph convolutional networks,” in
Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint
Conference on Natural Language Processing (EMNLP-IJCNLP), 2019, pp. 4568–4578, doi: 10.18653/v1/D19-1464.
[8] X. Hou et al., “Graph ensemble learning over multiple dependency trees for aspect-level sentiment classification,” in Proceedings
of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language
Technologies, 2021, pp. 2884–2894, doi: 10.18653/v1/2021.naacl-main.229.
[9] R. Li, H. Chen, F. Feng, Z. Ma, X. Wang, and E. Hovy, “Dual graph convolutional networks for aspect-based sentiment analysis,”
in Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint
Conference on Natural Language Processing (Volume 1: Long Papers), 2021, pp. 6319–6329, doi: 10.18653/v1/2021.acl-long.494.
[10] B. Huang and K. Carley, “Parameterized convolutional neural networks for aspect level sentiment classification,” in Proceedings
of the 2018 Conference on Empirical Methods in Natural Language Processing, 2018, pp. 1091–1096, doi: 10.18653/v1/D18-1136.
[11] C. Gan, L. Wang, Z. Zhang, and Z. Wang, “Sparse attention based separable dilated convolutional neural network for targeted
sentiment analysis,” Knowledge-Based Systems, vol. 188, p. 104827, Jan. 2020, doi: 10.1016/j.knosys.2019.06.035.
[12] X. Wang, F. Li, Z. Zhang, G. Xu, J. Zhang, and X. Sun, “A unified position-aware convolutional neural network for aspect based
sentiment analysis,” Neurocomputing, vol. 450, pp. 91–103, Aug. 2021, doi: 10.1016/j.neucom.2021.03.092.
[13] K. Sun, R. Zhang, S. Mensah, Y. Mao, and X. Liu, “Aspect-level sentiment analysis via convolution over dependency tree,” in
Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint
Conference on Natural Language Processing (EMNLP-IJCNLP), 2019, pp. 5679–5688, doi: 10.18653/v1/D19-1569.
[14] M. Zhang and T. Qian, “Convolution over hierarchical syntactic and lexical graphs for aspect level sentiment analysis,” in
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020, pp. 3540–3549, doi:
10.18653/v1/2020.emnlp-main.286.
[15] H. Tang, D. Ji, C. Li, and Q. Zhou, “Dependency graph enhanced dual-transformer structure for aspect-based sentiment
classification,” in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020, pp. 6578–6588,
doi: 10.18653/v1/2020.acl-main.588.
[16] K. Wang, W. Shen, Y. Yang, X. Quan, and R. Wang, “Relational graph attention network for aspect-based sentiment analysis,” in
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020, pp. 3229–3238, doi:
10.18653/v1/2020.acl-main.295.
[17] B. Zhang, X. Fu, C. Luo, Y. Ye, X. Li, and L. Jing, “Cross-domain aspect-based sentiment classification by exploiting domain-
invariant semantic-primary feature,” IEEE Transactions on Affective Computing, vol. 14, no. 4, pp. 3106–3119, Oct. 2023, doi:
10.1109/TAFFC.2023.3239540.
[18] M. Hu, S. Zhao, H. Guo, and Z. Su, “Hybrid regularizations for multi-aspect category sentiment analysis,” IEEE Transactions on
Affective Computing, vol. 14, no. 4, pp. 3294–3304, Oct. 2023, doi: 10.1109/TAFFC.2023.3236948.
[19] L. Dong, F. Wei, C. Tan, D. Tang, M. Zhou, and K. Xu, “Adaptive recursive neural network for target-dependent Twitter sentiment
classification,” in Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short
Papers), 2014, vol. 2, pp. 49–54, doi: 10.3115/v1/P14-2009.
[20] D. Tang, B. Qin, and T. Liu, “Aspect level sentiment classification with deep memory network,” in Proceedings of the 2016
Conference on Empirical Methods in Natural Language Processing, 2016, pp. 214–224, doi: 10.18653/v1/D16-1021.
[21] H. Peng, Y. Ma, Y. Li, and E. Cambria, “Learning multi-grained aspect target sequence for Chinese sentiment analysis,” Knowledge-
Based Systems, vol. 148, pp. 167–176, May 2018, doi: 10.1016/j.knosys.2018.02.034.
[22] B. Huang, Y. Ou, and K. M. Carley, “Aspect level sentiment classification with attention-over-attention neural networks,” in Social,
Cultural, and Behavioral Modeling, Cham: Springer, 2018, pp. 197–206, doi: 10.1007/978-3-319-93372-6_22.
[23] Q. Xu, L. Zhu, T. Dai, and C. Yan, “Aspect-based sentiment classification with multi-attention network,” Neurocomputing,
vol. 388, pp. 135–143, May 2020, doi: 10.1016/j.neucom.2020.01.024.
[24] N. Liu and B. Shen, “Aspect-based sentiment analysis with gated alternate neural network,” Knowledge-Based Systems, vol. 188,
p. 105010, Jan. 2020, doi: 10.1016/j.knosys.2019.105010.
[25] P. Chen, Z. Sun, L. Bing, and W. Yang, “Recurrent attention network on memory for aspect sentiment analysis,” in Proceedings of
the 2017 Conference on Empirical Methods in Natural Language Processing, 2017, pp. 452–461, doi: 10.18653/v1/D17-1047.
[26] F. Fan, Y. Feng, and D. Zhao, “Multi-grained attention network for aspectlevel sentiment classification,” in Proceedings of the 2018
Conference on Empirical Methods in Natural Language Processing, 2018, pp. 3433–3442, doi: 10.18653/v1/D18-1380.
[27] P. Lin, M. Yang, and J. Lai, “Deep mask memory network with semantic dependency and context moment for aspect level sentiment
classification,” in Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, Aug. 2019,
pp. 5088–5094, doi: 10.24963/ijcai.2019/707.
 ISSN: 2252-8938
Int J Artif Intell, Vol. 13, No. 2, June 2024: 1668-1678
1678
[28] Z. Gao, A. Feng, X. Song, and X. Wu, “Target-dependent sentiment classification with BERT,” IEEE Access, vol. 7,
pp. 154290–154299, 2019, doi: 10.1109/ACCESS.2019.2946594.
[29] Y. Song, J. Wang, T. Jiang, Z. Liu, and Y. Rao, “Targeted sentiment classification with attentional encoder network,” in Artificial
Neural Networks and Machine Learning – ICANN 2019: Text and Time Series, Cham: Springer, 2019, pp. 93–103, doi:
10.1007/978-3-030-30490-4_9.
[30] H. Xu, B. Liu, L. Shu, and P. S. Yu, “BERT post-training for review reading comprehension and aspect-based sentiment analysis,”
in Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human
Language Technologies, 2019, pp. 2324–2335.
[31] J. Devlin, M. W. Chang, K. Lee, and K. Toutanova, “BERT: Pre-training of deep bidirectional transformers for language
understanding,” Arxiv-Computer Science, pp. 1–16, 2019.
[32] Y. Zheng, R. Zhang, S. Mensah, and Y. Mao, “Replicate, Walk, and Stop on Syntax: An Effective Neural Network Model for
Aspect-Level Sentiment Classification,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 5,
pp. 9685–9692, Apr. 2020, doi: 10.1609/aaai.v34i05.6517.
[33] J. Zhou, J. X. Huang, Q. V. Hu, and L. He, “SK-GCN: modeling syntax and knowledge via graph convolutional network for aspect-
level sentiment classification,” Knowledge-Based Systems, vol. 205, p. 106292, Oct. 2020, doi: 10.1016/j.knosys.2020.106292.
[34] L. Huang, X. Sun, S. Li, L. Zhang, and H. Wang, “Syntax-aware graph attention network for aspect-level sentiment classification,”
in Proceedings of the 28th International Conference on Computational Linguistics, 2020, pp. 799–810, doi:
10.18653/v1/2020.coling-main.69.
BIOGRAPHIES OF AUTHORS
Abraham Rajan earned his Bachelor's of Engineering BE degree in CSE from
VTU, Belagavi in 2016. He has obtained his master's degree in M.Tech (CSE) from Reva
University in 2018. Currently he is a research scholar at Christ University doing his Ph.D. in
Computer Science and Engineering and also working as assistant professor in Nagarjuna
College of Engineering and Technology. He has attended many workshops and induction
programs conducted by various universities. His areas of interest are big data analytics and
cloud computing. He can be contacted at email: abinrajan1994@gmail.com.
Manohara Manur is an associate professor in the Department of Computer
Science and Engineering at School of Engineering and Technology of Christ University,
Bangalore with an experience of 22 years in Teaching. He is qualified in Bachelor and Master
Degrees in Computer Science and Engineering, and Ph.D. in Computer Science and
Engineering in the area of data mining and big data. His areas of interest are data mining,
computer vision, machine learning, artificial intelligence, internet of things, and image
processing. He can be contacted at this email: manohar.m@christuniversity.in.

More Related Content

PDF
Aspect based sentiment analysis using fine-tuned BERT model with deep context...
PDF
Sentiment Analysis Tasks and Approaches
PDF
Leveraging Large Language Models for Aspect-Based Sentiment Classification Us...
PDF
Leveraging Large Language Models for Aspect-Based Sentiment Classification Us...
PDF
D018212428
PDF
Neural Network Based Context Sensitive Sentiment Analysis
PPTX
PPTX
Aspect based sentiment analysis using fine-tuned BERT model with deep context...
Sentiment Analysis Tasks and Approaches
Leveraging Large Language Models for Aspect-Based Sentiment Classification Us...
Leveraging Large Language Models for Aspect-Based Sentiment Classification Us...
D018212428
Neural Network Based Context Sensitive Sentiment Analysis

Similar to Aspect based sentiment analysis using a novel ensemble deep network (20)

PDF
IRJET- Aspect based Sentiment Analysis on Financial Data using Transferred Le...
PDF
SENTIMENT ANALYSIS-AN OBJECTIVE VIEW
PDF
CICLing_2016_paper_52
PDF
Review of Sentiment Analysis: An Hybrid Approach
PDF
Insights to Problems, Research Trend and Progress in Techniques of Sentiment ...
PDF
A scalable, lexicon based technique for sentiment analysis
PDF
IRJET- Opinion Mining and Sentiment Analysis for Online Review
PDF
A Context-Based Algorithm For Sentiment Analysis
PDF
20120140506009
PDF
Aspect-Level Sentiment Analysis On Hotel Reviews
PDF
Attention-Based Multi-Channel Gated Recurrent Neural Networks A Novel Feature...
PDF
Dictionary Based Approach to Sentiment Analysis - A Review
PDF
Ijsrdv8 i10424
PDF
IRJET - Sentiment Analysis for Marketing and Product Review using a Hybrid Ap...
PDF
A BRIEF REVIEW OF SENTIMENT ANALYSIS METHODS
PDF
A BRIEF REVIEW OF SENTIMENT ANALYSIS METHODS
PDF
Graph embedding approach to analyze sentiments on cryptocurrency
PDF
Crime prediction using a hybrid sentiment analysis approach based on the bidi...
PDF
Analyzing sentiment system to specify polarity by lexicon-based
PDF
IRJET- Aspect based Sentiment Analysis on Financial Data using Transferred Le...
SENTIMENT ANALYSIS-AN OBJECTIVE VIEW
CICLing_2016_paper_52
Review of Sentiment Analysis: An Hybrid Approach
Insights to Problems, Research Trend and Progress in Techniques of Sentiment ...
A scalable, lexicon based technique for sentiment analysis
IRJET- Opinion Mining and Sentiment Analysis for Online Review
A Context-Based Algorithm For Sentiment Analysis
20120140506009
Aspect-Level Sentiment Analysis On Hotel Reviews
Attention-Based Multi-Channel Gated Recurrent Neural Networks A Novel Feature...
Dictionary Based Approach to Sentiment Analysis - A Review
Ijsrdv8 i10424
IRJET - Sentiment Analysis for Marketing and Product Review using a Hybrid Ap...
A BRIEF REVIEW OF SENTIMENT ANALYSIS METHODS
A BRIEF REVIEW OF SENTIMENT ANALYSIS METHODS
Graph embedding approach to analyze sentiments on cryptocurrency
Crime prediction using a hybrid sentiment analysis approach based on the bidi...
Analyzing sentiment system to specify polarity by lexicon-based
Ad

More from IAESIJAI (20)

PDF
Hybrid model detection and classification of lung cancer
PDF
Adaptive kernel integration in visual geometry group 16 for enhanced classifi...
PDF
Video forgery: An extensive analysis of inter-and intra-frame manipulation al...
PDF
Enhancing fall detection and classification using Jarratt‐butterfly optimizat...
PDF
Deep ensemble learning with uncertainty aware prediction ranking for cervical...
PDF
Event detection in soccer matches through audio classification using transfer...
PDF
Detecting road damage utilizing retinaNet and mobileNet models on edge devices
PDF
Optimizing deep learning models from multi-objective perspective via Bayesian...
PDF
Squeeze-excitation half U-Net and synthetic minority oversampling technique o...
PDF
A novel scalable deep ensemble learning framework for big data classification...
PDF
Exploring DenseNet architectures with particle swarm optimization: efficient ...
PDF
A transfer learning-based deep neural network for tomato plant disease classi...
PDF
U-Net for wheel rim contour detection in robotic deburring
PDF
Deep learning-based classifier for geometric dimensioning and tolerancing sym...
PDF
Enhancing fire detection capabilities: Leveraging you only look once for swif...
PDF
Accuracy of neural networks in brain wave diagnosis of schizophrenia
PDF
Depression detection through transformers-based emotion recognition in multiv...
PDF
A comparative analysis of optical character recognition models for extracting...
PDF
Enhancing financial cybersecurity via advanced machine learning: analysis, co...
PDF
Crop classification using object-oriented method and Google Earth Engine
Hybrid model detection and classification of lung cancer
Adaptive kernel integration in visual geometry group 16 for enhanced classifi...
Video forgery: An extensive analysis of inter-and intra-frame manipulation al...
Enhancing fall detection and classification using Jarratt‐butterfly optimizat...
Deep ensemble learning with uncertainty aware prediction ranking for cervical...
Event detection in soccer matches through audio classification using transfer...
Detecting road damage utilizing retinaNet and mobileNet models on edge devices
Optimizing deep learning models from multi-objective perspective via Bayesian...
Squeeze-excitation half U-Net and synthetic minority oversampling technique o...
A novel scalable deep ensemble learning framework for big data classification...
Exploring DenseNet architectures with particle swarm optimization: efficient ...
A transfer learning-based deep neural network for tomato plant disease classi...
U-Net for wheel rim contour detection in robotic deburring
Deep learning-based classifier for geometric dimensioning and tolerancing sym...
Enhancing fire detection capabilities: Leveraging you only look once for swif...
Accuracy of neural networks in brain wave diagnosis of schizophrenia
Depression detection through transformers-based emotion recognition in multiv...
A comparative analysis of optical character recognition models for extracting...
Enhancing financial cybersecurity via advanced machine learning: analysis, co...
Crop classification using object-oriented method and Google Earth Engine
Ad

Recently uploaded (20)

PPTX
Effective Security Operations Center (SOC) A Modern, Strategic, and Threat-In...
PPTX
KOM of Painting work and Equipment Insulation REV00 update 25-dec.pptx
PDF
Advanced methodologies resolving dimensionality complications for autism neur...
PDF
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
PDF
Encapsulation theory and applications.pdf
PPTX
Digital-Transformation-Roadmap-for-Companies.pptx
PPTX
Programs and apps: productivity, graphics, security and other tools
PDF
KodekX | Application Modernization Development
PDF
Spectral efficient network and resource selection model in 5G networks
PDF
MIND Revenue Release Quarter 2 2025 Press Release
PPTX
Cloud computing and distributed systems.
PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
PDF
Building Integrated photovoltaic BIPV_UPV.pdf
PDF
Machine learning based COVID-19 study performance prediction
PDF
NewMind AI Weekly Chronicles - August'25 Week I
PDF
Unlocking AI with Model Context Protocol (MCP)
PDF
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
PDF
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
PDF
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
PDF
Mobile App Security Testing_ A Comprehensive Guide.pdf
Effective Security Operations Center (SOC) A Modern, Strategic, and Threat-In...
KOM of Painting work and Equipment Insulation REV00 update 25-dec.pptx
Advanced methodologies resolving dimensionality complications for autism neur...
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
Encapsulation theory and applications.pdf
Digital-Transformation-Roadmap-for-Companies.pptx
Programs and apps: productivity, graphics, security and other tools
KodekX | Application Modernization Development
Spectral efficient network and resource selection model in 5G networks
MIND Revenue Release Quarter 2 2025 Press Release
Cloud computing and distributed systems.
Agricultural_Statistics_at_a_Glance_2022_0.pdf
Building Integrated photovoltaic BIPV_UPV.pdf
Machine learning based COVID-19 study performance prediction
NewMind AI Weekly Chronicles - August'25 Week I
Unlocking AI with Model Context Protocol (MCP)
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
Profit Center Accounting in SAP S/4HANA, S4F28 Col11
Build a system with the filesystem maintained by OSTree @ COSCUP 2025
Mobile App Security Testing_ A Comprehensive Guide.pdf

Aspect based sentiment analysis using a novel ensemble deep network

  • 1. IAES International Journal of Artificial Intelligence (IJ-AI) Vol. 13, No. 2, June 2024, pp. 1668~1678 ISSN: 2252-8938, DOI: 10.11591/ijai.v13.i2.pp1668-1678  1668 Journal homepage: http://ijai.iaescore.com Aspect based sentiment analysis using a novel ensemble deep network Abraham Rajan, Manohar Manur Department of Computer Science and Engineering, Christ University, Bangalore, India Article Info ABSTRACT Article history: Received Apr 5, 2023 Revised Nov 11, 2023 Accepted Dec 1, 2023 Aspect-based sentiment analysis (ABSA) is a fine-grained task in natural language processing, which aims to predict the sentiment polarity of several parts of a sentence or document. The essential aspect of sentiment polarity and global context have deep relationships that have not received enough attention. This research work design and develops a novel ensemble deep network (EDN) which comprises the various network and integrated to enhance the model performance. In the proposed work the words of the input sentence are converted into word vectors using the optimised bidirectional encoder representations from transformers (BERT) model and an optimised BERT-graph neural networks (GNN) model with convolutions is built that analyses the ABSA of the input sentence. The optimised GNN model with convolutions for context-based word representations is developed for the word-vector embedding. We propose a novel EDN for an ABSA model for optimised BERT over GNN with convolutions. The proposed ensemble deep network proposed system (EDN-PS) is evaluated with various existing techniques and results are plotted in terms of metrics for accuracy and F1- score, concluding that the proposed EDN-PS ensures better performance in comparison with the existing model. Keywords: Aspect-based sentiment analysis Ensemble deep network Graph neural network Sentiment analysis This is an open access article under the CC BY-SA license. Corresponding Author: Abraham Rajan Department of Computer Science and Engineering, Christ University Bangalore, India Email: abraham_ranjan@rediffmail.com 1. INTRODUCTION Through blogs, discussion forums, social networking sites, and e-commerce websites, users have expressed their opinions on a wide range of topics, including products, individuals, and organizations, during the past several years. This swapping of ideas is made possible by the fast development of web applications and the broad availability of inexpensive internet connections, both of which generate massive amounts of online data. Sentiment analysis uses this large quantity of data that is used to analyses major challenges. Before the widespread use of the internet as a resource for decision-making, due to the widespread use of the internet and the world wide web, it is now possible to read the opinions and perspectives of anonymous individuals from a range of cultures and places. Many users' shopping opinions are impacted by internet reviews by various other shoppers [1]. Systems that analyses user sentiment will automatically summarize user reviews, which may help a customer, make an opinion. Major advantages of sentiment analysis systems include scalability, the ability to summarize massive volumes of text, real-time analysis, and the capability to deliver fast assessments, and consistent criteria, which is the result of automation and lacks bias compared to user judgment. Organizations, whether public or private, require a sentiment analysis system; it is feasible to supplement standard recommendation systems using sentiment analysis. It assists the owners of the organization in understanding
  • 2. Int J Artif Intell ISSN: 2252-8938  Aspect based sentiment analysis using a novel ensemble deep network (Abraham Rajan) 1669 the sentiments of customers towards their items, however, this is suitable for market research and competition analysis, politics, public policy development, and legal research are further uses [2]. Although sentiment analysis may be undertaken at various levels, aspect-level sentiment analysis is the most useful. Researchers have conducted an extensive study on the identification of explicit sentiments that result in the generation of many alternatives. Due to its complexity, embedded characteristic recognition receives less attention; this is simultaneously a substantial portion of a sentence, which has various parameters, to detect sentiment analysis in understanding the concept. Sentiment analysis research focuses on making predictions at the phrase or document level by assessing the overall sentiment towards the entire phrase or content [2]–[4]. The opinion is based on the perhaps inaccurate assumption that the available data determines a single opinion towards a specific issue. In consequence, the need for aspect-based sentiment analysis (ABSA) is a method for differentiating more complex aspect-level thoughts and feelings, which has surged over the past decade [5], [6]. The target achieved for opinion mining on ABSA shifts its focus from being a phrase or document on a whole to being an entity or a particular characteristic of an entity. For example, an entity may be an online-sold product, with its features including its price and size. Refer to the following example review to understand the words used in aspect-level sentiment analysis: "the display quality of the Samsung TV is outstanding." The term outstanding in this context denotes a favorable opinion of the Samsung TV's display, which is the element of the Samsung TV that is being discussed. As illustrated in Figure 1, the three methods that may be employed to perform aspect-level sentiment analysis are aspect identification, sentiment evaluation concerning the specific aspect, and sentiment aggregation. Figure 1. Aspect level sentiment analysis Context-based methods and syntax-based methods are the two main areas of focus in the existing research for ABSA based on deep neural networks [5], [6]. To gather aspect word representations in a phrase, context-based approaches often employ convolutional neural networks (CNN) or recurrent neural networks in conjunction with attention processes. Graph convolutional networks (GCN) are frequently used to represent phrases in syntax-based models. Recent research on ABSA has relied widely on deep neural models due to their capability to capture the semantic features of specific aspects [1]–[4]. In the existing research on ABSA that uses deep neural networks, context-based techniques [5], [6] and syntax-based methods [7]–[9] are distinguished into two groups. Context-based algorithms frequently combine CNN or recurrent neural networks with attention mechanisms to collect aspect word representations in a sentence, in syntax-based models, GCN are typically employed to represent sentences. Finding several aspect-level sentiment components, including sentiment polarities, opinion terms, aspect words, and aspect categories, is the focus of this research work. Present ABSA models that have been developed to achieve better performance in a variety of tasks, which focuses on training and testing data, which is developed from the same distribution (e.g., the same domain or the same language). When the data distribution changes, it is usually important to retrain the ABSA model to maintain its performance. Huge quantities of labelled data are often too expensive or even impossible to collect for ABSA applications that require aspect-level annotations. Incorporating various other methods for building ABSA-based models, which generalize across various domains and languages, is cross-domain transfer or cross-lingual transfer, which requires adapting the trained model to unknown domains or unknown languages. − Convert the words of the input sentence into word-vectors using the optimized bidirectional encoder representations from transformers (BERT) model. − To extract the context-based word representation using the optimized graph neural network (GNN) model with convolutions over the word-vector embedding. − To develop a BERT-GNN model that analyses the ABSA of the input sentence.
  • 3.  ISSN: 2252-8938 Int J Artif Intell, Vol. 13, No. 2, June 2024: 1668-1678 1670 − To develop a proposed novel ensemble deep network (EDN) for an ABSA model for optimized BERT over GNN with convolution. This particular research is organized as follows: first section of the research starts with background work and significance of sentiment analysis in natural language processing (NLP); section concludes with research motivation and contribution. Second section of the research discuss various existing deep learning protocol along with its shortcoming; third section presents mathematical modelling of EDN. Thereafter, EDN is evaluated in fourth section. 2. RELATED WORK ABSA has been one of the key area of research in sentiment analysis, thus there has been plenty of research in past decade. Traditional approach lacks the deep understanding of the common sense among the sentences. This section discuss several deep learning mechanism for sentiment analysis. Huang and Carley [10] have developed a novel, reliable CNN for aspect-level sentiment analysis, and integrating this aspect data into a CNN with the aspect-level features results in developing a high-accuracy model. Long short-term memory (LSTM) networks are difficult to analyses and have low temporal efficiency, whereas CNN can only allocate local semantic data. The sparse attention-based convolutional neural network (SA-SDCCN) proposed by [11] deals with semantic and sentiment embedding that is added to an embedding tensor at the multichannel embedding layer, which leads to accurate representations of the input sequence. According to Wang et al. [12], a novel model known as unified position-aware convolutional neural network (UP-CNN), which initially proposed an aspect detection network with prior knowledge to accommodate the absence of aspect positions in ACSA, and has proposed an aspect mask to construct aspect-aware contextual representations to accommodate CNN in ABSA. Consequently, several NLP applications have successfully integrated GCNs, by using grammatical links in aspect-level sentiment analysis tasks, Zhang et al. [7] have developed the first aspect-specific graph convolutional network (ASGCN). Initially, the context information is collected using a network with bidirectional long short-term memory (Bi-LSTM), then aspect-based word information is collected using the masked approach. The ABSA approach developed by [13] is based on the convolution of dependency trees. Zhang and Qian [14] introduced a bidirectional GCN model consisting of a hierarchical syntactic network and a lexical graph for each phrase based on syntactic links and word co-occurrence data. Based on a dependency network, Tang et al. [15] suggested a modified double-transformer structure for aspect sentiment categorization that focuses on essential sentence components. Using relational graph-based attention networks, Wang et al. [16] have modified the dependent trees. The previously published GCN-based models perform well on the ABSA; however, they lacked sentiment data across context-related factors. In addition to this the interactive sentiment correlation between aspects recovered by GCNs on the dependency tree with external sentiment-based knowledge and integrate with MHSA to generate semantic information between context and aspects. According to Zhang et al. [17], transformer-based semantic-primary knowledge transferring (TSPKT) network for aspect-term sentiment analysis across domains makes use of semantic-primary information to facilitate the transfer of knowledge across multiple domains. Before semantic-primary information is collected from the S-Graph, it is generated using external semantic lexicons. The second postulate is that Aoa graphormer learns syntactically meaningful aspect-term phrases. Thirdly, the standard by LSTM classifier is developed by adding a knowledge-aware memory unit (KAMU) that holds the whole corpus of semantic primary input. According to Hu et al. [18], this approach focuses on the above-mentioned two issues by combining aspect and task-level regularizations. Aspect-level regularizations limit the attention-based weights to reduce noise. Sparse regularization is more suited for sentences with a single aspect than orthogonal regularization for sentences with various aspects. Here they have proposed a task-level regularization by adding aspect-level category identification as an orthogonal auxiliary task to generate sentiment-dominant features. 3. PROPOSED METHODOLOGY The proposed novel EDN for ABSA model for optimized BERT over GNN with convolutions. Figure 2 shows the proposed architecture. The pooling layer here is responsible to extract features of similar size by selecting the maximum number from each feature vector. Further, this connected layer fine-tunes the model considering the sentiments taken from the previous layer. Finally determining the sentiments from the softmax layer. − Transfer the words of the input sentence into words using the optimized BERT model. − To extract the context-based word representation using the optimized GNN model over the word-vector embedding.
  • 4. Int J Artif Intell ISSN: 2252-8938  Aspect based sentiment analysis using a novel ensemble deep network (Abraham Rajan) 1671 − Build a BERT-GNN model with convolutions that analyses the ABSA of the input sentence. Consider a given set of sentiments D = {d1 , d2, … . . , dn} for l ∈ L, let Sen = {sen1 , sen2, … . . , senn} be a set of sentences for k sentiments. For sen ∈ Sen be the set of words sen = {l1 , l2, … . . , ln} for one sentence denoted assen. In the proposed architecture for a novel EDN for aspect-based sentiment, analysis the input is given in the form of sentiments, which is then converted into a sentence–word-vector. This is further fed into the optimized BERT model where it is converted into word embedding by the optimized BERT model. The optimized LSTM model here extract the context-word representations, the context-based words are fed into the GNN with convolution, which converts the node matrices into the Attention mechanism wherein the further bias is reduced while, processing the GNN model with convolutions. Figure 2. Proposed architecture 3.1. Preliminaries and problem definition A fixed set of c sentiments D = {d1 , d2, … . . , dn} in context with the target entity, for a target sentiment dk. Let fxy be the y − th aspect for the given entity, gxy is the essential feature relevant to the aspect fxy and Sxy determines the sentiment of the aspect fxy. The main aim here is to develop an optimized GNN model for aspect-level sentiment analysis. The main objective here is subjected to the mapping function as mentioned in (1). The proposed model incorporates the following steps, to transform the words of the input provided into word vectors along with the optimized-BERT model. G: (gxy) = { 1, if Sxy is positive 0, if Sxy is neutral −1, if Sxy is negative (1) 3.1.1. Word-vector generation The word vector model automatically inserts a cls at the start of the sentence as well as two separate symbols in the initial phase and ending phase of the term to determine the sentiment. The word sequence here is determined by converting a specific sentence into a separate form v = [cls] + v + [diff] + fx + [diff]. This sequence of words is then fed into this model to obtain the vector for each word Oi as shown in (2). Here the Opt − BERT model (Oi) is the word vector extracted via the pre-trained model, as well as dimo is the dimension for the word vector. The input sequence is determined from the input sequence through c, the sequence obtained through the G = {g1, g2, … , gk}.
  • 5.  ISSN: 2252-8938 Int J Artif Intell, Vol. 13, No. 2, June 2024: 1668-1678 1672 gx = Opt − BERT(Oi) ∈ Hdimo∗1 (2) 3.1.2. Word–sentence–vector converter The word representations are depicted as each word converted into an aggregated form through the entire vector that is condensed into the entire input vector sequence. This study deploys an optimized LSTM model over the word-vector embedding for word-vector context basis word representation. Optimised LSTM by learning the context information through understanding the meaning of words by reading the input given in two different ways. Context-based representations are made according to the following steps. − Step 1: first step is dealing with input layer; the word-vector sequence is represented as G = {g1, g2, … , gk} given as input to the optimised LSTM model. − Step 2: second step is dealing with optimized LSTM this layer is responsible to co-ordinate the context- based information by the remaining words in the direction of the sentence in two different ways for the specific-target word. The optimized LSTM model incorporates a dual-LSTM model, left-to-right model (→ Opt − LSTM ) model, and right-to-left model (← Opt − LSTM), which is responsible to encode the sentence from left-to-right as (→ jx(x = [1, k])) and right-to-left (← jx(x = [1, k])) as shown in (3). → jx = (→ Opt − LSTM(Lg→ j gx + g→ j → j → j x−1 + e→ j ) ∈ Npj (3) ← jx = (→ Opt − LSTM(Lg← j gx + g← j ← j → j x−1 + e← j ) ∈ Npj (4) jx = (L→ j j → jx + L→ j j ← jx + ej ) (5) Here L is the weight matrix, by considering the Lg→ j amidst the input given and the forward hidden vectors, jx for representing the context. jx+1 is the hidden vector represented for jx and jk+1 = 0. jx−1 for the previously hidden vector for jx and j0 = 0. − Step 3: the context-based word-vector matrix denoted as G = {g1, g2, … , gk}∈ Nk∗pj, here pj depicts the size of the vector in the matrix G. 3.2. Building a graph neural network with convolutions − Step 1: build a sentence GNN A sentence-based GNN with convolution denoted as SGNN = ( N, V, AM) including a set of nodes in correspondence with the k words in a given sentence, a set E of edges denotes the dependencies of the adjacent node pairs for syntactic dependency as well as an adjacency matrix AM∈Z|N|∗|N| denoting the node relations denoted as shown in (6): Fxy = { 1, if Nx, Ny ∈ N, and vxy ∈ V 1, if Nx = Ny, 0, else (6) Additionally, this graph SGNN has a feature matrix embedded with nodes denoted as Y = [G] ∈ N|N|∗pj, where each row is denoted as Gx which is responsible for the context-based representation of word node Nx ∈ V. − Step 2: generation of node-embedding Representing the nodes: the matrices are generated as P ∈ N|N|∗pj and F ∈ N|N|∗pj are given as input into a GNN network with convolution shown to generate the nodes represented as (7): H = G(t) = ϑ(PG(b−1) L(b) + q(b) (7) Here b depicts the number of layers in the GNN network with convolutions layer, P ∈ N|N|∗ph; G(0) = Y, here ϑ is a non-linear activation function as ReLU. L(b) ∈ N|N |∗ph for the transformation layer in the b − th layer. q(b) is the bias associated with GNN layers associated with convolution, respectively: P = K−1/2 TAM−1/2 (8) P is the symmetric matrix of AM where T is the degree of the matrix AM whereas; Kxx = ∑ Fxy y (9)
  • 6. Int J Artif Intell ISSN: 2252-8938  Aspect based sentiment analysis using a novel ensemble deep network (Abraham Rajan) 1673 − Step 3: GNN with convolution transformation In this step, the noise is reduced and the bias is associated with processing the GNN network. A position attention mechanism, which captures the necessary parts in the sentence irrespective of the sentiment as shown in (10): Ix = ∑ ϑx |N| x=1 Hx (10) Here, ϑx = exponential(CFtanh (LF[Hx⊕wx]+qF ∑ |N| h=1 exponential(CFtanh (LF[Hh⊕wh]+qF (11) Here LF ∈ N( pj + pv) ∗ (pj + pv), qF ∈ N( pj + pv), and CF ∈ N pj + pv for learning matrix. ⊕ is the concatenation operator. wx ∈ Npv is the weight associated with the x − th word formulated as in (12): wx = { |x−initiate| k , x ≤ initiate 0, initiate ≤ x ≤ terminate |x−terminate| k , x > terminate (12) Here initiate and terminate determines the starting and ending phase of the sentiment in the sentence. 3.3. Constructing the classifier The dimension of the node embedding is reduced by the convolution layer by generating the feature vectors ux, which determines the sliding of the filter R ∈ N r∗ph with length r from x to x + r − 1 to extract the necessary information shown as (13): ux = ReLU(RꭈMx:x+|N|−1 + q) (13) Here x = [1, |N|] determines the order of the node representation M; ꭈ is the convolution operator, the activation function is denoted by ReLU and q is the bias term. The feature vectors are constructed by node representations as shown in (14): u = [u1,u2, … … , u|N|] (14) − Step 1: max-pooling layer The pooling layer builds the feature vectors for the same size by picking the maximum number of each vector uk, the main aim here is to determine the size of feature vectors uk, which depends on the size of each matrix M and R. The size will differ if the length of the sentence and the filter sizes are different, new features such as u, are defined as in (15): u = [u1,u2, … … , u|N|] (15) − Step 2: fine-tuning of the layers This layer is responsible for fine-tuning the characteristics of the previous layer, which necessarily determines the sentiments as shown in (16): Y = SM(HM . u + q) (16) Here HM ∈ Lb∗|N| and q ∈ Hb is the weight matrix and bias of the layer. b is the number of sentiment phases. Training the optimized GNN model with convolution to reduce the entropy loss values. 3.4. Optimizing loss of ensemble deep network To train the CNN model over optimized LSTM model integrated with GNN network with convolutions to reduce the loss function denoted as shown in (17) Loss = − ∑ yk logyk^ + δ||γ||2 q k (17)
  • 7.  ISSN: 2252-8938 Int J Artif Intell, Vol. 13, No. 2, June 2024: 1668-1678 1674 Here q denotes the sentiment classes for the training model, yk denotes the distribution of true sentiment for the k − th class and yk^ denotes the distribution of the sentiment to be predicted for the k − th class, δ denotes the regulation co-efficient, and γ is the parameter set from the previous layer. 4. PERFORMANCE EVALUATION The task of solving ABSA is more complex than conventional text-level sentiment analysis. It focuses on recognizing the characteristics or features of an item described in a text and the perspective has taken on each characteristic. It has been proven that applying optimized BERT to ABSA as a sentence-pair classification task and developing an auxiliary sentence yields better results than the existing state-of-art models and baseline models. Further, we have evaluated our proposed ensemble deep network proposed system (EDN-PS) model with existing state-of-art methods and baseline models in terms of accuracy and F1-score. 4.1. Dataset details To evaluate the efficiency of the proposed model to test the performance of the association for computational linguistics (ACL) twitter social dataset [19]. The sentiment polarity is determined by dividing the dataset into three categories: positive, negative and neutral. Table 1 shows the dataset details. The details of the various methodologies used for comparison with the proposed EDN-PS model. Table 1. Dataset details Dataset Positive Negative Neutral Total Training Testing Training Testing Training Testing Training Testing Twitter 1,561 173 1,560 173 3,127 346 3,127 346 4.2. Compared models To comprehensively evaluate our proposed model, we compared it to certain baseline models and state-of-the-art (SOTA) models from different times. The eight baseline models are evaluated are mentioned here. The seventeen models listed below, including the existing ones, have always performed at SOTA level. 4.2.1. Baseline models: − IAN [4]: constructs attentive representations of the target context and engages in interactive attentive learning of the target and context, respectively. − MemNet [20]: determines the importance of various context words for a single feature and utilizes this information as the semantic representation of the sentence to distinguish the most essential context information of a particular trait. − ATSM-S [21]: classifies the sentiment of extremely complex and coarsely written language by specifically modelling the desired qualities at three distinct granularity levels, followed by categorizing the sentiment. − AOA [22]: aspects and sentences are represented in a decentralized manner, with explicit capture of the relationship between context phrases and aspects. The AOA module simultaneously teaches the model aspects and sentence representation, with an emphasis on the main parts of the phrase. − ASGCN-DG [7]: possesses a structure of undirected dependency. The adjacency matrix is produced using DSPT words to avoid the limitations of attention processes and CNN-based models and depends on an ordered dependency tree. Two networks are distinguished by their adjacency matrices. − MAN [23]: a multi-attention network (MAN) was developed to alleviate the problem of information loss when a coarse-level attention mechanism is employed to simulate aspects. − UP-CNN-BERT [12]: this model proposes a novel unified location-aware CNN that constructs an aspect-aware contextual representation for ABSA using an aspect mask to match the CNN, to compensate for the absence of aspect locations in ACSA. 4.2.2. State-of-art models at different times − ATAE-LSTM [3]: by using LSTM in conjunction with the attention mechanism to do sentence-level semantic modelling to address the issue of aspect-level sentiment analysis. This is accomplished by evaluating the significance of diverse contextual information to a certain attribute. − GANN [24]: the most important local components of the overall emotional signals are recovered using convolution and pooling algorithms. This is done while monitoring the sentiment signal semantic linkages, the distance between each context word and aspect target, and the sequence information.
  • 8. Int J Artif Intell ISSN: 2252-8938  Aspect based sentiment analysis using a novel ensemble deep network (Abraham Rajan) 1675 − RAM [25]: it increases resistance to irrelevant data by using a method called multiple attention to identify sentiment features from a distance. − MGAN [26]: employs a combination of coarse- and fine-grained attention to capture the interactions between aspects and contexts at the word level and an aspect alignment loss to explain the aspect-level interactions between aspects that share a shared context. − CDT [13]: the BiLSTM-learned embedding are amplified using GCN after learning the feature representation of sentences with BiLSTM. GCN directly adjusts a phrase's dependency tree. − DMMN-SDCM [27]: integrates information from semantic analysis into the memory network and assigns an auxiliary task to learn the sentiment distribution of the entire phrase to offer the necessary background knowledge for sentiment classification of the target aspect. − TD-BERT [28]: a max-pooling layer and an FCL are added to the BERT model to create the necessary sentiment classification model. − AEN-BERT [29]: uses attention-based encoding to model the link between context and target, examines the issue of inaccurate labels, and suggests label smoothing regularization. The idea also applies to Bert's prior condition. − BERT-PT [30]: after training BERT, a technique is introduced to compensate for this lack of domain and task-related knowledge and to improve the training's results by utilizing the remaining quantity of data. − BERT-BASE [31]: in all stages, attempts are made to simultaneously condition on both left and right contexts to pre-train deep bidirectional representations from the unlabeled text. The pre-trained BERT model may be enhanced to produce existing models for a variety of purposes with just one additional output layer. We employ ABSA for multilingual learning. − RepWalk [32]: finds the context words that have a great impact on the ability of aspect words to predict the sentiment by randomly searching the grammatical graph. To determine the importance of each word in a phrase using grammatical knowledge. − SK-GCN (BERT) [33]: combines two modelling strategies for both GCN's syntactic dependency trees and common sense information to specifically enhance phrase representation. − SAGAT-BERT [34]: the BERT language model's external pre-training data and understanding of syntax perception are applied to the dependent tree structure using the graph attention network, which aids in replicating the interaction between the context and aspects. − ASGCN-BERT [7]: uses the syntactic dependent structure of the phrase to solve the long-standing multi- word reliability issue of ABSA. − R-GAT+BERT [16]: by using this methodology and pruning the traditional dependency parse tree, a unified aspect-oriented dependency tree structure depending on the desired aspect is produced. It is also recommended to encode a novel tree structure for sentiment prediction using a relational graph attention network (R-GAT). − LCFS-BERT-CDM [1]: this method recommends employing syntactic relative distance to reduce the impact of unrelated words with unstable syntactic ties to aspects to improve the aspect-based sentiment classifier’s accuracy. To improve the efficiency of the aspect extractor, it also incorporates part of speech embedding, syntactic relation embedding, and context embedding (e.g., BERT and Roberta). 4.3. Experimental analyses The proposed EDN-ABSA is compared with various state-of-art models and baseline models. Graph is plotted for accuracy and F1-score for ACL Twitter Dataset. the Baseline models used for comparison here are IAN, MemNet, ATMS-S, AOA, ASGCN-DG, ASGCN-DT, MAN, and UP-CNN-BERT and state-of-art models at different times are ATAE-LSTM, GANN, RAM, MGAN, CDT, DMMN-SDCM, TD-BERT, AEN-BERT, BERT-PT, BERT-BASE, RepWalk, SK-GCN(BERT), SAGAT-BERT, ASGCN-BERT, R-GAT-BERT, LCFS-BERT-CDM, and LCFS-BERT-CDW. 4.3.1 Accuracy The accuracy is evaluated for various existing state-of-art models and baseline models and a graph is plotted against the proposed model we can conclude that the proposed EDN-ABSA model performs better in comparison with the existing system. Figure 3 shows the accuracy comparison graphically. Table 2 shows the accuracy comparison. The accuracy value for the baseline method for ASGCN-DT is 71.53 and whereas for IAN is 72.5 and the maximum value among the baseline methods is MAN methodology, which gives an accuracy value of 76.56. Whereas the accuracy value for the state-of-art-model for a methodology for ATAE-LSTM is 68.64 and for the RepWalk methodology, it is 74.4 and the maximum value for the state-of-art-model at different times is TD-BERT value is 76.7, whereas the proposed EDN-PS model outperforms other existing state-of-art and baseline models with an accuracy value of 78.5.
  • 9.  ISSN: 2252-8938 Int J Artif Intell, Vol. 13, No. 2, June 2024: 1668-1678 1676 4.4. F1-score The F1-score is evaluated for various existing state-of-art models and baseline models and a graph is plotted against the proposed model we can conclude that the proposed EDN-PS model performs better in comparison with the existing models. Figure 4 shows the F1-score comparison graphically. Table 3 shows the F1-score comparison. The F1-score value for the baseline method for ASGCN-DT is 69.68, which is the least, and whereas for MemNet is 69.9 and the maximum value among the baseline methods is R-GAT-BERT methodology, which gives an accuracy value of 74.88. Whereas the F1-score value for the state-of-art-model at different times for methodology for MemNet is 69.9, and for SAGAT-BERT methodology it is 74.17 and the maximum value for the state-of-art-model at different times is R-GAT-BERT value is 74.88, whereas the proposed EDN-PS model outperforms other existing state-of-art and baseline models with an F1-score value of 76.6. Figure 3. Accuracy comparison Figure 4. F1-score comparison Table 2. Accuracy comparison Methodology Accuracy ATAE-LSTM 68.64 RAM 69.36 UP-CNN-BERT 70.95 MemNet 71.48 ASGCN-DT 71.53 ASGCN-DG 72.15 AOA 72.3 IAN 72.5 MGAN 72.54 ASGCN-BERT 73.19 GANN 74.25 RepWalk 74.4 CDT 74.66 AEN-BERT 74.71 CFS-BERT-CDM 74.82 SK-GCN-BERT 75 SAGAT-BERT 75.4 LGCF-CDM 75.69 GCF-CDM-CDW 75.69 CFS-BERT-CDW 75.83 R-GAT-BERT 76.15 LGCF-CDW 76.27 TD-BERT 76.7 MAN 76.56 EDN-ABSA 78.56 Table 3. F1-score comparison Methodology F1-score ATAE-LSTM 66.6 RAM 67.3 ASGCN-DT 69.68 MemNet 69.9 AOA 70.2 ASGCN-DG 70.4 MGAN 70.81 IAN 70.81 GANN 72.01 MAN 72.19 ASGCN-BERT 72.54 RepWalk 72.6 SK-GCN-BERT 73.01 AEN-BERT 73.13 CFS-BERT-CDM 73.5 CDT 73.66 SAGAT-BERT 74.17 LGCF-CDM 74.31 CFS-BERT-CDW 74.4 LGCF-CDW 74.53 GCF-CDM-CDW 74.77 R-GAT-BERT 74.88 EDN-ABSA 76.6 5. CONCLUSION ABSA, a method of text analysis, divides data into groups and evaluates the sentiment connected to each group. ABSA may assess customer feedback by linking different sentiments with specific qualities of an item or service. The existing research is focused on the correlation between the sentiment polarity of various aspects and the local context. The essential deep correlations between the global context and aspects are not focused on attention. The proposed EDN-ABSA is compared with various state-of-art models and baseline models and a graph is plotted for accuracy and F1-score for ACL Twitter Dataset whereas the proposed
  • 10. Int J Artif Intell ISSN: 2252-8938  Aspect based sentiment analysis using a novel ensemble deep network (Abraham Rajan) 1677 EDN-PS model outperforms other existing state-of-art and baseline models with an accuracy value of 78.5 and F1-score value of 76.6. REFERENCES [1] M. H. Phan and P. O. Ogunbona, “Modelling context and syntactical features for aspect-based sentiment analysis,” in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020, pp. 3211–3220, doi: 10.18653/v1/2020.acl- main.293. [2] N. Majumder, S. Poria, A. Gelbukh, M. S. Akhtar, E. Cambria, and A. Ekbal, “IARM: inter-aspect relation modeling with memory networks in aspect-based sentiment analysis,” in Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2018, pp. 3402–3411, doi: 10.18653/v1/D18-1377. [3] Y. Wang, M. Huang, X. Zhu, and L. Zhao, “Attention-based LSTM for aspect level sentiment classification,” in Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, 2016, pp. 606–615, doi: 10.18653/v1/D16-1058. [4] D. Ma, S. Li, X. Zhang, and H. Wang, “Interactive attention networks for aspect-level sentiment classification,” in Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, Aug. 2017, pp. 4068–4074, doi: 10.24963/ijcai.2017/568. [5] B. Xu, X. Wang, B. Yang, and Z. Kang, “Target embedding and position attention with LSTM for aspect based sentiment analysis,” in Proceedings of the 2020 5th International Conference on Mathematics and Artificial Intelligence, Apr. 2020, pp. 93–97, doi: 10.1145/3395260.3395280. [6] L. Xu, L. Bing, W. Lu, and F. Huang, “Aspect sentiment classification with aspect-specific opinion spans,” in Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020, pp. 3561–3567, doi: 10.18653/v1/2020.emnlp-main.288. [7] C. Zhang, Q. Li, and D. Song, “Aspect-based sentiment classification with aspect-specific graph convolutional networks,” in Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 2019, pp. 4568–4578, doi: 10.18653/v1/D19-1464. [8] X. Hou et al., “Graph ensemble learning over multiple dependency trees for aspect-level sentiment classification,” in Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2021, pp. 2884–2894, doi: 10.18653/v1/2021.naacl-main.229. [9] R. Li, H. Chen, F. Feng, Z. Ma, X. Wang, and E. Hovy, “Dual graph convolutional networks for aspect-based sentiment analysis,” in Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), 2021, pp. 6319–6329, doi: 10.18653/v1/2021.acl-long.494. [10] B. Huang and K. Carley, “Parameterized convolutional neural networks for aspect level sentiment classification,” in Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2018, pp. 1091–1096, doi: 10.18653/v1/D18-1136. [11] C. Gan, L. Wang, Z. Zhang, and Z. Wang, “Sparse attention based separable dilated convolutional neural network for targeted sentiment analysis,” Knowledge-Based Systems, vol. 188, p. 104827, Jan. 2020, doi: 10.1016/j.knosys.2019.06.035. [12] X. Wang, F. Li, Z. Zhang, G. Xu, J. Zhang, and X. Sun, “A unified position-aware convolutional neural network for aspect based sentiment analysis,” Neurocomputing, vol. 450, pp. 91–103, Aug. 2021, doi: 10.1016/j.neucom.2021.03.092. [13] K. Sun, R. Zhang, S. Mensah, Y. Mao, and X. Liu, “Aspect-level sentiment analysis via convolution over dependency tree,” in Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), 2019, pp. 5679–5688, doi: 10.18653/v1/D19-1569. [14] M. Zhang and T. Qian, “Convolution over hierarchical syntactic and lexical graphs for aspect level sentiment analysis,” in Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), 2020, pp. 3540–3549, doi: 10.18653/v1/2020.emnlp-main.286. [15] H. Tang, D. Ji, C. Li, and Q. Zhou, “Dependency graph enhanced dual-transformer structure for aspect-based sentiment classification,” in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020, pp. 6578–6588, doi: 10.18653/v1/2020.acl-main.588. [16] K. Wang, W. Shen, Y. Yang, X. Quan, and R. Wang, “Relational graph attention network for aspect-based sentiment analysis,” in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2020, pp. 3229–3238, doi: 10.18653/v1/2020.acl-main.295. [17] B. Zhang, X. Fu, C. Luo, Y. Ye, X. Li, and L. Jing, “Cross-domain aspect-based sentiment classification by exploiting domain- invariant semantic-primary feature,” IEEE Transactions on Affective Computing, vol. 14, no. 4, pp. 3106–3119, Oct. 2023, doi: 10.1109/TAFFC.2023.3239540. [18] M. Hu, S. Zhao, H. Guo, and Z. Su, “Hybrid regularizations for multi-aspect category sentiment analysis,” IEEE Transactions on Affective Computing, vol. 14, no. 4, pp. 3294–3304, Oct. 2023, doi: 10.1109/TAFFC.2023.3236948. [19] L. Dong, F. Wei, C. Tan, D. Tang, M. Zhou, and K. Xu, “Adaptive recursive neural network for target-dependent Twitter sentiment classification,” in Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 2014, vol. 2, pp. 49–54, doi: 10.3115/v1/P14-2009. [20] D. Tang, B. Qin, and T. Liu, “Aspect level sentiment classification with deep memory network,” in Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, 2016, pp. 214–224, doi: 10.18653/v1/D16-1021. [21] H. Peng, Y. Ma, Y. Li, and E. Cambria, “Learning multi-grained aspect target sequence for Chinese sentiment analysis,” Knowledge- Based Systems, vol. 148, pp. 167–176, May 2018, doi: 10.1016/j.knosys.2018.02.034. [22] B. Huang, Y. Ou, and K. M. Carley, “Aspect level sentiment classification with attention-over-attention neural networks,” in Social, Cultural, and Behavioral Modeling, Cham: Springer, 2018, pp. 197–206, doi: 10.1007/978-3-319-93372-6_22. [23] Q. Xu, L. Zhu, T. Dai, and C. Yan, “Aspect-based sentiment classification with multi-attention network,” Neurocomputing, vol. 388, pp. 135–143, May 2020, doi: 10.1016/j.neucom.2020.01.024. [24] N. Liu and B. Shen, “Aspect-based sentiment analysis with gated alternate neural network,” Knowledge-Based Systems, vol. 188, p. 105010, Jan. 2020, doi: 10.1016/j.knosys.2019.105010. [25] P. Chen, Z. Sun, L. Bing, and W. Yang, “Recurrent attention network on memory for aspect sentiment analysis,” in Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, 2017, pp. 452–461, doi: 10.18653/v1/D17-1047. [26] F. Fan, Y. Feng, and D. Zhao, “Multi-grained attention network for aspectlevel sentiment classification,” in Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2018, pp. 3433–3442, doi: 10.18653/v1/D18-1380. [27] P. Lin, M. Yang, and J. Lai, “Deep mask memory network with semantic dependency and context moment for aspect level sentiment classification,” in Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, Aug. 2019, pp. 5088–5094, doi: 10.24963/ijcai.2019/707.
  • 11.  ISSN: 2252-8938 Int J Artif Intell, Vol. 13, No. 2, June 2024: 1668-1678 1678 [28] Z. Gao, A. Feng, X. Song, and X. Wu, “Target-dependent sentiment classification with BERT,” IEEE Access, vol. 7, pp. 154290–154299, 2019, doi: 10.1109/ACCESS.2019.2946594. [29] Y. Song, J. Wang, T. Jiang, Z. Liu, and Y. Rao, “Targeted sentiment classification with attentional encoder network,” in Artificial Neural Networks and Machine Learning – ICANN 2019: Text and Time Series, Cham: Springer, 2019, pp. 93–103, doi: 10.1007/978-3-030-30490-4_9. [30] H. Xu, B. Liu, L. Shu, and P. S. Yu, “BERT post-training for review reading comprehension and aspect-based sentiment analysis,” in Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 2019, pp. 2324–2335. [31] J. Devlin, M. W. Chang, K. Lee, and K. Toutanova, “BERT: Pre-training of deep bidirectional transformers for language understanding,” Arxiv-Computer Science, pp. 1–16, 2019. [32] Y. Zheng, R. Zhang, S. Mensah, and Y. Mao, “Replicate, Walk, and Stop on Syntax: An Effective Neural Network Model for Aspect-Level Sentiment Classification,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 5, pp. 9685–9692, Apr. 2020, doi: 10.1609/aaai.v34i05.6517. [33] J. Zhou, J. X. Huang, Q. V. Hu, and L. He, “SK-GCN: modeling syntax and knowledge via graph convolutional network for aspect- level sentiment classification,” Knowledge-Based Systems, vol. 205, p. 106292, Oct. 2020, doi: 10.1016/j.knosys.2020.106292. [34] L. Huang, X. Sun, S. Li, L. Zhang, and H. Wang, “Syntax-aware graph attention network for aspect-level sentiment classification,” in Proceedings of the 28th International Conference on Computational Linguistics, 2020, pp. 799–810, doi: 10.18653/v1/2020.coling-main.69. BIOGRAPHIES OF AUTHORS Abraham Rajan earned his Bachelor's of Engineering BE degree in CSE from VTU, Belagavi in 2016. He has obtained his master's degree in M.Tech (CSE) from Reva University in 2018. Currently he is a research scholar at Christ University doing his Ph.D. in Computer Science and Engineering and also working as assistant professor in Nagarjuna College of Engineering and Technology. He has attended many workshops and induction programs conducted by various universities. His areas of interest are big data analytics and cloud computing. He can be contacted at email: abinrajan1994@gmail.com. Manohara Manur is an associate professor in the Department of Computer Science and Engineering at School of Engineering and Technology of Christ University, Bangalore with an experience of 22 years in Teaching. He is qualified in Bachelor and Master Degrees in Computer Science and Engineering, and Ph.D. in Computer Science and Engineering in the area of data mining and big data. His areas of interest are data mining, computer vision, machine learning, artificial intelligence, internet of things, and image processing. He can be contacted at this email: manohar.m@christuniversity.in.