TY - GEN
T1 - An Unsupervised Deep Learning Model for Aspect Retrieving Using Transformer Encoder
AU - Dey, Atanu
AU - Jenamani, Mamata
AU - De, Arijit
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024.
PY - 2024/6/13
Y1 - 2024/6/13
N2 - We introduce a deep-learning-based aspect extraction model called RATE, which stands for Retrieving of Aspects using Transformer Encoder. When doing unsupervised aspect-based sentiment analysis, the process of retrieving aspects is both critical and challenging. Most prior efforts use some kind of topic modeling to extract aspect only. Despite their efficacy, these techniques seldom provide consistent outcomes for highly coherent aspects. Even though some approaches address these issues by employing an attention-based deep neural model, their performance is hindered by their single-headed attention mechanism. Thus, RATE is designed to improve the performance of extracting coherent aspects from the text by using multi-headed attention mechanism with transformer encoder, negative samplings, and word embeddings. This model promotes the proximity in the embedding space of words that arise in similar contexts, as opposed to topic models and other techniques that often presume independently created words. To further enhance the coherence of the aspects, we use multi-headed attention technique in the encoder of the RATE architecture to downplay unimportant words during training. The RATE model outperforms the highest performing unsupervised baseline ones in terms of precision (8.34%), recall (0.94%), and f1-score (5.44%) on the ACOS-Laptop dataset. Again, on the ACOS-Restaurant dataset, RATE enhances precision, recall, and f1-score by 1.4%, 8.87%, and 4.31%, respectively, for finding more significant and coherent aspects.
AB - We introduce a deep-learning-based aspect extraction model called RATE, which stands for Retrieving of Aspects using Transformer Encoder. When doing unsupervised aspect-based sentiment analysis, the process of retrieving aspects is both critical and challenging. Most prior efforts use some kind of topic modeling to extract aspect only. Despite their efficacy, these techniques seldom provide consistent outcomes for highly coherent aspects. Even though some approaches address these issues by employing an attention-based deep neural model, their performance is hindered by their single-headed attention mechanism. Thus, RATE is designed to improve the performance of extracting coherent aspects from the text by using multi-headed attention mechanism with transformer encoder, negative samplings, and word embeddings. This model promotes the proximity in the embedding space of words that arise in similar contexts, as opposed to topic models and other techniques that often presume independently created words. To further enhance the coherence of the aspects, we use multi-headed attention technique in the encoder of the RATE architecture to downplay unimportant words during training. The RATE model outperforms the highest performing unsupervised baseline ones in terms of precision (8.34%), recall (0.94%), and f1-score (5.44%) on the ACOS-Laptop dataset. Again, on the ACOS-Restaurant dataset, RATE enhances precision, recall, and f1-score by 1.4%, 8.87%, and 4.31%, respectively, for finding more significant and coherent aspects.
KW - Aspect extraction
KW - Attention mechanism
KW - Encoder
KW - Transformer
KW - Unsupervised deep neural network
KW - Word embedding
UR - http://www.scopus.com/inward/record.url?scp=85197469692&partnerID=8YFLogxK
U2 - 10.1007/978-3-031-62277-9_18
DO - 10.1007/978-3-031-62277-9_18
M3 - Conference contribution
SN - 978-3-031-62276-2
T3 - Lecture Notes in Networks and Systems
SP - 303
EP - 317
BT - Lecture Notes in Networks and Systems
A2 - Arai, Kohei
ER -