A Deep Dive into Interactive Attention Networks: Are Computers On Their Way to Replacing Human Sentiment Analysis?

Category Computer Science

tldr #

Researchers at Anhui University of Science & Technology in China have recently developed a new model, the Lightweight Multilayer Interactive Attention Network (LMIAN), to effectively complete Aspect-based Sentiment Analysis (ABSA) tasks. Initial evaluations show the model being able to effectively identify the sentiment polarity of words in sentences with an accuracy of over 90%, and with less GPU memory than other networks. This model has potential applications in chatbot-based customer service and review-based recommender systems.


content #

In recent years, computer scientists have been trying to develop effective models for sentiment analysis. These models are designed to analyze sentences or longer texts and autonomously determine their underlying emotional tone.

In addition to expressing an overall emotional tone, texts can contain individual words that are charged with positive or negative emotions, as well as neutral words. For example, if we are complaining about something, the word describing the thing in question will most likely be charged with a negative emotion.

Interactive attention networks have been used to identify the sentiment polarity of individual words in sentences with an accuracy of over 90%.

The task of identifying the 'sentiment polarity' (i.e., the negative or positive emotional connotations) of specific words in a sentence is known as aspect-based sentiment analysis (ABSA). While this task can be intuitive and simple for humans, it can be far more challenging to tackle using computational models.

Researchers at Anhui University of Science & Technology in China have recently developed a new model designed to effectively complete ABSA tasks. This model, introduced in a paper in Connection Science, is based on a machine learning algorithm known as lightweight multilayer interactive attention network (LMIAN). "Existing studies have recognized the value of interactive learning in ABSA and have developed various methods to precisely model aspect words and their contexts through interactive learning," Wenjun Zheng, Shunxiang Zhang and their colleagues wrote in their paper. "However, these methods mostly take a shallow interactive way to model aspect words and their contexts, which may lead to the lack of complex sentiment information. To solve this issue, we propose a Lightweight Multilayer Interactive Attention Network (LMIAN) for ABSA." .

The team's Lightweight Multilayer Interactive Attention Network (LMIAN) consumes less GPU memory than other networks.

Interactive attention networks can learn to pay attention to specific words in a sentence via an attention mechanism, which allows them to link these words to the overall linguistic context in which they are found. As part of their recent study, Zheng, Zhang and their colleagues trained one of these models to specifically learn the complex emotional information hidden in a text, so that it can then detect the sentiment polarity of specific words.

The model achieved promising result in six publicly available sentiment analysis datasets (containing texts in Chinese and English).

"We first employ a pre-trained language model to initialize word embedding vectors," Zheng, Zhang and their colleagues wrote in their paper. "Second, an interactive computational layer is designed to build correlations between aspect words and their contexts. Such correlation degree is calculated by multiple computational layers with neural attention models. Third, we use a parameter-sharing strategy among the computational layers. This allows the model to learn complex sentiment features with lower memory costs." .

The model uses a pre-trained language model to initialize the word embedding vectors.

The researchers trained and evaluated their model using six publicly available sentiment analysis datasets containing texts in Chinese and English. These texts are essentially online reviews of products or services, with an associated object and a sentiment label (i.e., positive, negative or neutral), which ultimately allowed the team to train their model.

In initial evaluations on these datasets, the team's LMIAN achieved highly promising results, identifying the sentiment polarity of words in sentences with accuracies over 90%, while cosuming less GPU memory than other networks. In the future, its performance could be improved furter, potentially being used in various applications such as chatbot-based customer service and review-based recommender systems.

The model employs a parameter-sharing strategy among multiple computational layers to learn complex sentiment features.

hashtags #
worddensity #

Share