Aspect-based sentiment analysis (ABSA) provides an opportunity to systematically generate user's opinions of specific aspects to enrich the idea creation process in the early stage of product/service design process. Yet, the current ABSA task has two major limitations. First, existing research mostly focusing on the subsets of ABSA task, e.g. aspect-sentiment extraction, extract aspect, opinion, and sentiment in a unified model is still an open problem. Second, the implicit opinion and sentiment are ignored in the current ABSA task. This article tackles these gaps by (1) creating a new annotated dataset comprised of five types of labels, including aspect, category, opinion, sentiment, and implicit indicator (ACOSI) and (2) developing a unified model which could extract all five types of labels simultaneously in a generative manner. Numerical experiments conducted on the manually labeled dataset originally scraped from three major e-Commerce retail stores for apparel and footwear products indicate the performance, scalability, and potentials of the framework developed. Several directions are provided for future exploration in the area of automated aspect-based sentiment analysis for user-centered design.
more »
« less
A Design Knowledge Guided Position Encoding Methodology for Implicit Need Identification From User Reviews
Aspect-based sentiment analysis (ABSA) enables a systematic identification of user opinions on particular aspects, thus enhancing the idea creation process in the initial stages of product/service design. Attention-based large language models (LLMs) like BERT and T5 have proven powerful in ABSA tasks. Yet, several key limitations remain, both regarding the ABSA task and the capabilities of attention-based models. First, existing research mainly focuses on relatively simpler ABSA tasks such as aspect-based sentiment analysis, while the task of extracting aspect, opinion, and sentiment in a unified model remains largely unaddressed. Second, current ABSA tasks overlook implicit opinions and sentiments. Third, most attention-based LLMs like BERT use position encoding in a linear projected manner or through split-position relations in word distance schemes, which could lead to relation biases during the training process. This article addresses these gaps by (1) creating a new annotated dataset with five types of labels, including aspect, category, opinion, sentiment, and implicit indicator (ACOSI), (2) developing a unified model capable of extracting all five types of labels simultaneously in a generative manner, and (3) designing a new position encoding method in the attention-based model. The numerical experiments conducted on a manually labeled dataset scraped from three major e-Commerce retail stores for apparel and footwear products demonstrate the performance, scalability, and potential of the framework developed. The article concludes with recommendations for future research on automated need finding and sentiment analysis for user-centered design.
more »
« less
- Award ID(s):
- 2050052
- PAR ID:
- 10590698
- Publisher / Repository:
- American Society of Mechanical Engineers
- Date Published:
- ISBN:
- 978-0-7918-8729-5
- Format(s):
- Medium: X
- Location:
- Boston, Massachusetts, USA
- Sponsoring Org:
- National Science Foundation
More Like this
-
-
Proc. 2023 ACM Int. Conf. on Web Search and Data Mining (Ed.)Target-oriented opinion summarization is to profile a target by extracting user opinions from multiple related documents. Instead of simply mining opinion ratings on a target (e.g., a restaurant) or on multiple aspects (e.g., food, service) of a target, it is desirable to go deeper, to mine opinion on fine-grained sub-aspects (e.g., fish). However, it is expensive to obtain high-quality annotations at such fine-grained scale. This leads to our proposal of a new framework, FineSum, which advances the frontier of opinion analysis in three aspects: (1) minimal supervision, where no document-summary pairs are provided, only aspect names and a few aspect/sentiment keywords are available; (2) fine-grained opinion analysis, where sentiment analysis drills down to a specific subject or characteristic within each general aspect; and (3) phrase-based summarization, where short phrases are taken as basic units for summarization, and semantically coherent phrases are gathered to improve the consistency and comprehensiveness of summary. Given a large corpus with no annotation, FineSum first automatically identifies potential spans of opinion phrases, and further reduces the noise in identification results using aspect and sentiment classifiers. It then constructs multiple fine-grained opinion clusters under each aspect and sentiment. Each cluster expresses uniform opinions towards certain sub-aspects (e.g., “fish” in “food” aspect) or characteristics (e.g., “Mexican” in “food” aspect). To accomplish this, we train a spherical word embedding space to explicitly represent different aspects and sentiments. We then distill the knowledge from embedding to a contextualized phrase classifier, and perform clustering using the contextualized opinion-aware phrase embedding. Both automatic evaluations on the benchmark and quantitative human evaluation validate the effectiveness of our approach.more » « less
-
This paper focuses on learning domain oriented language models driven by end tasks, which aims to combine the worlds of both general-purpose language models (such as ELMo and BERT) and domain-specific language understanding. We propose DomBERT, an extension of BERT to learn from both in-domain corpus and relevant domain corpora. This helps in learning domain language models with low-resources. Experiments are conducted on an assortment of tasks in aspect based sentiment analysis (ABSA), demonstrating promising results.more » « less
-
null (Ed.)Aspect category detection (ACD) is one of the challenging sub-tasks in aspect-based sentiment analysis. The goal of this task is to detect implicit or explicit aspect categories from the sentences of user-generated reviews. Since annotation over the aspects is time-consuming, the amount of labeled data is limited for super-vised learning. In this paper, we study contextual representations of reviews using the BERT model to better extract useful features from text segments in the reviews, and train a supervised classifier with a small amount of labeled data for the ACD task. Experimental results obtained on Amazon reviews of six product domains show that our method is effective in some domains.more » « less
-
null (Ed.)Targeted opinion word extraction (TOWE) is a sub-task of aspect based sentiment analysis (ABSA) which aims to find the opinion words for a given aspect-term in a sentence. Despite their success for TOWE, the current deep learning models fail to exploit the syntactic information of the sentences that have been proved to be useful for TOWE in the prior research. In this work, we propose to incorporate the syntactic structures of the sentences into the deep learning models for TOWE, leveraging the syntax-based opinion possibility scores and the syntactic connections between the words. We also introduce a novel regularization technique to improve the performance of the deep learning models based on the representation distinctions between the words in TOWE. The proposed model is extensively analyzed and achieves the state-of-the-art performance on four benchmark datasets.more » « less
An official website of the United States government

