A Comparison of Potentials and Limitations of Transformer Models for Aspect-Based Medical Sentiment Analysis
Identifiers
10.1007/978-3-032-00652-3_25
Date Issued
2025
Author(s)
Deng, Yihan
Editor(s)
Cafolla, Daniele
Rittman, Timothy
Ni, Hao
Type
Book Chapter
Language
English
Abstract
Transformer models have garnered significant attention for various tasks, such as sentiment analysis. From basic transformer architectures to generative models, numerous approaches have been applied and tested for sentiment analysis. However, the limitations and challenges associated with these models have not yet been adequately assessed. In this study, we aim to explore both the potential and the limitations of two types of transformer-based models (RoBERTa-XLM and GPT-2) for a spect-based sentiment analysis. Our evaluation is conducted using data from the Swiss Sleep Database. We conclude, that techniques like fine-tuning, data balancing, expert-driven normalization, and negationaware processing techniques are essential to improve their performance in medical contexts.
Publisher DOI
Volume
16038
Publisher
Springer
Submitter
Denecke, Kerstin
Citation apa
Deng, Y., & Denecke, K. (2025). A Comparison of Potentials and Limitations of Transformer Models for Aspect-Based Medical Sentiment Analysis (D. Cafolla, T. Rittman, & H. Ni, Eds.; Vol. 16038). Springer. https://arbor.bfh.ch/handle/arbor/45577
File(s)![Thumbnail Image]()
Loading...
restricted
Name
AIiH_SentimentPaper.pdf
Description
Version published
License
Publisher
Size
916.81 KB
Format
Adobe PDF
Checksum (MD5)
a6f3e33e887b73835637ef1ddc1b5368
