Transformers in health: a systematic review on architectures for longitudinal data analysis
Version
Published
Date Issued
2024
Author(s)
Type
Article
Language
English
Abstract
Transformers are state-of-the-art technology to support diverse Natural Language Processing (NLP) tasks, such as language translation and word/sentence predictions. The main advantage of transformers is their ability to obtain high accuracies when processing long sequences since they avoid the vanishing gradient problem and use the attention mechanism to maintain the focus on the information that matters. These features are fostering the use of transformers in other domains beyond NLP. This paper employs a systematic protocol to identify and analyze studies that propose new transformers’ architectures for processing longitudinal health datasets, which are often dense, and specifically focused on physiological, symptoms, functioning, and other daily life data. Our analysis considered 21 of 456 initial papers, collecting evidence to characterize how recent studies modified or extended these architectures to handle longitudinal multifeatured health representations or provide better ways to generate outcomes. Our findings suggest, for example, that the main efforts are focused on methods to integrate multiple vocabularies, encode input data, and represent temporal notions among longitudinal dependencies. We comprehensively discuss these and other findings, addressing major issues that are still open to efficiently deploy transformers architectures for longitudinal multifeatured healthcare data analysis.
Publisher DOI
Journal
Artificial Intelligence Review
ISSN
0269-2821
Volume
57
Publisher
Springer
Submitter
Kurpicz-Briki, Mascha
Citation apa
Siebra, C. A., Kurpicz-Briki, M., & Wac, K. (2024). Transformers in health: a systematic review on architectures for longitudinal data analysis. In Artificial Intelligence Review (Vol. 57). Springer. https://doi.org/10.24451/dspace/11562
File(s)![Thumbnail Image]()
Loading...
open access
Name
Siebra2024b.pdf
Description
Version published
License
Attribution 4.0 International
Size
3.28 MB
Format
Adobe PDF
Checksum (MD5)
1b8abb2e9cc978e9aa20770038b3af95
