MTS-LOF: Medical Time-Series Representation Learning via Occlusion-Invariant Features
Affiliation
Department of Electrical & Computer Engineering, University of ArizonaBIO5 Institute, The University of Arizona
Issue Date
2024-03-05Keywords
Brain modelingComputer vision
health monitoring
masked autoencoder
Medical diagnostic imaging
Medical services
Medical time series
representation learning
Representation learning
self-supervised learning
Self-supervised learning
Time series analysis
time series classification
transformer
Metadata
Show full item recordCitation
H. Li, A. S. Carreon-Rascon, X. Chen, G. Yuan and A. Li, "MTS-LOF: Medical Time-Series Representation Learning via Occlusion-Invariant Features," in IEEE Journal of Biomedical and Health Informatics, doi: 10.1109/JBHI.2024.3373439.Rights
@ 2024 IEEE.Collection Information
This item from the UA Faculty Publications collection is made available by the University of Arizona with support from the University of Arizona Libraries. If you have questions, please contact us at repository@u.library.arizona.edu.Abstract
Medical time series data are indispensable in healthcare, providing critical insights for disease diagnosis, treatment planning, and patient management. The exponential growth in data complexity, driven by advanced sensor technologies, has presented challenges related to data labeling. Self-supervised learning (SSL) has emerged as a transformative approach to address these challenges, eliminating the need for extensive human annotation. In this study, we introduce a novel framework for Medical Time Series Representation Learning, known as MTS-LOF. MTS-LOF leverages the strengths of Joint-Embedding SSL and Masked Autoencoder (MAE) methods, offering a unique approach to representation learning for medical time series data. By combining these techniques, MTS-LOF enhances the potential of healthcare applications by providing more sophisticated, context-rich representations. Additionally, MTS-LOF employs a multi-masking strategy to facilitate occlusion-invariant feature learning. This approach allows the model to create multiple views of the data by masking portions of it. By minimizing the discrepancy between the representations of these masked patches and the fully visible patches, MTS-LOF learns to capture rich contextual information within medical time series datasets. The results of experiments conducted on diverse medical time series datasets demonstrate the superiority of MTS-LOF over other methods. These findings hold promise for significantly enhancing healthcare applications by improving representation learning. Furthermore, our work delves into the integration of Joint-Embedding SSL and MAE techniques, shedding light on the intricate interplay between temporal and structural dependencies in healthcare data. This understanding is crucial, as it allows us to grasp the complexities of healthcare data analysis.Note
Immediate accessEISSN
2168-2208PubMed ID
38442055Version
Final accepted manuscriptae974a485f413a2113503eed53cd6c53
10.1109/JBHI.2024.3373439
Scopus Count
Collections
Related articles
- A Generic Self-Supervised Framework of Learning Invariant Discriminative Features.
- Authors: Ntelemis F, Jin Y, Thomas SA
- Issue date: 2024 Sep
- Enhancing diagnostic deep learning via self-supervised pretraining on large-scale, unlabeled non-medical images.
- Authors: Tayebi Arasteh S, Misera L, Kather JN, Truhn D, Nebelung S
- Issue date: 2024 Feb 8
- Self-Supervised Contrastive Representation Learning for Semi-Supervised Time-Series Classification.
- Authors: Eldele E, Ragab M, Chen Z, Wu M, Kwoh CK, Li X, Guan C
- Issue date: 2023 Dec
- Learning the heterogeneous representation of brain's structure from serial SEM images using a masked autoencoder.
- Authors: Cheng A, Shi J, Wang L, Zhang R
- Issue date: 2023
- Voxel-wise adversarial semi-supervised learning for medical image segmentation.
- Authors: Lee CE, Park H, Shin YG, Chung M
- Issue date: 2022 Nov