Contrastive Weakly Supervised Representation Learning for Robust and Annotation-Efficient Human Activity Recognition with Wearable Sensors
DOI:
https://doi.org/10.63891/j-mart.v2i1.132Keywords:
Human activity recognition, weakly supervised learning, wearable sensors, contrastive learning, self-supervised representation learning, annotation efficiency, robust deep learningAbstract
Wearable sensors and Human Activity Recognition (HAR) have turned into the technology of the foundation of healthcare monitoring and smart living environments, analytics in sports, and systems of rehabilitation. Although there are notable improvements in deep learning architecture time-series modeling, the overwhelming majority of the state-of-the-art HAR systems depend on the presence of fully annotated datasets, which are expensive, require human labor, and are not always available on large scale. Moreover, wearable sensor data is prone to noise, subject variability, domain shifts, and inconsistent labelling and thus, in real-world applications, wearable sensor data results in worse generalization performance. The current paper suggests a weakly supervised paradigm of representation learning that is contrastive, and it is expected to be more robust and much less reliant on annotation. By combining contrastive self-supervised tasks with weakly labeled input, the proposed method will learn both discriminative and invariant time-series representations using multivariate time-series signals produced by wearable devices. It has a framework that includes data augmentation policies, consistency regularization, pseudo-label refinement, and noise-sensitive training systems to reduce the effects of small and low-quality annotations. The experimental studies of the proposed method in different label scarcity and simulated noise levels show that the proposed approach has better performance than fully supervised and purely self-supervised baselines. The findings demonstrate significant improvements in the quality of representations, robustness and efficiency of annotations, and the way to scalable and economical systems of HAR. This paper can be used in the further development of annotation-efficient deep learning techniques in sensor-based smart systems.
References
[1] Sheng, T.; Huber, M. Weakly supervised multi-task representation learning for human activity analysis using wearables. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2020, 4, 57.
[2] Sheng, T.; Huber, M. Siamese Networks for Weakly Supervised Human Activity Recognition. In Proceedings of the 2019 IEEE International Conference on Systems, Man and Cybernetics (SMC), Bari, Italy, 6–9 October 2019.
[3] Azizi, S., Culp, L., Freyberg, J., Mustafa, B., Baur, S., Kornblith, S., ... & Natarajan, V. (2022). Robust and efficient medical imaging with self-supervision. arXiv preprint arXiv:2205.09723.
[4] Sheng, T.; Huber, M. Unsupervised Embedding Learning for Human Activity Recognition Using Wearable Sensor Data. In Proceedings of the FLAIRS Conference, Daytona Beach, FL, USA, 20–23 May 2020.
[5] Lin, C. H., Tseng, Y. H., Wu, P. C., Huang, C. Y., & Lai, M. Y. (2025). Self-supervised Fine-grained Image Recognition Method Based on Multi-scale Attention and Contrastive Learning. International Journal of Advance in Applied Science Research, 4(4), 1-8.
[6] Sheng, T.; Huber, M. Consistency Based Weakly Self-Supervised Learning for Human Activity Recognition with Wearables. In Proceedings of the AAAI-22 Workshop on 4. Human-Centric Self-Supervised Learning, Virtual, 22 February–1 March 2022. HC-SSL’22.
[7] Gupta, U., Paluru, N., Nankani, D., Kulkarni, K., & Awasthi, N. (2024). A comprehensive review on efficient artificial intelligence models for classification of abnormal cardiac rhythms using electrocardiograms. Heliyon, 10(5).
[8] Ding, W., Zhou, T., Huang, J., Jiang, S., Hou, T., & Lin, C. T. (2024). Fmdnn: A fuzzy-guided multigranular deep neural network for histopathological image classification. IEEE Transactions on Fuzzy Systems, 32(8), 4709-4723.
[9] Saleh, P. (2025). Advanced Unsupervised Analysis of Spatio-Temporal Pattern Analytics (Doctoral dissertation, Université d'Ottawa/University of Ottawa).
[10] He, Y., Huang, F., Jiang, X., Nie, Y., Wang, M., Wang, J., & Chen, H. (2024). Foundation model for advancing healthcare: challenges, opportunities and future directions. IEEE Reviews in Biomedical Engineering, 18, 172-191.
[11] Lia, Y., Chen, W., Zhang, F., Li, M., Wang, X., Liu, Z., ... & Zhao, J. (2025). Domain Shift and Generalization Strategies for Foundation Models in Clinical Imaging. Authorea Preprints.
[12] Marcinkevičs, R., Ozkan, E., & Vogt, J. E. (2022). Introduction to machine learning for physicians: a survival guide for data deluge. arXiv preprint arXiv:2212.12303.
[13] Vogt, J. E., Ozkan, E., & Marcinkeviĉs, R. (2023). Introduction to Machine Learning for Physicians: A Survival Guide for Data Deluge. In Digital Medicine (pp. 3-34). Jenny Stanford Publishing.
[14] Sheng, T.; Huber, M. Reducing Label Dependency in Human Activity Recognition with Wearables: From Supervised Learning to Novel Weakly Self-Supervised Approaches. Sensors 2025, 25, 4032.
Downloads
Published
Issue
Section
License
Copyright (c) 2026 Phillip Thomas

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.

