EAAI Journal 2026 Journal Article
Exercise quality assessment in monocular video streaming
- Yongchang Zhang
- Boxuan Xu
- Zhaowen Lin
- Junjie Li
- Anlong Ming
The recent proliferation of home-based exercise content has garnered significant attention. This has led to an increasing demand for Artificial Intelligence (AI) devices capable of automatically assessing exercise quality and providing guidance. However, existing real-time exercise quality assessment algorithms require instructors and learners to share similar camera views. Furthermore, these methods often rely on pre-labeled data, support a limited number of exercise actions, and offer restricted feedback. Among the numerous videos where data is hard to pre-label, instructors and learners may have unrestricted camera views and inconsistent body shape, while instructors may demonstrate unpredicted actions. To address the aforementioned challenges, we propose a method for Exercise Quality Assessment in Monocular Videos (MV-EQA), which incorporates a Skeleton Mapping and View Aligning (SMVA) module, a Multi-Feature Dynamic Time Warping (MF-DTW) module, and online/offline Exercise Quality Assessment (EQA) modules. Specifically, SMVA utilizes a lightweight encoder–decoder network based on transformer architecture that effectively handles differences in view and skeleton between learners and instructors while preserving inherent variations in their movements; MF-DTW utilizes multiple body information for temporal alignment; online/offline EQA modules enable online feedback (scoring with visual comparison) and offline feedback (reviews with comments). Extensive experiments indicate the superiority of our approach over other methods in EQA tasks. The code is available at link.