AI Chat Paper
Note: Please note that the following content is generated by AMiner AI. SciOpen does not take any responsibility related to this content.
{{lang === 'zh_CN' ? '文章概述' : 'Summary'}}
{{lang === 'en_US' ? '中' : 'Eng'}}
Chat more with AI
PDF (15.4 MB)
Collect
Submit Manuscript AI Chat Paper
Show Outline
Outline
Show full outline
Hide outline
Outline
Show full outline
Hide outline
Open Access

UniCount: Mining Large-Scale Video Data for Universal Repetitive Action Counting

Big Data Institute, Central South University, Changsha 410083, China
School of Computer Science and Engineering, Central South University, Changsha 410083, China
Transsion Co., Ltd., Shanghai 200135, China
School of Electrical and Automation Engineering, Nanjing Normal Univeisity, Nanjing 210023, China
Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China
Show Author Information

Abstract

We introduce the Open Sequential Repetitive Action Counting (OSRAC) task, which aims to count all repetitions and locate transition boundaries of sequential actions from large-scale video data, without relying on predefined action categories. Unlike the Repetitive Action Counting (RAC) task that focuses on a single-action assumption, OSRAC handles diverse and alternating repetitive action sequences in real-world scenarios, which is fundamentally more challenging. To this end, we propose UniCount, a universal system capable of counting multiple sequential repetitive actions from video data. Specifically, UniCount designs three primary modules: the Universal Repetitive Pattern Learner (URPL) to capture general repetitive patterns in alternating actions, Temporal Action Boundary Discriminator (TABD) to locate the action transition boundaries, and Dual Density Map Estimator (DDME) to achieve action counting and repetition segmentation. We also design a novel actionness loss to improve the detection of action transitions. To support this task, we conduct in-depth data analysis on existing RAC datasets and construct several OSRAC benchmarks (i.e., MUCFRep, MRepCount, and MInfiniteRep) by developing a pipeline on data processing and mining. We further perform comprehensive experiments to evaluate the effectiveness of UniCount. On MInfiniteRep, UniCount substantially improves the Off-By-One Accuracy (OBOA) from 0.39 to 0.78 and decreases the Mean Absolute Error (MAE) from 0.29 to 0.14 compared to counterparts. UniCount also achieves superior performance in open-set data, showcasing its universality.

References

【1】
【1】
 
 
Big Data Mining and Analytics
Pages 1112-1126

{{item.num}}

Comments on this article

Go to comment

< Back to all reports

Review Status: {{reviewData.commendedNum}} Commended , {{reviewData.revisionRequiredNum}} Revision Required , {{reviewData.notCommendedNum}} Not Commended Under Peer Review

Review Comment

Close
Close
Cite this article:
Tang Y, Zhang D, Luo W, et al. UniCount: Mining Large-Scale Video Data for Universal Repetitive Action Counting. Big Data Mining and Analytics, 2025, 8(5): 1112-1126. https://doi.org/10.26599/BDMA.2025.9020017

1211

Views

127

Downloads

0

Crossref

0

Web of Science

0

Scopus

0

CSCD

Received: 19 December 2024
Revised: 24 January 2025
Accepted: 08 February 2025
Published: 14 July 2025
© The author(s) 2025.

The articles published in this open access journal are distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/).