TY - GEN
T1 - Pixel-REfocused Navigated Tri-margin for Semi-Supervised Action Detection
AU - Liu, Wenxuan
AU - Zhao, Shilei
AU - Han, Xiyu
AU - Yi, Aoyu
AU - Jiang, Kui
AU - Wang, Zheng
AU - Zhong, Xian
N1 - Publisher Copyright:
© 2024 Copyright held by the owner/author(s).
PY - 2024/10/28
Y1 - 2024/10/28
N2 - This paper identifies a novel issue, termed pixel activation-uncertainty, in semi-supervised action detection, which highlights the difficulty in distinguishing between action and background boundaries due to active motion. To address this, we propose an effective pipeline called Pixel-Refocused Navigated Tri-margin (PRENT), which adaptively leverages class-explicit knowledge. PRENT emphasizes maintaining region consistency by updating pseudo-label selection with each training epoch, ensuring continuous improvement. We introduce a class-explicit tri-margin as a soft solution to manage uncertain boundaries within latent buffer regions. This technique refines the buffer zone based on the unique characteristics of each category, thereby addressing the varying challenges in localizing actions and backgrounds. Experimental results on various benchmarks and training settings demonstrate the superiority of our method compared to state-of-the-art methods.
AB - This paper identifies a novel issue, termed pixel activation-uncertainty, in semi-supervised action detection, which highlights the difficulty in distinguishing between action and background boundaries due to active motion. To address this, we propose an effective pipeline called Pixel-Refocused Navigated Tri-margin (PRENT), which adaptively leverages class-explicit knowledge. PRENT emphasizes maintaining region consistency by updating pseudo-label selection with each training epoch, ensuring continuous improvement. We introduce a class-explicit tri-margin as a soft solution to manage uncertain boundaries within latent buffer regions. This technique refines the buffer zone based on the unique characteristics of each category, thereby addressing the varying challenges in localizing actions and backgrounds. Experimental results on various benchmarks and training settings demonstrate the superiority of our method compared to state-of-the-art methods.
KW - class-explicit tri-margin
KW - pixel activation-uncertainty
KW - region consistency
KW - semi-supervised action detection
UR - https://www.scopus.com/pages/publications/85210872288
U2 - 10.1145/3688865.3689478
DO - 10.1145/3688865.3689478
M3 - 会议稿件
AN - SCOPUS:85210872288
T3 - HCMA 2024 - Proceedings of the 5th International Workshop on Human-centric Multimedia Analysis, Co-Located with: MM 2024
SP - 23
EP - 31
BT - HCMA 2024 - Proceedings of the 5th International Workshop on Human-centric Multimedia Analysis, Co-Located with
PB - Association for Computing Machinery, Inc
T2 - 5th International Workshop on Human-centric Multimedia Analysis, HCMA 2024
Y2 - 28 October 2024 through 1 November 2024
ER -