Home / Journals / CMC / Online First / doi:10.32604/cmc.2025.059006
Special Issues
Table of Content

Open Access

ARTICLE

Token Masked Pose Transformers Are Efficient Learners

Xinyi Song1, Haixiang Zhang1,*, Shaohua Li2
1 School of Computer Science and Technology, Zhejiang Sci-Tech University, Hangzhou, 310018, China
2 College of Artificial Intelligence, Nankai University, Tianjin, 300350, China
* Corresponding Author: Haixiang Zhang. Email: email
(This article belongs to the Special Issue: Computer Vision and Image Processing: Feature Selection, Image Enhancement and Recognition)

Computers, Materials & Continua https://doi.org/10.32604/cmc.2025.059006

Received 25 September 2024; Accepted 10 January 2025; Published online 17 March 2025

Abstract

In recent years, Transformer has achieved remarkable results in the field of computer vision, with its built-in attention layers effectively modeling global dependencies in images by transforming image features into token forms. However, Transformers often face high computational costs when processing large-scale image data, which limits their feasibility in real-time applications. To address this issue, we propose Token Masked Pose Transformers (TMPose), constructing an efficient Transformer network for pose estimation. This network applies semantic-level masking to tokens and employs three different masking strategies to optimize model performance, aiming to reduce computational complexity. Experimental results show that TMPose reduces computational complexity by 61.1% on the COCO validation dataset, with negligible loss in accuracy. Additionally, our performance on the MPII dataset is also competitive. This research not only enhances the accuracy of pose estimation but also significantly reduces the demand for computational resources, providing new directions for further studies in this field. Code is available at: (accessed on 9 January 2025).

Keywords

Pattern recognition; image processing; neural network; pose transformer
  • 153

    View

  • 54

    Download

  • 0

    Like

Share Link