Logo BSU

Please use this identifier to cite or link to this item: https://elib.bsu.by/handle/123456789/306232
Title: Human Pose Estimation using SimCC and Swin Transformer
Authors: Li, Tongrui
Ablameyko, Sergey
Keywords: ЭБ БГУ::ЕСТЕСТВЕННЫЕ И ТОЧНЫЕ НАУКИ::Кибернетика
ЭБ БГУ::ЕСТЕСТВЕННЫЕ И ТОЧНЫЕ НАУКИ::Математика
Issue Date: 2023
Publisher: Minsk : BSU
Citation: Pattern Recognition and Information Processing (PRIP’2023). Artificial Universe: New Horisont : Proceedings of the 16 th International Conference, Belarus, Minsk, October 17–19, 2023 / Belarusian State University : eds. A. Nedzved, A. Belotserkovsky. – Minsk : BSU, 2023. – Pp. 197-201.
Abstract: 2D Human Pose Estimation is an important task in computer vision. In recent years, methods using deep learning for human pose estimation have been proposed one after another and achieved good results. Among existing models, the built-in attention layer in Transformer enables the model to effectively capture long-range relationships and also reveal the dependencies on which predicted key points depend. SimCC formulates keypoint localization as a classification problem, dividing the horizontal and vertical axes into equal-width numbered bins, and discretizing continuous coordinates into integer bin labels. We propose a new model that combines the Swin Transformer training model to predict the bin where the key points are located, so as to achieve the purpose of predicting key points. This method can achieve better results than other models and can achieve sub-pixel positioning accuracy and low quantization error
URI: https://elib.bsu.by/handle/123456789/306232
ISBN: 978-985-881-522-6
Sponsorship: This paper was funded by the China Scholarship Council.
Licence: info:eu-repo/semantics/openAccess
Appears in Collections:2023. Pattern Recognition and Information Processing (PRIP’2023). Artificial Intelliverse: Expanding Horizons

Files in This Item:
File Description SizeFormat 
197-201.pdf6,4 MBAdobe PDFView/Open
Show full item record Google Scholar



Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.