メインナビゲーションにスキップ 検索にスキップ メインコンテンツにスキップ

Temporal Attention for Robust Multiple Object Pose Tracking

研究成果: 書籍/レポート タイプへの寄稿会議への寄与

抄録

Estimating the pose of multiple objects has improved substantially since deep learning became widely used. However, the performance deteriorates when the objects are highly similar in appearance or when occlusions are present. This issue is usually addressed by leveraging temporal information that takes previous frames as priors to improve the robustness of estimation. Existing methods are either computationally expensive by using multiple frames, or are inefficiently integrated with ad hoc procedures. In this paper, we perform computationally efficient object association between two consecutive frames via attention through a video sequence. Furthermore, instead of heatmap-based approaches, we adopt a coordinate classification strategy that excludes post-processing, where the network is built in an end-to-end fashion. Experiments on real data show that our approach achieves state-of-the-art results on PoseTrack datasets.

本文言語英語
ホスト出版物のタイトルNeural Information Processing - 30th International Conference, ICONIP 2023, Proceedings
編集者Biao Luo, Long Cheng, Zheng-Guang Wu, Hongyi Li, Chaojie Li
出版社Springer Science and Business Media Deutschland GmbH
ページ551-561
ページ数11
ISBN(印刷版)9789819980697
DOI
出版ステータス出版済み - 2024
外部発表はい
イベント30th International Conference on Neural Information Processing, ICONIP 2023 - Changsha, 中国
継続期間: 20-11-202323-11-2023

出版物シリーズ

名前Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
14450 LNCS
ISSN(印刷版)0302-9743
ISSN(電子版)1611-3349

会議

会議30th International Conference on Neural Information Processing, ICONIP 2023
国/地域中国
CityChangsha
Period20-11-2323-11-23

All Science Journal Classification (ASJC) codes

  • 理論的コンピュータサイエンス
  • コンピュータサイエンス一般

フィンガープリント

「Temporal Attention for Robust Multiple Object Pose Tracking」の研究トピックを掘り下げます。これらがまとまってユニークなフィンガープリントを構成します。

引用スタイル