Szczegóły publikacji

Opis bibliograficzny

Pose guided feature learning for 3D object tracking on RGB videos / Mateusz MAJCHER, Bogdan KWOLEK // W: VISIGRAPP 2022 [Dokument elektroniczny] : 17th international joint conference on Computer Vision, imaging and computer graphics theory and applications : online streming, 6–8 February, 2022. Vol. 5, VISAPP / eds. Giovanni Maria Farinella, Petia Radeva, Kadi Bouatouch. — Wersja do Windows. — Dane tekstowe. — [Portugal] : SciTePress Digital Library, [2022]. — (VISIGRAPP ; ISSN 2184-5921). — e-ISBN: 978-989-758-555-5. — S. 574–581. — Wymagania systemowe: Adobe Reader. — Tryb dostępu: https://www.scitepress.org/PublicationsDetail.aspx?ID=sQ2Xb7o... [2022-05-04]. — Dostęp po zalogowaniu

Autorzy (2)

Słowa kluczowe

detection of object keypoints3D object posepose tracking

Dane bibliometryczne

ID BaDAP140060
Data dodania do BaDAP2022-09-19
DOI10.5220/0010886800003124
Rok publikacji2022
Typ publikacjimateriały konferencyjne (aut.)
Otwarty dostęptak
Creative Commons
KonferencjaJoint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications 2022
Czasopismo/seriaVISIGRAPP

Abstract

In this work we propose a new approach to 3D object pose tracking in sequences of RGB images acquired by a calibrated camera. A single hourglass neural network that has been trained to detect fiducial keypoints on a set of objects delivers heatmaps representing 2D locations of the keypoints. Given a calibrated camera model and a sparse object model consisting of 3D locations of the keypoints, the keypoints in hypothesized object poses are projected onto 2D plane and then matched with the heatmaps. A quaternion particle filter with a probabilistic observation model that uses such a matching is employed to maintain 3D object pose distribution. A single Siamese neural network is trained for a set of objects on keypoints from the current and previous frame in order to generate a particle in the predicted 3D object pose. The filter draws particles to predict the current pose using its a priori knowledge about the object velocity and includes the predicted 3D object pose by the neural network in a priori distribution. Thus, the hypothesized 3D object poses are generated using both a priori knowledge about the object velocity in 3D and keypoint-based geometric reasoning as well as relative transformations in the image plane. In an extended algorithm we combine the set of propagated particles with an optimized particle, whose pose is determined by Levenberg-Marguardt.

Publikacje, które mogą Cię zainteresować

fragment książki
#135376Data dodania: 24.9.2021
Fiducial points-supported object pose tracking on RGB images via particle filtering with heuristic optimization / Mateusz MAJCHER, Bogdan KWOLEK // W: VISIGRAPP 2021 [Dokument elektroniczny] : proceedings of the 16th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications. Vol. 5, VISAPP / eds. Giovanni Maria Farinella, [et al.]. — [Lisbon] : SCITEPRESS - Science and Technology Publications, [2021]. — (VISIGRAPP ; ISSN 2184-5921). — e-ISBN: 978-989-758-488-6. — S. 919–926. — Wymagania systemowe: Adobe Reader. — Tryb dostępu: https://www.scitepress.org/Link.aspx?doi=10.5220/001023710919... [2021-07-23]. — Bibliogr., Abstr. — Dostęp do pełnego tekstu po zalogowaniu
fragment książki
#128181Data dodania: 5.5.2020
3D model-based 6D object pose tracking on RGB images using particle filtering and heuristic optimization / Mateusz MAJCHER, Bogdan KWOLEK // W: VISIGRAPP 2020 [Dokument elektroniczny] : proceedings of the 15th international joint conference on Computer Vision, Imaging and Computer Graphics Theory and Applications. Vol. 5, VISAPP / eds. Giovanni Maria Farinella, Petia Radeva, Jose Braz. — Wersja do Windows. — Dane tekstowe. — [Lisbon] : SCITEPRESS - Science and Technology Publications, cop. 2020. — (VISIGRAPP ; ISSN 2184-5921). — e-ISBN: 978-989-758-402-2. — S. 690–697. — Wymagania systemowe: Adobe Reader. — Tryb dostępu: https://www.scitepress.org/PublicationsDetail.aspx?ID=Sc6i5xP... [2020-03-30]. — Bibliogr. s. 697, Abstr. — Dostęp po zalogowaniu