Handy: Towards a High Fidelity 3D Hand Shape and Appearance Model
Over the last few years, with the advent of virtual and augmented reality, an enormous amount of research has been focused on modeling, tracking and reconstructing human hands. Given their power to express human behavior, hands have been a very important, but challenging component of the human body....
Saved in:
Published in | Proceedings (IEEE Computer Society Conference on Computer Vision and Pattern Recognition. Online) pp. 4670 - 4680 |
---|---|
Main Authors | , , , , |
Format | Conference Proceeding |
Language | English |
Published |
IEEE
01.06.2023
|
Subjects | |
Online Access | Get full text |
ISSN | 1063-6919 |
DOI | 10.1109/CVPR52729.2023.00453 |
Cover
Abstract | Over the last few years, with the advent of virtual and augmented reality, an enormous amount of research has been focused on modeling, tracking and reconstructing human hands. Given their power to express human behavior, hands have been a very important, but challenging component of the human body. Currently, most of the state-of-the-art reconstruction and pose estimation methods rely on the low polygon MANO model. Apart from its low polygon count, MANO model was trained with only 31 adult subjects, which not only limits its expressive power but also imposes unnecessary shape reconstruction constraints on pose estimation methods. Moreover, hand appearance remains almost unexplored and neglected from the majority of hand reconstruction methods. In this work, we propose "Handy", a large-scale model of the human hand, modeling both shape and appearance composed of over 1200 subjects which we make publicly available for the benefit of the research community. In contrast to current models, our proposed hand model was trained on a dataset with large diversity in age, gender, and ethnicity, which tackles the limitations of MANO and accurately reconstructs out-of-distribution samples. In order to create a high quality texture model, we trained a powerful GAN, which preserves high frequency details and is able to generate high resolution hand textures. To showcase the capabilities of the proposed model, we built a synthetic dataset of textured hands and trained a hand pose estimation network to reconstruct both the shape and appearance from single images. As it is demonstrated in an extensive series of quantitative as well as qualitative experiments, our model proves to be robust against the state-of-the-art and realistically captures the 3D hand shape and pose along with a high frequency detailed texture even in adverse "in-the-wild" conditions. |
---|---|
AbstractList | Over the last few years, with the advent of virtual and augmented reality, an enormous amount of research has been focused on modeling, tracking and reconstructing human hands. Given their power to express human behavior, hands have been a very important, but challenging component of the human body. Currently, most of the state-of-the-art reconstruction and pose estimation methods rely on the low polygon MANO model. Apart from its low polygon count, MANO model was trained with only 31 adult subjects, which not only limits its expressive power but also imposes unnecessary shape reconstruction constraints on pose estimation methods. Moreover, hand appearance remains almost unexplored and neglected from the majority of hand reconstruction methods. In this work, we propose "Handy", a large-scale model of the human hand, modeling both shape and appearance composed of over 1200 subjects which we make publicly available for the benefit of the research community. In contrast to current models, our proposed hand model was trained on a dataset with large diversity in age, gender, and ethnicity, which tackles the limitations of MANO and accurately reconstructs out-of-distribution samples. In order to create a high quality texture model, we trained a powerful GAN, which preserves high frequency details and is able to generate high resolution hand textures. To showcase the capabilities of the proposed model, we built a synthetic dataset of textured hands and trained a hand pose estimation network to reconstruct both the shape and appearance from single images. As it is demonstrated in an extensive series of quantitative as well as qualitative experiments, our model proves to be robust against the state-of-the-art and realistically captures the 3D hand shape and pose along with a high frequency detailed texture even in adverse "in-the-wild" conditions. |
Author | Ploumpis, Stylianos Moschoglou, Stylianos Potamias, Rolandos Alexandros Zafeiriou, Stefanos Triantafyllou, Vasileios |
Author_xml | – sequence: 1 givenname: Rolandos Alexandros surname: Potamias fullname: Potamias, Rolandos Alexandros email: r.potamias@imperial.ac.uk organization: Imperial College London – sequence: 2 givenname: Stylianos surname: Ploumpis fullname: Ploumpis, Stylianos email: s.ploumpis@imperial.ac.uk organization: Imperial College London – sequence: 3 givenname: Stylianos surname: Moschoglou fullname: Moschoglou, Stylianos email: s.moschoglou@imperial.ac.uk organization: Imperial College London – sequence: 4 givenname: Vasileios surname: Triantafyllou fullname: Triantafyllou, Vasileios organization: Cosmos Designs Ltd – sequence: 5 givenname: Stefanos surname: Zafeiriou fullname: Zafeiriou, Stefanos email: s.zafeiriou@imperial.ac.uk organization: Imperial College London |
BookMark | eNotjs1Kw0AURkdRsNa-QRfzAql37vxl3IVojVBRtLgtN5lbG6lJSAqSt7dFV-dbHD7Otbho2oaFmCtYKAXhNv94fbPoMSwQUC8AjNVnYhZ8SLUFDQpDei4mCpxOXFDhSsyG4QsANCrlQjoRWUFNHO_kuv2hPg6SZFF_7uSyjryvD6PU9_JkyPcddSxPK-s6pp6aiuVze7RuxOWW9gPP_jkV6-XDOi-S1cvjU56tkhrBHBKPjKxdrComwIqPDcAxxG1KZSyNKUvyDKnyW-tRV2TLYNB44ypyqfV6KuZ_tzUzb7q-_qZ-3ChA0MY6_QtdjEtQ |
CODEN | IEEPAD |
ContentType | Conference Proceeding |
DBID | 6IE 6IH CBEJK RIE RIO |
DOI | 10.1109/CVPR52729.2023.00453 |
DatabaseName | IEEE Electronic Library (IEL) Conference Proceedings IEEE Proceedings Order Plan (POP) 1998-present by volume IEEE Xplore All Conference Proceedings IEEE Xplore Digital Libary (IEL) IEEE Proceedings Order Plans (POP) 1998-present |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: RIE name: IEEE Electronic Library (IEL) url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/ sourceTypes: Publisher |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Applied Sciences |
EISBN | 9798350301298 |
EISSN | 1063-6919 |
EndPage | 4680 |
ExternalDocumentID | 10203456 |
Genre | orig-research |
GrantInformation_xml | – fundername: EPSRC grantid: EP/S010203/1 funderid: 10.13039/501100000266 |
GroupedDBID | 6IE 6IH 6IL 6IN AAWTH ABLEC ADZIZ ALMA_UNASSIGNED_HOLDINGS BEFXN BFFAM BGNUA BKEBE BPEOZ CBEJK CHZPO IEGSK IJVOP OCL RIE RIL RIO |
ID | FETCH-LOGICAL-i204t-72e2e36dccea02ce0030ed9df8abdb44bba7e0817f5723ca5b9424746ca68573 |
IEDL.DBID | RIE |
IngestDate | Wed Aug 27 02:56:32 EDT 2025 |
IsPeerReviewed | false |
IsScholarly | true |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-i204t-72e2e36dccea02ce0030ed9df8abdb44bba7e0817f5723ca5b9424746ca68573 |
PageCount | 11 |
ParticipantIDs | ieee_primary_10203456 |
PublicationCentury | 2000 |
PublicationDate | 2023-June |
PublicationDateYYYYMMDD | 2023-06-01 |
PublicationDate_xml | – month: 06 year: 2023 text: 2023-June |
PublicationDecade | 2020 |
PublicationTitle | Proceedings (IEEE Computer Society Conference on Computer Vision and Pattern Recognition. Online) |
PublicationTitleAbbrev | CVPR |
PublicationYear | 2023 |
Publisher | IEEE |
Publisher_xml | – name: IEEE |
SSID | ssj0003211698 |
Score | 2.28038 |
Snippet | Over the last few years, with the advent of virtual and augmented reality, an enormous amount of research has been focused on modeling, tracking and... |
SourceID | ieee |
SourceType | Publisher |
StartPage | 4670 |
SubjectTerms | body Generative adversarial networks gesture High frequency Humans: Face movement pose Pose estimation Reconstruction algorithms Shape Solid modeling Three-dimensional displays |
Title | Handy: Towards a High Fidelity 3D Hand Shape and Appearance Model |
URI | https://ieeexplore.ieee.org/document/10203456 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3NS8MwFA9uJ0_zY-I3OXhtbZM0Sb3JdAzBMXTKbiMfLyhKN7buoH-9SdtNFARvoSRpSPJ4vyS_33sIXUinpHEpiVxKRcQssZECriMniUlsngFNg8D5fsgHT-xukk0asXqlhQGAinwGcShWb_l2ZlbhqsxbOEmo9_gt1PL7rBZrbS5UqD_K8Fw28rg0yS97z6OHjHj0GIcc4XFAL_RHEpXKh_Q7aLj-e00deYtXpY7N56_AjP8e3g7qfsv18GjjiHbRFhR7qNPgS9xY73IfXQ9UYT-u8Liiyi6xwoHlgfsh0pUH45je4FADP76oOeBQ8n14SwgbA4ekae9dNO7fjnuDqEmhEL2ShJWRIECAcmsMqIQYCDYNNrdOKm01Y1orAR4VCJcJQo3KdM4IE4wbxWUm6AFqF7MCDhGmShJtqTXgmzHLcm2cS8EfnzIgktMj1A0zMp3XQTKm68k4_uP7CdoOq1Kzrk5Ru1ys4Mz791KfV-v6BWl-pAA |
linkProvider | IEEE |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1dS8MwFA06H_Rpfkz8Ng--trZJ2qa-yXRM3cbQKnsb-bhBUbrhugf99SZtN1EQfAulTUuSyzlJz7kXoTNuBFcmJJ4JaeIxTbQnIJae4UQFOo2Ahs7g3B_E3Ud2O4pGtVm99MIAQCk-A981y3_5eqLm7qjMRjgJqEX8VbRmgZ9FlV1reaRC7WYmTnltkAuD9Lz9NLyPiOWPvqsS7jv-Qn-UUSlRpNNEg8X7K_HIqz8vpK8-f6Vm_PcHbqLWt2EPD5dQtIVWIN9GzZph4jp-Zzvosity_XGBs1IsO8MCO50H7rhcV5aOY3qF3R344VlMAbuW7cPGglsa2JVNe2uhrHOdtbteXUTBeyEBK7yEAAEaa6VABESBi2rQqTZcSC0Zk1IkYHlBYqKEUCUimTLCEhYrEfMoobuokU9y2EOYCk6kplqBfYxplkplTAh2AxUB4THdRy03IuNplSZjvBiMgz-un6L1btbvjXs3g7tDtOFmqNJgHaFG8T6HY4v2hTwp5_gLM6SnTQ |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=Proceedings+%28IEEE+Computer+Society+Conference+on+Computer+Vision+and+Pattern+Recognition.+Online%29&rft.atitle=Handy%3A+Towards+a+High+Fidelity+3D+Hand+Shape+and+Appearance+Model&rft.au=Potamias%2C+Rolandos+Alexandros&rft.au=Ploumpis%2C+Stylianos&rft.au=Moschoglou%2C+Stylianos&rft.au=Triantafyllou%2C+Vasileios&rft.date=2023-06-01&rft.pub=IEEE&rft.eissn=1063-6919&rft.spage=4670&rft.epage=4680&rft_id=info:doi/10.1109%2FCVPR52729.2023.00453&rft.externalDocID=10203456 |