Handy: Towards a High Fidelity 3D Hand Shape and Appearance Model

Over the last few years, with the advent of virtual and augmented reality, an enormous amount of research has been focused on modeling, tracking and reconstructing human hands. Given their power to express human behavior, hands have been a very important, but challenging component of the human body....

Full description

Saved in:
Bibliographic Details
Published inProceedings (IEEE Computer Society Conference on Computer Vision and Pattern Recognition. Online) pp. 4670 - 4680
Main Authors Potamias, Rolandos Alexandros, Ploumpis, Stylianos, Moschoglou, Stylianos, Triantafyllou, Vasileios, Zafeiriou, Stefanos
Format Conference Proceeding
LanguageEnglish
Published IEEE 01.06.2023
Subjects
Online AccessGet full text
ISSN1063-6919
DOI10.1109/CVPR52729.2023.00453

Cover

Abstract Over the last few years, with the advent of virtual and augmented reality, an enormous amount of research has been focused on modeling, tracking and reconstructing human hands. Given their power to express human behavior, hands have been a very important, but challenging component of the human body. Currently, most of the state-of-the-art reconstruction and pose estimation methods rely on the low polygon MANO model. Apart from its low polygon count, MANO model was trained with only 31 adult subjects, which not only limits its expressive power but also imposes unnecessary shape reconstruction constraints on pose estimation methods. Moreover, hand appearance remains almost unexplored and neglected from the majority of hand reconstruction methods. In this work, we propose "Handy", a large-scale model of the human hand, modeling both shape and appearance composed of over 1200 subjects which we make publicly available for the benefit of the research community. In contrast to current models, our proposed hand model was trained on a dataset with large diversity in age, gender, and ethnicity, which tackles the limitations of MANO and accurately reconstructs out-of-distribution samples. In order to create a high quality texture model, we trained a powerful GAN, which preserves high frequency details and is able to generate high resolution hand textures. To showcase the capabilities of the proposed model, we built a synthetic dataset of textured hands and trained a hand pose estimation network to reconstruct both the shape and appearance from single images. As it is demonstrated in an extensive series of quantitative as well as qualitative experiments, our model proves to be robust against the state-of-the-art and realistically captures the 3D hand shape and pose along with a high frequency detailed texture even in adverse "in-the-wild" conditions.
AbstractList Over the last few years, with the advent of virtual and augmented reality, an enormous amount of research has been focused on modeling, tracking and reconstructing human hands. Given their power to express human behavior, hands have been a very important, but challenging component of the human body. Currently, most of the state-of-the-art reconstruction and pose estimation methods rely on the low polygon MANO model. Apart from its low polygon count, MANO model was trained with only 31 adult subjects, which not only limits its expressive power but also imposes unnecessary shape reconstruction constraints on pose estimation methods. Moreover, hand appearance remains almost unexplored and neglected from the majority of hand reconstruction methods. In this work, we propose "Handy", a large-scale model of the human hand, modeling both shape and appearance composed of over 1200 subjects which we make publicly available for the benefit of the research community. In contrast to current models, our proposed hand model was trained on a dataset with large diversity in age, gender, and ethnicity, which tackles the limitations of MANO and accurately reconstructs out-of-distribution samples. In order to create a high quality texture model, we trained a powerful GAN, which preserves high frequency details and is able to generate high resolution hand textures. To showcase the capabilities of the proposed model, we built a synthetic dataset of textured hands and trained a hand pose estimation network to reconstruct both the shape and appearance from single images. As it is demonstrated in an extensive series of quantitative as well as qualitative experiments, our model proves to be robust against the state-of-the-art and realistically captures the 3D hand shape and pose along with a high frequency detailed texture even in adverse "in-the-wild" conditions.
Author Ploumpis, Stylianos
Moschoglou, Stylianos
Potamias, Rolandos Alexandros
Zafeiriou, Stefanos
Triantafyllou, Vasileios
Author_xml – sequence: 1
  givenname: Rolandos Alexandros
  surname: Potamias
  fullname: Potamias, Rolandos Alexandros
  email: r.potamias@imperial.ac.uk
  organization: Imperial College London
– sequence: 2
  givenname: Stylianos
  surname: Ploumpis
  fullname: Ploumpis, Stylianos
  email: s.ploumpis@imperial.ac.uk
  organization: Imperial College London
– sequence: 3
  givenname: Stylianos
  surname: Moschoglou
  fullname: Moschoglou, Stylianos
  email: s.moschoglou@imperial.ac.uk
  organization: Imperial College London
– sequence: 4
  givenname: Vasileios
  surname: Triantafyllou
  fullname: Triantafyllou, Vasileios
  organization: Cosmos Designs Ltd
– sequence: 5
  givenname: Stefanos
  surname: Zafeiriou
  fullname: Zafeiriou, Stefanos
  email: s.zafeiriou@imperial.ac.uk
  organization: Imperial College London
BookMark eNotjs1Kw0AURkdRsNa-QRfzAql37vxl3IVojVBRtLgtN5lbG6lJSAqSt7dFV-dbHD7Otbho2oaFmCtYKAXhNv94fbPoMSwQUC8AjNVnYhZ8SLUFDQpDei4mCpxOXFDhSsyG4QsANCrlQjoRWUFNHO_kuv2hPg6SZFF_7uSyjryvD6PU9_JkyPcddSxPK-s6pp6aiuVze7RuxOWW9gPP_jkV6-XDOi-S1cvjU56tkhrBHBKPjKxdrComwIqPDcAxxG1KZSyNKUvyDKnyW-tRV2TLYNB44ypyqfV6KuZ_tzUzb7q-_qZ-3ChA0MY6_QtdjEtQ
CODEN IEEPAD
ContentType Conference Proceeding
DBID 6IE
6IH
CBEJK
RIE
RIO
DOI 10.1109/CVPR52729.2023.00453
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Proceedings Order Plan (POP) 1998-present by volume
IEEE Xplore All Conference Proceedings
IEEE Xplore Digital Libary (IEL)
IEEE Proceedings Order Plans (POP) 1998-present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
Discipline Applied Sciences
EISBN 9798350301298
EISSN 1063-6919
EndPage 4680
ExternalDocumentID 10203456
Genre orig-research
GrantInformation_xml – fundername: EPSRC
  grantid: EP/S010203/1
  funderid: 10.13039/501100000266
GroupedDBID 6IE
6IH
6IL
6IN
AAWTH
ABLEC
ADZIZ
ALMA_UNASSIGNED_HOLDINGS
BEFXN
BFFAM
BGNUA
BKEBE
BPEOZ
CBEJK
CHZPO
IEGSK
IJVOP
OCL
RIE
RIL
RIO
ID FETCH-LOGICAL-i204t-72e2e36dccea02ce0030ed9df8abdb44bba7e0817f5723ca5b9424746ca68573
IEDL.DBID RIE
IngestDate Wed Aug 27 02:56:32 EDT 2025
IsPeerReviewed false
IsScholarly true
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-i204t-72e2e36dccea02ce0030ed9df8abdb44bba7e0817f5723ca5b9424746ca68573
PageCount 11
ParticipantIDs ieee_primary_10203456
PublicationCentury 2000
PublicationDate 2023-June
PublicationDateYYYYMMDD 2023-06-01
PublicationDate_xml – month: 06
  year: 2023
  text: 2023-June
PublicationDecade 2020
PublicationTitle Proceedings (IEEE Computer Society Conference on Computer Vision and Pattern Recognition. Online)
PublicationTitleAbbrev CVPR
PublicationYear 2023
Publisher IEEE
Publisher_xml – name: IEEE
SSID ssj0003211698
Score 2.28038
Snippet Over the last few years, with the advent of virtual and augmented reality, an enormous amount of research has been focused on modeling, tracking and...
SourceID ieee
SourceType Publisher
StartPage 4670
SubjectTerms body
Generative adversarial networks
gesture
High frequency
Humans: Face
movement
pose
Pose estimation
Reconstruction algorithms
Shape
Solid modeling
Three-dimensional displays
Title Handy: Towards a High Fidelity 3D Hand Shape and Appearance Model
URI https://ieeexplore.ieee.org/document/10203456
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3NS8MwFA9uJ0_zY-I3OXhtbZM0Sb3JdAzBMXTKbiMfLyhKN7buoH-9SdtNFARvoSRpSPJ4vyS_33sIXUinpHEpiVxKRcQssZECriMniUlsngFNg8D5fsgHT-xukk0asXqlhQGAinwGcShWb_l2ZlbhqsxbOEmo9_gt1PL7rBZrbS5UqD_K8Fw28rg0yS97z6OHjHj0GIcc4XFAL_RHEpXKh_Q7aLj-e00deYtXpY7N56_AjP8e3g7qfsv18GjjiHbRFhR7qNPgS9xY73IfXQ9UYT-u8Liiyi6xwoHlgfsh0pUH45je4FADP76oOeBQ8n14SwgbA4ekae9dNO7fjnuDqEmhEL2ShJWRIECAcmsMqIQYCDYNNrdOKm01Y1orAR4VCJcJQo3KdM4IE4wbxWUm6AFqF7MCDhGmShJtqTXgmzHLcm2cS8EfnzIgktMj1A0zMp3XQTKm68k4_uP7CdoOq1Kzrk5Ru1ys4Mz791KfV-v6BWl-pAA
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1dS8MwFA06H_Rpfkz8Ng--trZJ2qa-yXRM3cbQKnsb-bhBUbrhugf99SZtN1EQfAulTUuSyzlJz7kXoTNuBFcmJJ4JaeIxTbQnIJae4UQFOo2Ahs7g3B_E3Ud2O4pGtVm99MIAQCk-A981y3_5eqLm7qjMRjgJqEX8VbRmgZ9FlV1reaRC7WYmTnltkAuD9Lz9NLyPiOWPvqsS7jv-Qn-UUSlRpNNEg8X7K_HIqz8vpK8-f6Vm_PcHbqLWt2EPD5dQtIVWIN9GzZph4jp-Zzvosity_XGBs1IsO8MCO50H7rhcV5aOY3qF3R344VlMAbuW7cPGglsa2JVNe2uhrHOdtbteXUTBeyEBK7yEAAEaa6VABESBi2rQqTZcSC0Zk1IkYHlBYqKEUCUimTLCEhYrEfMoobuokU9y2EOYCk6kplqBfYxplkplTAh2AxUB4THdRy03IuNplSZjvBiMgz-un6L1btbvjXs3g7tDtOFmqNJgHaFG8T6HY4v2hTwp5_gLM6SnTQ
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=Proceedings+%28IEEE+Computer+Society+Conference+on+Computer+Vision+and+Pattern+Recognition.+Online%29&rft.atitle=Handy%3A+Towards+a+High+Fidelity+3D+Hand+Shape+and+Appearance+Model&rft.au=Potamias%2C+Rolandos+Alexandros&rft.au=Ploumpis%2C+Stylianos&rft.au=Moschoglou%2C+Stylianos&rft.au=Triantafyllou%2C+Vasileios&rft.date=2023-06-01&rft.pub=IEEE&rft.eissn=1063-6919&rft.spage=4670&rft.epage=4680&rft_id=info:doi/10.1109%2FCVPR52729.2023.00453&rft.externalDocID=10203456