SOMA: Solving Optical Marker-Based MoCap Automatically
Marker-based optical motion capture (mocap) is the "gold standard" method for acquiring accurate 3D human motion in computer vision, medicine, and graphics. The raw output of these systems are noisy and incomplete 3D points or short tracklets of points. To be useful, one must associate the...
Saved in:
Published in | Proceedings / IEEE International Conference on Computer Vision pp. 11097 - 11106 |
---|---|
Main Authors | , |
Format | Conference Proceeding |
Language | English |
Published |
IEEE
01.10.2021
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | Marker-based optical motion capture (mocap) is the "gold standard" method for acquiring accurate 3D human motion in computer vision, medicine, and graphics. The raw output of these systems are noisy and incomplete 3D points or short tracklets of points. To be useful, one must associate these points with corresponding markers on the captured subject; i.e. "labelling". Given these labels, one can then "solve" for the 3D skeleton or body surface mesh. Commercial auto-labeling tools require a specific calibration procedure at capture time, which is not possible for archival data. Here we train a novel neural network called SOMA, which takes raw mocap point clouds with varying numbers of points, labels them at scale without any calibration data, independent of the capture technology, and requiring only minimal human intervention. Our key insight is that, while labeling point clouds is highly ambiguous, the 3D body provides strong constraints on the solution that can be exploited by a learning-based method. To enable learning, we generate massive training sets of simulated noisy and ground truth mocap markers animated by 3D bodies from AMASS. SOMA exploits an architecture with stacked self-attention elements to learn the spatial structure of the 3D body and an optimal transport layer to constrain the assignment (labeling) problem while rejecting outliers. We extensively evaluate SOMA both quantitatively and qualitatively. SOMA is more accurate and robust than existing state of the art research methods and can be applied where commercial systems cannot. We automatically label over 8 hours of archival mocap data across 4 different datasets captured using various technologies and output SMPL-X body models. The model and data is released for research purposes at https://soma.is.tue.mpg.de/. |
---|---|
AbstractList | Marker-based optical motion capture (mocap) is the "gold standard" method for acquiring accurate 3D human motion in computer vision, medicine, and graphics. The raw output of these systems are noisy and incomplete 3D points or short tracklets of points. To be useful, one must associate these points with corresponding markers on the captured subject; i.e. "labelling". Given these labels, one can then "solve" for the 3D skeleton or body surface mesh. Commercial auto-labeling tools require a specific calibration procedure at capture time, which is not possible for archival data. Here we train a novel neural network called SOMA, which takes raw mocap point clouds with varying numbers of points, labels them at scale without any calibration data, independent of the capture technology, and requiring only minimal human intervention. Our key insight is that, while labeling point clouds is highly ambiguous, the 3D body provides strong constraints on the solution that can be exploited by a learning-based method. To enable learning, we generate massive training sets of simulated noisy and ground truth mocap markers animated by 3D bodies from AMASS. SOMA exploits an architecture with stacked self-attention elements to learn the spatial structure of the 3D body and an optimal transport layer to constrain the assignment (labeling) problem while rejecting outliers. We extensively evaluate SOMA both quantitatively and qualitatively. SOMA is more accurate and robust than existing state of the art research methods and can be applied where commercial systems cannot. We automatically label over 8 hours of archival mocap data across 4 different datasets captured using various technologies and output SMPL-X body models. The model and data is released for research purposes at https://soma.is.tue.mpg.de/. |
Author | Ghorbani, Nima Black, Michael J. |
Author_xml | – sequence: 1 givenname: Nima surname: Ghorbani fullname: Ghorbani, Nima email: nghorbani@tuebingen.mpg.de organization: Max Planck Institute for Intelligent Systems,Tübingen,Germany – sequence: 2 givenname: Michael J. surname: Black fullname: Black, Michael J. email: black@tuebingen.mpg.de organization: Max Planck Institute for Intelligent Systems,Tübingen,Germany |
BookMark | eNotjtFKwzAUhqMouM09gV70BVpPTpIm8a6WqYOVXky9HUlzKtWuLe0U9vYO9eqH74OPf84uur4jxm45JJyDvVvn-Zs0FjFBQJ7AiYkztrTa8DRVEg1Hdc5mKAzEWoG8YvNp-gAQFk06Y-m2LLL7aNu33033HpXDoalcGxVu_KQxfnAThajoczdE2deh37tf3R6v2WXt2omW_7tgr4-rl_w53pRP6zzbxA0X4hBT5XXlrD99qGzgSknnlBHEwaMLaCQHChIrRaEmLbxHm0LQQWluPNZWLNjNX7chot0wNns3HndWcwAD4gdh8UeG |
CODEN | IEEPAD |
ContentType | Conference Proceeding |
DBID | 6IE 6IH CBEJK RIE RIO |
DOI | 10.1109/ICCV48922.2021.01093 |
DatabaseName | IEEE Electronic Library (IEL) Conference Proceedings IEEE Proceedings Order Plan (POP) 1998-present by volume IEEE Xplore All Conference Proceedings IEEE Electronic Library (IEL) IEEE Proceedings Order Plans (POP) 1998-present |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: RIE name: IEEE Electronic Library (IEL) url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/ sourceTypes: Publisher |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Applied Sciences |
EISBN | 9781665428125 1665428120 |
EISSN | 2380-7504 |
EndPage | 11106 |
ExternalDocumentID | 9710080 |
Genre | orig-research |
GroupedDBID | 29O 6IE 6IF 6IH 6IK 6IL 6IM 6IN AAJGR AAWTH ACGFS ADZIZ ALMA_UNASSIGNED_HOLDINGS BEFXN BFFAM BGNUA BKEBE BPEOZ CBEJK CHZPO IPLJI M43 OCL RIE RIL RIO RNS |
ID | FETCH-LOGICAL-i133t-ecb7ca9b125c9d1554aa583e10b2ad28410ed42c5edfe73bb2960d7d5718b2f93 |
IEDL.DBID | RIE |
IngestDate | Wed Aug 27 02:25:36 EDT 2025 |
IsPeerReviewed | false |
IsScholarly | true |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-i133t-ecb7ca9b125c9d1554aa583e10b2ad28410ed42c5edfe73bb2960d7d5718b2f93 |
PageCount | 10 |
ParticipantIDs | ieee_primary_9710080 |
PublicationCentury | 2000 |
PublicationDate | 2021-Oct. |
PublicationDateYYYYMMDD | 2021-10-01 |
PublicationDate_xml | – month: 10 year: 2021 text: 2021-Oct. |
PublicationDecade | 2020 |
PublicationTitle | Proceedings / IEEE International Conference on Computer Vision |
PublicationTitleAbbrev | ICCV |
PublicationYear | 2021 |
Publisher | IEEE |
Publisher_xml | – name: IEEE |
SSID | ssj0039286 |
Score | 2.271866 |
Snippet | Marker-based optical motion capture (mocap) is the "gold standard" method for acquiring accurate 3D human motion in computer vision, medicine, and graphics.... |
SourceID | ieee |
SourceType | Publisher |
StartPage | 11097 |
SubjectTerms | Computer vision Data models Datasets and evaluation Gestures and body pose Learning systems Motion and tracking Point cloud compression Three-dimensional displays Training Training data |
Title | SOMA: Solving Optical Marker-Based MoCap Automatically |
URI | https://ieeexplore.ieee.org/document/9710080 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV09T8MwELVKJ6YCLeJbHhhJ2jiOHbOVCFSQQpFKUbfKHxcJUTVVSQb49dhJWgRiYIsyxNGd7HfvfO8OoUseSUIZkV5MNFiCwjJPUSU8GnIgccaU5k7gnD6y0ZQ-zKJZC11ttTAAUBWfge8eq7t8k-vSpcr6ompFYwn6jiVutVZrc-pamI9ZI40LBqJ_nyQvNBbEaa1I4LsLoPDHAJUKP-46KN2sXJeNvPlloXz9-asp439_bQ_1vpV6-GmLQfuoBcsD1GlCS9xs3PcuYpNxOrzGk3zhEgh4vKpS2NgpdWDt3VgoMzjNE7nCw7LIqzaucrH46KHp3e1zMvKakQneqyWbhQdacS2FsmGLFsbFClJGcQjBQBFpLBQFAzCU6AhMBjxUilgGY7iJLEQpkonwELWX-RKOEHYj01XAjBxoQ61DBXCmM5oFIbffIfoYdZ0Z5qu6K8a8scDJ369P0a5zRF0Gd4baxbqEcwvnhbqo_PgFtYyejg |
linkProvider | IEEE |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV07T8MwELaqMsDEo0W88cBI0sZx7JitRFQtNC1SW9St8isSomqqkgzw67GTtAjEwGZ5sK072d_d-b47AG5owBEmiDshkto4KCRxBBbMwT7VKEyIkNQSnOMh6U3x4yyY1cDtlgujtS6Sz7Rrh8VfvkplbkNlLVaUojEO-o7B_cAr2Vqbd9cAfUgqcpzXZq1-FL3gkCHLtkKea7-A_B8tVAoE6e6DeLN3mTjy5uaZcOXnr7KM_z3cAWh-c_Xg8xaFDkFNL4_AfmVcwurqvjcAGY_izh0cpwsbQoCjVRHEhparo9fOvQEzBeM04ivYybO0KOTKF4uPJph2HyZRz6maJjivxt3MHC0FlZwJY7hIpqy1wHkQ-tprC8SVASOvrRVGMtAq0dQXAhkfRlEVGJASKGH-Magv06U-AdA2TRceUbwtFTYqZZoSmeDE86lZB8lT0LBimK_KuhjzSgJnf09fg93eJB7MB_3h0znYs0opk-IuQD1b5_rSgHsmrgqdfgEvpaHX |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=Proceedings+%2F+IEEE+International+Conference+on+Computer+Vision&rft.atitle=SOMA%3A+Solving+Optical+Marker-Based+MoCap+Automatically&rft.au=Ghorbani%2C+Nima&rft.au=Black%2C+Michael+J.&rft.date=2021-10-01&rft.pub=IEEE&rft.eissn=2380-7504&rft.spage=11097&rft.epage=11106&rft_id=info:doi/10.1109%2FICCV48922.2021.01093&rft.externalDocID=9710080 |