SOMA: Solving Optical Marker-Based MoCap Automatically

Marker-based optical motion capture (mocap) is the "gold standard" method for acquiring accurate 3D human motion in computer vision, medicine, and graphics. The raw output of these systems are noisy and incomplete 3D points or short tracklets of points. To be useful, one must associate the...

Full description

Saved in:
Bibliographic Details
Published inProceedings / IEEE International Conference on Computer Vision pp. 11097 - 11106
Main Authors Ghorbani, Nima, Black, Michael J.
Format Conference Proceeding
LanguageEnglish
Published IEEE 01.10.2021
Subjects
Online AccessGet full text

Cover

Loading…
Abstract Marker-based optical motion capture (mocap) is the "gold standard" method for acquiring accurate 3D human motion in computer vision, medicine, and graphics. The raw output of these systems are noisy and incomplete 3D points or short tracklets of points. To be useful, one must associate these points with corresponding markers on the captured subject; i.e. "labelling". Given these labels, one can then "solve" for the 3D skeleton or body surface mesh. Commercial auto-labeling tools require a specific calibration procedure at capture time, which is not possible for archival data. Here we train a novel neural network called SOMA, which takes raw mocap point clouds with varying numbers of points, labels them at scale without any calibration data, independent of the capture technology, and requiring only minimal human intervention. Our key insight is that, while labeling point clouds is highly ambiguous, the 3D body provides strong constraints on the solution that can be exploited by a learning-based method. To enable learning, we generate massive training sets of simulated noisy and ground truth mocap markers animated by 3D bodies from AMASS. SOMA exploits an architecture with stacked self-attention elements to learn the spatial structure of the 3D body and an optimal transport layer to constrain the assignment (labeling) problem while rejecting outliers. We extensively evaluate SOMA both quantitatively and qualitatively. SOMA is more accurate and robust than existing state of the art research methods and can be applied where commercial systems cannot. We automatically label over 8 hours of archival mocap data across 4 different datasets captured using various technologies and output SMPL-X body models. The model and data is released for research purposes at https://soma.is.tue.mpg.de/.
AbstractList Marker-based optical motion capture (mocap) is the "gold standard" method for acquiring accurate 3D human motion in computer vision, medicine, and graphics. The raw output of these systems are noisy and incomplete 3D points or short tracklets of points. To be useful, one must associate these points with corresponding markers on the captured subject; i.e. "labelling". Given these labels, one can then "solve" for the 3D skeleton or body surface mesh. Commercial auto-labeling tools require a specific calibration procedure at capture time, which is not possible for archival data. Here we train a novel neural network called SOMA, which takes raw mocap point clouds with varying numbers of points, labels them at scale without any calibration data, independent of the capture technology, and requiring only minimal human intervention. Our key insight is that, while labeling point clouds is highly ambiguous, the 3D body provides strong constraints on the solution that can be exploited by a learning-based method. To enable learning, we generate massive training sets of simulated noisy and ground truth mocap markers animated by 3D bodies from AMASS. SOMA exploits an architecture with stacked self-attention elements to learn the spatial structure of the 3D body and an optimal transport layer to constrain the assignment (labeling) problem while rejecting outliers. We extensively evaluate SOMA both quantitatively and qualitatively. SOMA is more accurate and robust than existing state of the art research methods and can be applied where commercial systems cannot. We automatically label over 8 hours of archival mocap data across 4 different datasets captured using various technologies and output SMPL-X body models. The model and data is released for research purposes at https://soma.is.tue.mpg.de/.
Author Ghorbani, Nima
Black, Michael J.
Author_xml – sequence: 1
  givenname: Nima
  surname: Ghorbani
  fullname: Ghorbani, Nima
  email: nghorbani@tuebingen.mpg.de
  organization: Max Planck Institute for Intelligent Systems,Tübingen,Germany
– sequence: 2
  givenname: Michael J.
  surname: Black
  fullname: Black, Michael J.
  email: black@tuebingen.mpg.de
  organization: Max Planck Institute for Intelligent Systems,Tübingen,Germany
BookMark eNotjtFKwzAUhqMouM09gV70BVpPTpIm8a6WqYOVXky9HUlzKtWuLe0U9vYO9eqH74OPf84uur4jxm45JJyDvVvn-Zs0FjFBQJ7AiYkztrTa8DRVEg1Hdc5mKAzEWoG8YvNp-gAQFk06Y-m2LLL7aNu33033HpXDoalcGxVu_KQxfnAThajoczdE2deh37tf3R6v2WXt2omW_7tgr4-rl_w53pRP6zzbxA0X4hBT5XXlrD99qGzgSknnlBHEwaMLaCQHChIrRaEmLbxHm0LQQWluPNZWLNjNX7chot0wNns3HndWcwAD4gdh8UeG
CODEN IEEPAD
ContentType Conference Proceeding
DBID 6IE
6IH
CBEJK
RIE
RIO
DOI 10.1109/ICCV48922.2021.01093
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Proceedings Order Plan (POP) 1998-present by volume
IEEE Xplore All Conference Proceedings
IEEE Electronic Library (IEL)
IEEE Proceedings Order Plans (POP) 1998-present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
Discipline Applied Sciences
EISBN 9781665428125
1665428120
EISSN 2380-7504
EndPage 11106
ExternalDocumentID 9710080
Genre orig-research
GroupedDBID 29O
6IE
6IF
6IH
6IK
6IL
6IM
6IN
AAJGR
AAWTH
ACGFS
ADZIZ
ALMA_UNASSIGNED_HOLDINGS
BEFXN
BFFAM
BGNUA
BKEBE
BPEOZ
CBEJK
CHZPO
IPLJI
M43
OCL
RIE
RIL
RIO
RNS
ID FETCH-LOGICAL-i133t-ecb7ca9b125c9d1554aa583e10b2ad28410ed42c5edfe73bb2960d7d5718b2f93
IEDL.DBID RIE
IngestDate Wed Aug 27 02:25:36 EDT 2025
IsPeerReviewed false
IsScholarly true
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-i133t-ecb7ca9b125c9d1554aa583e10b2ad28410ed42c5edfe73bb2960d7d5718b2f93
PageCount 10
ParticipantIDs ieee_primary_9710080
PublicationCentury 2000
PublicationDate 2021-Oct.
PublicationDateYYYYMMDD 2021-10-01
PublicationDate_xml – month: 10
  year: 2021
  text: 2021-Oct.
PublicationDecade 2020
PublicationTitle Proceedings / IEEE International Conference on Computer Vision
PublicationTitleAbbrev ICCV
PublicationYear 2021
Publisher IEEE
Publisher_xml – name: IEEE
SSID ssj0039286
Score 2.271866
Snippet Marker-based optical motion capture (mocap) is the "gold standard" method for acquiring accurate 3D human motion in computer vision, medicine, and graphics....
SourceID ieee
SourceType Publisher
StartPage 11097
SubjectTerms Computer vision
Data models
Datasets and evaluation
Gestures and body pose
Learning systems
Motion and tracking
Point cloud compression
Three-dimensional displays
Training
Training data
Title SOMA: Solving Optical Marker-Based MoCap Automatically
URI https://ieeexplore.ieee.org/document/9710080
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV09T8MwELVKJ6YCLeJbHhhJ2jiOHbOVCFSQQpFKUbfKHxcJUTVVSQb49dhJWgRiYIsyxNGd7HfvfO8OoUseSUIZkV5MNFiCwjJPUSU8GnIgccaU5k7gnD6y0ZQ-zKJZC11ttTAAUBWfge8eq7t8k-vSpcr6ompFYwn6jiVutVZrc-pamI9ZI40LBqJ_nyQvNBbEaa1I4LsLoPDHAJUKP-46KN2sXJeNvPlloXz9-asp439_bQ_1vpV6-GmLQfuoBcsD1GlCS9xs3PcuYpNxOrzGk3zhEgh4vKpS2NgpdWDt3VgoMzjNE7nCw7LIqzaucrH46KHp3e1zMvKakQneqyWbhQdacS2FsmGLFsbFClJGcQjBQBFpLBQFAzCU6AhMBjxUilgGY7iJLEQpkonwELWX-RKOEHYj01XAjBxoQ61DBXCmM5oFIbffIfoYdZ0Z5qu6K8a8scDJ369P0a5zRF0Gd4baxbqEcwvnhbqo_PgFtYyejg
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV07T8MwELaqMsDEo0W88cBI0sZx7JitRFQtNC1SW9St8isSomqqkgzw67GTtAjEwGZ5sK072d_d-b47AG5owBEmiDshkto4KCRxBBbMwT7VKEyIkNQSnOMh6U3x4yyY1cDtlgujtS6Sz7Rrh8VfvkplbkNlLVaUojEO-o7B_cAr2Vqbd9cAfUgqcpzXZq1-FL3gkCHLtkKea7-A_B8tVAoE6e6DeLN3mTjy5uaZcOXnr7KM_z3cAWh-c_Xg8xaFDkFNL4_AfmVcwurqvjcAGY_izh0cpwsbQoCjVRHEhparo9fOvQEzBeM04ivYybO0KOTKF4uPJph2HyZRz6maJjivxt3MHC0FlZwJY7hIpqy1wHkQ-tprC8SVASOvrRVGMtAq0dQXAhkfRlEVGJASKGH-Magv06U-AdA2TRceUbwtFTYqZZoSmeDE86lZB8lT0LBimK_KuhjzSgJnf09fg93eJB7MB_3h0znYs0opk-IuQD1b5_rSgHsmrgqdfgEvpaHX
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=Proceedings+%2F+IEEE+International+Conference+on+Computer+Vision&rft.atitle=SOMA%3A+Solving+Optical+Marker-Based+MoCap+Automatically&rft.au=Ghorbani%2C+Nima&rft.au=Black%2C+Michael+J.&rft.date=2021-10-01&rft.pub=IEEE&rft.eissn=2380-7504&rft.spage=11097&rft.epage=11106&rft_id=info:doi/10.1109%2FICCV48922.2021.01093&rft.externalDocID=9710080