Towards Graph Representation Learning Based Surgical Workflow Anticipation
Surgical workflow anticipation can give predictions on what steps to conduct or what instruments to use next, which is an essential part of the computer-assisted intervention system for surgery, e.g. workflow reasoning in robotic surgery. However, current approaches are limited to their insufficient...
Saved in:
Main Authors | , , |
---|---|
Format | Journal Article |
Language | English |
Published |
07.08.2022
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | Surgical workflow anticipation can give predictions on what steps to conduct
or what instruments to use next, which is an essential part of the
computer-assisted intervention system for surgery, e.g. workflow reasoning in
robotic surgery. However, current approaches are limited to their insufficient
expressive power for relationships between instruments. Hence, we propose a
graph representation learning framework to comprehensively represent instrument
motions in the surgical workflow anticipation problem. In our proposed graph
representation, we maps the bounding box information of instruments to the
graph nodes in the consecutive frames and build inter-frame/inter-instrument
graph edges to represent the trajectory and interaction of the instruments over
time. This design enhances the ability of our network on modeling both the
spatial and temporal patterns of surgical instruments and their interactions.
In addition, we design a multi-horizon learning strategy to balance the
understanding of various horizons indifferent anticipation tasks, which
significantly improves the model performance in anticipation with various
horizons. Experiments on the Cholec80 dataset demonstrate the performance of
our proposed method can exceed the state-of-the-art method based on richer
backbones, especially in instrument anticipation (1.27 v.s. 1.48 for inMAE;
1.48 v.s. 2.68 for eMAE). To the best of our knowledge, we are the first to
introduce a spatial-temporal graph representation into surgical workflow
anticipation. |
---|---|
AbstractList | Surgical workflow anticipation can give predictions on what steps to conduct
or what instruments to use next, which is an essential part of the
computer-assisted intervention system for surgery, e.g. workflow reasoning in
robotic surgery. However, current approaches are limited to their insufficient
expressive power for relationships between instruments. Hence, we propose a
graph representation learning framework to comprehensively represent instrument
motions in the surgical workflow anticipation problem. In our proposed graph
representation, we maps the bounding box information of instruments to the
graph nodes in the consecutive frames and build inter-frame/inter-instrument
graph edges to represent the trajectory and interaction of the instruments over
time. This design enhances the ability of our network on modeling both the
spatial and temporal patterns of surgical instruments and their interactions.
In addition, we design a multi-horizon learning strategy to balance the
understanding of various horizons indifferent anticipation tasks, which
significantly improves the model performance in anticipation with various
horizons. Experiments on the Cholec80 dataset demonstrate the performance of
our proposed method can exceed the state-of-the-art method based on richer
backbones, especially in instrument anticipation (1.27 v.s. 1.48 for inMAE;
1.48 v.s. 2.68 for eMAE). To the best of our knowledge, we are the first to
introduce a spatial-temporal graph representation into surgical workflow
anticipation. |
Author | Shum, Hubert P. H Zhang, Xiatian Moubayed, Noura Al |
Author_xml | – sequence: 1 givenname: Xiatian surname: Zhang fullname: Zhang, Xiatian – sequence: 2 givenname: Noura Al surname: Moubayed fullname: Moubayed, Noura Al – sequence: 3 givenname: Hubert P. H surname: Shum fullname: Shum, Hubert P. H |
BackLink | https://doi.org/10.48550/arXiv.2208.03824$$DView paper in arXiv |
BookMark | eNotz7FOwzAUhWEPMEDhAZjwCyS49nXjjqWCAoqEBJEYo2v7ulgEJ3IChbdHBKaz_DrSd8qOUp-IsYulKMFoLa4wf8XPUkphSqGMhBP20PQHzH7ku4zDK3-iIdNIacIp9onXhDnFtOfXOJLnzx95Hx12_KXPb6HrD3yTpujiMNdn7DhgN9L5_y5Yc3vTbO-K-nF3v93UBa4qKFB4o-16ufLagBMhCJJAJC05VXlSxmJQhNU6CKVNQGsBPAgJlXIetFULdvl3O2PaIcd3zN_tL6qdUeoHg2BKgw |
ContentType | Journal Article |
Copyright | http://arxiv.org/licenses/nonexclusive-distrib/1.0 |
Copyright_xml | – notice: http://arxiv.org/licenses/nonexclusive-distrib/1.0 |
DBID | AKY GOX |
DOI | 10.48550/arxiv.2208.03824 |
DatabaseName | arXiv Computer Science arXiv.org |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: GOX name: arXiv.org url: http://arxiv.org/find sourceTypes: Open Access Repository |
DeliveryMethod | fulltext_linktorsrc |
ExternalDocumentID | 2208_03824 |
GroupedDBID | AKY GOX |
ID | FETCH-LOGICAL-a674-a0d85b916d584c0ff0e24ee2bec37de38baf3ea79f0358fabb44d402473cd45b3 |
IEDL.DBID | GOX |
IngestDate | Mon Jan 08 05:40:35 EST 2024 |
IsDoiOpenAccess | true |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-a674-a0d85b916d584c0ff0e24ee2bec37de38baf3ea79f0358fabb44d402473cd45b3 |
OpenAccessLink | https://arxiv.org/abs/2208.03824 |
ParticipantIDs | arxiv_primary_2208_03824 |
PublicationCentury | 2000 |
PublicationDate | 2022-08-07 |
PublicationDateYYYYMMDD | 2022-08-07 |
PublicationDate_xml | – month: 08 year: 2022 text: 2022-08-07 day: 07 |
PublicationDecade | 2020 |
PublicationYear | 2022 |
Score | 1.8519592 |
SecondaryResourceType | preprint |
Snippet | Surgical workflow anticipation can give predictions on what steps to conduct
or what instruments to use next, which is an essential part of the... |
SourceID | arxiv |
SourceType | Open Access Repository |
SubjectTerms | Computer Science - Computer Vision and Pattern Recognition Computer Science - Learning |
Title | Towards Graph Representation Learning Based Surgical Workflow Anticipation |
URI | https://arxiv.org/abs/2208.03824 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdV1NSwMxEB3anryIolI_ycHr4jbJNrvHKraloIKusLcls0lEkK1sW-vPd5JdqRevSS7zBvImyZsXgOsqMTGaCikDykTSZCJCq2SUETkR341QpL45-eFxPH-ViyIpesB-e2F08_3-1foD4-qGcy91FCmXfehz7iVbs6eifZwMVlzd-t06qjHD0B-SmB7AflfdsUmbjkPo2foIFnmQpq7YzJtDs-egPe1afmrWGZy-sVviE8NeNk3Yi5i_xXYfyy2b1Dvh8zHk0_v8bh51XxhEekwR69ikCVIFZijuKnYutlxaywk4oYwVKWonrFaZi0WSOo0opaETnVSiMjJBcQKDelnbITBpCGuHiciIUG0sUDiJmVNSq5EhTE9hGAIvP1uXitJjUgZMzv6fOoc97vX8XgOhLmCwbjb2klh2jVcB6h91wn63 |
link.rule.ids | 228,230,783,888 |
linkProvider | Cornell University |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Towards+Graph+Representation+Learning+Based+Surgical+Workflow+Anticipation&rft.au=Zhang%2C+Xiatian&rft.au=Moubayed%2C+Noura+Al&rft.au=Shum%2C+Hubert+P.+H&rft.date=2022-08-07&rft_id=info:doi/10.48550%2Farxiv.2208.03824&rft.externalDocID=2208_03824 |