Towards Graph Representation Learning Based Surgical Workflow Anticipation

Surgical workflow anticipation can give predictions on what steps to conduct or what instruments to use next, which is an essential part of the computer-assisted intervention system for surgery, e.g. workflow reasoning in robotic surgery. However, current approaches are limited to their insufficient...

Full description

Saved in:
Bibliographic Details
Main Authors Zhang, Xiatian, Moubayed, Noura Al, Shum, Hubert P. H
Format Journal Article
LanguageEnglish
Published 07.08.2022
Subjects
Online AccessGet full text

Cover

Loading…
Abstract Surgical workflow anticipation can give predictions on what steps to conduct or what instruments to use next, which is an essential part of the computer-assisted intervention system for surgery, e.g. workflow reasoning in robotic surgery. However, current approaches are limited to their insufficient expressive power for relationships between instruments. Hence, we propose a graph representation learning framework to comprehensively represent instrument motions in the surgical workflow anticipation problem. In our proposed graph representation, we maps the bounding box information of instruments to the graph nodes in the consecutive frames and build inter-frame/inter-instrument graph edges to represent the trajectory and interaction of the instruments over time. This design enhances the ability of our network on modeling both the spatial and temporal patterns of surgical instruments and their interactions. In addition, we design a multi-horizon learning strategy to balance the understanding of various horizons indifferent anticipation tasks, which significantly improves the model performance in anticipation with various horizons. Experiments on the Cholec80 dataset demonstrate the performance of our proposed method can exceed the state-of-the-art method based on richer backbones, especially in instrument anticipation (1.27 v.s. 1.48 for inMAE; 1.48 v.s. 2.68 for eMAE). To the best of our knowledge, we are the first to introduce a spatial-temporal graph representation into surgical workflow anticipation.
AbstractList Surgical workflow anticipation can give predictions on what steps to conduct or what instruments to use next, which is an essential part of the computer-assisted intervention system for surgery, e.g. workflow reasoning in robotic surgery. However, current approaches are limited to their insufficient expressive power for relationships between instruments. Hence, we propose a graph representation learning framework to comprehensively represent instrument motions in the surgical workflow anticipation problem. In our proposed graph representation, we maps the bounding box information of instruments to the graph nodes in the consecutive frames and build inter-frame/inter-instrument graph edges to represent the trajectory and interaction of the instruments over time. This design enhances the ability of our network on modeling both the spatial and temporal patterns of surgical instruments and their interactions. In addition, we design a multi-horizon learning strategy to balance the understanding of various horizons indifferent anticipation tasks, which significantly improves the model performance in anticipation with various horizons. Experiments on the Cholec80 dataset demonstrate the performance of our proposed method can exceed the state-of-the-art method based on richer backbones, especially in instrument anticipation (1.27 v.s. 1.48 for inMAE; 1.48 v.s. 2.68 for eMAE). To the best of our knowledge, we are the first to introduce a spatial-temporal graph representation into surgical workflow anticipation.
Author Shum, Hubert P. H
Zhang, Xiatian
Moubayed, Noura Al
Author_xml – sequence: 1
  givenname: Xiatian
  surname: Zhang
  fullname: Zhang, Xiatian
– sequence: 2
  givenname: Noura Al
  surname: Moubayed
  fullname: Moubayed, Noura Al
– sequence: 3
  givenname: Hubert P. H
  surname: Shum
  fullname: Shum, Hubert P. H
BackLink https://doi.org/10.48550/arXiv.2208.03824$$DView paper in arXiv
BookMark eNotz7FOwzAUhWEPMEDhAZjwCyS49nXjjqWCAoqEBJEYo2v7ulgEJ3IChbdHBKaz_DrSd8qOUp-IsYulKMFoLa4wf8XPUkphSqGMhBP20PQHzH7ku4zDK3-iIdNIacIp9onXhDnFtOfXOJLnzx95Hx12_KXPb6HrD3yTpujiMNdn7DhgN9L5_y5Yc3vTbO-K-nF3v93UBa4qKFB4o-16ufLagBMhCJJAJC05VXlSxmJQhNU6CKVNQGsBPAgJlXIetFULdvl3O2PaIcd3zN_tL6qdUeoHg2BKgw
ContentType Journal Article
Copyright http://arxiv.org/licenses/nonexclusive-distrib/1.0
Copyright_xml – notice: http://arxiv.org/licenses/nonexclusive-distrib/1.0
DBID AKY
GOX
DOI 10.48550/arxiv.2208.03824
DatabaseName arXiv Computer Science
arXiv.org
DatabaseTitleList
Database_xml – sequence: 1
  dbid: GOX
  name: arXiv.org
  url: http://arxiv.org/find
  sourceTypes: Open Access Repository
DeliveryMethod fulltext_linktorsrc
ExternalDocumentID 2208_03824
GroupedDBID AKY
GOX
ID FETCH-LOGICAL-a674-a0d85b916d584c0ff0e24ee2bec37de38baf3ea79f0358fabb44d402473cd45b3
IEDL.DBID GOX
IngestDate Mon Jan 08 05:40:35 EST 2024
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-a674-a0d85b916d584c0ff0e24ee2bec37de38baf3ea79f0358fabb44d402473cd45b3
OpenAccessLink https://arxiv.org/abs/2208.03824
ParticipantIDs arxiv_primary_2208_03824
PublicationCentury 2000
PublicationDate 2022-08-07
PublicationDateYYYYMMDD 2022-08-07
PublicationDate_xml – month: 08
  year: 2022
  text: 2022-08-07
  day: 07
PublicationDecade 2020
PublicationYear 2022
Score 1.8519592
SecondaryResourceType preprint
Snippet Surgical workflow anticipation can give predictions on what steps to conduct or what instruments to use next, which is an essential part of the...
SourceID arxiv
SourceType Open Access Repository
SubjectTerms Computer Science - Computer Vision and Pattern Recognition
Computer Science - Learning
Title Towards Graph Representation Learning Based Surgical Workflow Anticipation
URI https://arxiv.org/abs/2208.03824
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdV1NSwMxEB3anryIolI_ycHr4jbJNrvHKraloIKusLcls0lEkK1sW-vPd5JdqRevSS7zBvImyZsXgOsqMTGaCikDykTSZCJCq2SUETkR341QpL45-eFxPH-ViyIpesB-e2F08_3-1foD4-qGcy91FCmXfehz7iVbs6eifZwMVlzd-t06qjHD0B-SmB7AflfdsUmbjkPo2foIFnmQpq7YzJtDs-egPe1afmrWGZy-sVviE8NeNk3Yi5i_xXYfyy2b1Dvh8zHk0_v8bh51XxhEekwR69ikCVIFZijuKnYutlxaywk4oYwVKWonrFaZi0WSOo0opaETnVSiMjJBcQKDelnbITBpCGuHiciIUG0sUDiJmVNSq5EhTE9hGAIvP1uXitJjUgZMzv6fOoc97vX8XgOhLmCwbjb2klh2jVcB6h91wn63
link.rule.ids 228,230,783,888
linkProvider Cornell University
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Towards+Graph+Representation+Learning+Based+Surgical+Workflow+Anticipation&rft.au=Zhang%2C+Xiatian&rft.au=Moubayed%2C+Noura+Al&rft.au=Shum%2C+Hubert+P.+H&rft.date=2022-08-07&rft_id=info:doi/10.48550%2Farxiv.2208.03824&rft.externalDocID=2208_03824