Towards Real-Time Monocular Depth Estimation for Robotics: A Survey
As an essential component for many autonomous driving and robotic activities such as ego-motion estimation, obstacle avoidance and scene understanding, monocular depth estimation (MDE) has attracted great attention from the computer vision and robotics communities. Over the past decades, a large num...
Saved in:
Published in | IEEE transactions on intelligent transportation systems Vol. 23; no. 10; pp. 16940 - 16961 |
---|---|
Main Authors | , , , |
Format | Journal Article |
Language | English |
Published |
New York
IEEE
01.10.2022
The Institute of Electrical and Electronics Engineers, Inc. (IEEE) |
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | As an essential component for many autonomous driving and robotic activities such as ego-motion estimation, obstacle avoidance and scene understanding, monocular depth estimation (MDE) has attracted great attention from the computer vision and robotics communities. Over the past decades, a large number of methods have been developed. To the best of our knowledge, however, there is not a comprehensive survey of MDE. This paper aims to bridge this gap by reviewing 197 relevant articles published between 1970 and 2021. In particular, we provide a comprehensive survey of MDE covering various methods, introduce the popular performance evaluation metrics and summarize publically available datasets. We also summarize available open-source implementations of some representative methods and compare their performances. Furthermore, we review the application of MDE in some important robotic tasks. Finally, we conclude this paper by presenting some promising directions for future research. This survey is expected to assist readers to navigate this research field. |
---|---|
AbstractList | As an essential component for many autonomous driving and robotic activities such as ego-motion estimation, obstacle avoidance and scene understanding, monocular depth estimation (MDE) has attracted great attention from the computer vision and robotics communities. Over the past decades, a large number of methods have been developed. To the best of our knowledge, however, there is not a comprehensive survey of MDE. This paper aims to bridge this gap by reviewing 197 relevant articles published between 1970 and 2021. In particular, we provide a comprehensive survey of MDE covering various methods, introduce the popular performance evaluation metrics and summarize publically available datasets. We also summarize available open-source implementations of some representative methods and compare their performances. Furthermore, we review the application of MDE in some important robotic tasks. Finally, we conclude this paper by presenting some promising directions for future research. This survey is expected to assist readers to navigate this research field. |
Author | Abbass, Hussein A. Dong, Xingshuai Anavatti, Sreenatha G. Garratt, Matthew A. |
Author_xml | – sequence: 1 givenname: Xingshuai orcidid: 0000-0003-3900-1038 surname: Dong fullname: Dong, Xingshuai email: xingshuai.dong@student.adfa.edu.au organization: School of Engineering and Information Technology, University of New South Wales, Canberra, ACT, Australia – sequence: 2 givenname: Matthew A. orcidid: 0000-0003-0222-430X surname: Garratt fullname: Garratt, Matthew A. email: mattgarratt@gmail.com organization: School of Engineering and Information Technology, University of New South Wales, Canberra, ACT, Australia – sequence: 3 givenname: Sreenatha G. orcidid: 0000-0002-4754-8191 surname: Anavatti fullname: Anavatti, Sreenatha G. email: agsrenat@adfa.edu.au organization: School of Engineering and Information Technology, University of New South Wales, Canberra, ACT, Australia – sequence: 4 givenname: Hussein A. orcidid: 0000-0002-8837-0748 surname: Abbass fullname: Abbass, Hussein A. email: h.abbass@unsw.edu.au organization: School of Engineering and Information Technology, University of New South Wales, Canberra, ACT, Australia |
BookMark | eNo9kN9LwzAQgINMcE7_APEl4HPnXbo0jW9jTh1MhK0-hyxNsGNrZtIq--9t2fDpjuO7X981GdS-toTcIYwRQT4Wi2I9ZsDYOMUMxAQvyBA5zxMAzAZ9ziaJBA5X5DrGbVedcMQhmRX-V4cy0pXVu6So9pa--9qbdqcDfbaH5ovOY1PtdVP5mjof6MpvfFOZ-ESndN2GH3u8IZdO76K9PccR-XyZF7O3ZPnxuphNl4lhMm0SocsSuXA8kxq1s8zZknMnnJV5XuapNJyZDEu2AeC5k9hxkBprGN84AZiOyMNp7iH479bGRm19G-pupWKCpbL7SUBH4YkywccYrFOH0N0fjgpB9a5U70r1rtTZVddzf-qprLX_vBQcZM7TPwqPZjI |
CODEN | ITISFG |
CitedBy_id | crossref_primary_10_3390_s24134282 crossref_primary_10_1007_s11227_023_05359_0 crossref_primary_10_3390_s23218821 crossref_primary_10_1145_3650039 crossref_primary_10_3390_s22249830 crossref_primary_10_1016_j_mlwa_2022_100416 crossref_primary_10_1109_JSSC_2022_3218278 crossref_primary_10_1002_aisy_202300382 crossref_primary_10_1007_s11042_023_17966_3 crossref_primary_10_1007_s00521_023_09025_7 crossref_primary_10_1007_s10489_023_04872_2 crossref_primary_10_1007_s10489_023_04851_7 crossref_primary_10_1016_j_cviu_2024_104014 crossref_primary_10_1109_TITS_2023_3341457 crossref_primary_10_3390_rs15235552 crossref_primary_10_1109_LSENS_2023_3309264 crossref_primary_10_3390_app13179940 crossref_primary_10_1016_j_asoc_2024_111873 crossref_primary_10_1371_journal_pcbi_1011796 crossref_primary_10_3390_s22145353 crossref_primary_10_1016_j_ins_2024_120874 crossref_primary_10_3390_electronics13122418 crossref_primary_10_1016_j_measurement_2023_112481 crossref_primary_10_1016_j_optlaseng_2024_108313 crossref_primary_10_1038_s44182_024_00010_3 |
Cites_doi | 10.1109/CVPR.2016.438 10.1109/CVPR.2019.01002 10.1109/CVPR.2018.00474 10.1109/CVPR.2015.7298897 10.1109/ICCV.2019.00225 10.1007/978-3-030-01270-0_7 10.1109/ROBIO49542.2019.8961504 10.1109/ICRA.2019.8793512 10.1109/IVS.2015.7225750 10.1007/978-3-030-11009-3_20 10.24963/ijcai.2019/98 10.1049/cp.2016.1453 10.1109/TVCG.2019.2921336 10.1109/LRA.2020.3017478 10.1007/978-3-030-20870-7_41 10.1007/978-3-030-58529-7_34 10.1109/3DV.2018.00044 10.1109/CVPR46437.2021.00400 10.1007/978-3-030-20893-6_19 10.1109/ICIP.2019.8803170 10.1109/CVPR.2016.350 10.1109/ICCV.2015.336 10.1109/CVPR.2018.00043 10.1109/WACV.2019.00226 10.1007/978-3-642-33715-4_54 10.1109/CVPR.2017.261 10.1007/978-3-030-01267-0_4 10.1109/3DTV.2010.5506603 10.1109/WACV45572.2020.9093334 10.29121/granthaalayah.v5.i4RACSIT.2017.3342 10.1109/TITS.2017.2750080 10.1109/CVPR42600.2020.00481 10.1109/TPAMI.2008.132 10.1109/ICCV.2019.00181 10.1109/IROS.2012.6385773 10.1109/LRA.2018.2800083 10.1007/978-3-030-58565-5_35 10.1007/978-3-030-01249-6_43 10.1109/CVPR.2014.97 10.1109/CVPR.2019.01003 10.1109/ICCV.2019.00393 10.1109/IROS45743.2020.9341435 10.1109/ICRA.2017.7989590 10.1109/LRA.2019.2896963 10.1109/TPAMI.2019.2936024 10.1109/ICRA.2019.8794425 10.1109/TMM.2018.2859034 10.1109/ICRA.2014.6907325 10.1109/TIP.2020.2968751 10.1109/TCSVT.2020.2973068 10.1109/CVPR.2015.7298715 10.1109/ICCV.2019.00397 10.1109/CVPR.2019.00570 10.1007/978-3-319-23778-7_14 10.1109/CVPR.2019.00099 10.1109/TPAMI.2002.1033214 10.1109/ICCV.2015.401 10.1109/WACV.2019.00116 10.1109/CVPR.2016.90 10.1016/j.robot.2015.11.011 10.1007/978-3-319-10605-2_54 10.1109/ITSC.2019.8917177 10.1109/CVPR.2016.470 10.1007/s10846-020-01205-0 10.1109/TITS.2014.2308977 10.1109/CVPR.2015.7298965 10.1109/CVPR42600.2020.00069 10.1109/ICASSP.2019.8683235 10.1109/TCSVT.2017.2740321 10.1109/IROS45743.2020.9341448 10.1109/ICRA.2019.8794404 10.1109/CVPR.2016.584 10.1109/3DV.2018.00017 10.1109/CVPR.2017.699 10.1109/CVPRW.2019.00081 10.7551/mitpress/3877.003.0009 10.1109/TPAMI.2014.2316835 10.1109/IROS45743.2020.9340802 10.1109/ICRA.2012.6224734 10.1007/978-3-030-01237-3_50 10.1109/ICAR46387.2019.8981652 10.1007/978-3-030-01252-6_2 10.1109/CVPR.2018.00214 10.1109/CVPR42600.2020.00054 10.1109/CVPR.2018.00212 10.1007/978-3-319-14249-4_71 10.1109/TPAMI.2015.2505283 10.1109/ICRA40945.2020.9197481 10.1109/ICRA.2018.8460184 10.1109/CVPR42600.2020.00256 10.1109/3DV.2016.32 10.1109/IVS.2018.8500683 10.1109/CVPR.2010.5539823 10.1109/CVPR.2017.243 10.1007/s13042-018-0891-9 10.1109/ICCV.2019.00227 10.1109/3DV.2017.00081 10.1007/978-3-030-11012-3_4 10.1109/CVPR.2019.00343 10.1109/TMM.2019.2941776 10.1109/ICIP.2019.8803544 10.1109/TRO.2017.2705103 10.1109/IROS.2018.8594243 10.1145/1102351.1102426 10.1109/CVPR.2018.00745 10.1109/ICRA.2012.6224636 10.1109/TITS.2019.2955598 10.1109/CVPR.2017.272 10.1109/CVPR.2018.00296 10.1109/ACCESS.2020.3000773 10.1109/CVPR.2018.00037 10.1007/978-3-030-01228-1_3 10.1109/ICRA.2017.7989752 10.1007/978-3-030-01225-0_3 10.1109/CVPR.2015.7298655 10.1109/IROS.2018.8593814 10.1109/ICIP.2019.8803100 10.1109/CVPR.2017.695 10.1109/CVPR.2017.25 10.1109/ROBIO.2015.7418959 10.1109/ICRA.2017.7989537 10.1109/CVPR.2018.00594 10.1109/WACV.2016.7477636 10.1007/s13042-020-01251-y 10.1609/aaai.v33i01.33018001 10.1007/978-3-319-46484-8_45 10.1109/CVPR42600.2020.00917 10.1007/11861898_48 10.1109/ICCV.2017.292 10.1109/CVPR42600.2020.00097 10.1109/DSC50466.2020.00028 10.1109/CVPR42600.2020.00339 10.1109/ICCV.2015.304 10.1109/ICASSP.2018.8462433 10.1016/j.isprsjprs.2020.06.004 10.1109/ICRA.2019.8794220 10.1109/ICRA.2019.8793637 10.1109/ICRA.2019.8794182 10.1109/ICARCV.2016.7838832 10.1177/0278364913491297 10.1109/ICCV.2019.00578 10.1109/ACCESS.2020.2983149 10.1109/ICCP.2015.7312626 10.1109/ICCV.2019.01012 10.1109/ICCV.2017.625 10.1007/978-3-030-01231-1_28 10.1109/CVPR.2018.00218 10.1109/TPAMI.2019.2940948 10.1109/3DV.2019.00081 10.1109/CVPR.2011.5995480 10.1109/3DV.2018.00073 10.1109/CVPR.2017.238 10.1109/CVPR.2019.01252 10.1109/CVPR.2014.19 10.1016/j.patcog.2018.05.029 10.1007/978-3-030-01252-6_30 10.1109/CVPRW.2018.00066 10.1109/CCIOT48581.2019.8980350 10.1109/IVS.2014.6856419 10.1109/CVPR.2017.28 10.1109/IROS.2016.7759632 10.1109/CVPR.2017.700 10.1109/ICCVW.2017.108 |
ContentType | Journal Article |
Copyright | Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2022 |
Copyright_xml | – notice: Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2022 |
DBID | 97E RIA RIE AAYXX CITATION 7SC 7SP 8FD FR3 JQ2 KR7 L7M L~C L~D |
DOI | 10.1109/TITS.2022.3160741 |
DatabaseName | IEEE All-Society Periodicals Package (ASPP) 2005-present IEEE All-Society Periodicals Package (ASPP) 1998–Present IEEE Xplore CrossRef Computer and Information Systems Abstracts Electronics & Communications Abstracts Technology Research Database Engineering Research Database ProQuest Computer Science Collection Civil Engineering Abstracts Advanced Technologies Database with Aerospace Computer and Information Systems Abstracts Academic Computer and Information Systems Abstracts Professional |
DatabaseTitle | CrossRef Civil Engineering Abstracts Technology Research Database Computer and Information Systems Abstracts – Academic Electronics & Communications Abstracts ProQuest Computer Science Collection Computer and Information Systems Abstracts Engineering Research Database Advanced Technologies Database with Aerospace Computer and Information Systems Abstracts Professional |
DatabaseTitleList | Civil Engineering Abstracts |
Database_xml | – sequence: 1 dbid: RIE name: IEEE Xplore url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/ sourceTypes: Publisher |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Engineering |
EISSN | 1558-0016 |
EndPage | 16961 |
ExternalDocumentID | 10_1109_TITS_2022_3160741 9750985 |
Genre | orig-research |
GrantInformation_xml | – fundername: University of New South Wales Tuition Fee Scholarship funderid: 10.13039/501100001773 |
GroupedDBID | -~X 0R~ 29I 4.4 5GY 5VS 6IK 97E AAJGR AASAJ ABQJQ ABTAH ABVLG ACGFO ACGFS ACIWK ACNCT AENEX AETIX AIBXA AKJIK ALMA_UNASSIGNED_HOLDINGS ATWAV BEFXN BFFAM BGNUA BKEBE BPEOZ CS3 DU5 EBS EJD HZ~ H~9 IFIPE IPLJI JAVBF LAI M43 O9- OCL P2P PQQKQ RIA RIE RIG RNS XFK ZY4 AAYXX CITATION 7SC 7SP 8FD FR3 JQ2 KR7 L7M L~C L~D |
ID | FETCH-LOGICAL-c293t-7add157f569a1afe2fed55f7fe988d839c52c61d2b0058f9169a03cec25bf7013 |
IEDL.DBID | RIE |
ISSN | 1524-9050 |
IngestDate | Thu Oct 10 16:47:34 EDT 2024 Fri Aug 23 03:47:10 EDT 2024 Wed Jun 26 19:24:57 EDT 2024 |
IsPeerReviewed | true |
IsScholarly | true |
Issue | 10 |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-c293t-7add157f569a1afe2fed55f7fe988d839c52c61d2b0058f9169a03cec25bf7013 |
ORCID | 0000-0003-3900-1038 0000-0002-8837-0748 0000-0002-4754-8191 0000-0003-0222-430X |
PQID | 2723901470 |
PQPubID | 75735 |
PageCount | 22 |
ParticipantIDs | ieee_primary_9750985 crossref_primary_10_1109_TITS_2022_3160741 proquest_journals_2723901470 |
PublicationCentury | 2000 |
PublicationDate | 2022-10-01 |
PublicationDateYYYYMMDD | 2022-10-01 |
PublicationDate_xml | – month: 10 year: 2022 text: 2022-10-01 day: 01 |
PublicationDecade | 2020 |
PublicationPlace | New York |
PublicationPlace_xml | – name: New York |
PublicationTitle | IEEE transactions on intelligent transportation systems |
PublicationTitleAbbrev | TITS |
PublicationYear | 2022 |
Publisher | IEEE The Institute of Electrical and Electronics Engineers, Inc. (IEEE) |
Publisher_xml | – name: IEEE – name: The Institute of Electrical and Electronics Engineers, Inc. (IEEE) |
References | ref56 ref58 ref53 ref52 ref55 ref54 ref51 ref50 ref45 ref48 ref47 ref42 ref43 Chavan (ref28) 2015; 2 ref49 ref8 ref7 ref9 ref4 ref3 ref6 ref5 Zhang (ref117) 2019 ref100 ref101 ref40 ref35 ref37 ref36 ref31 ref30 Vyshna (ref34) 2019; 5 Dosovitskiy (ref91) 2020 ref39 ref38 Guizilini (ref119) 2020 Saxena (ref13) Laga (ref33) 2019 ref24 ref23 ref26 ref25 ref22 ref21 ref27 Peng (ref137) 2019 Horn (ref73) 1970 ref128 ref129 ref97 ref126 ref96 ref127 ref99 ref124 ref98 ref125 ref93 ref133 Zheng (ref143) ref92 ref134 ref95 ref131 ref94 ref132 Alhashim (ref86) 2018 ref130 ref89 ref139 ref85 ref138 ref88 ref135 ref136 Barrow (ref74) 1978; 2 ref82 ref144 ref81 ref145 ref84 ref142 ref83 Bian (ref123) 2019 ref140 ref141 ref80 ref79 ref108 ref78 ref109 ref106 ref107 ref75 ref104 ref105 ref77 ref102 ref76 ref103 ref71 ref70 ref112 ref72 ref110 ref68 ref67 ref69 ref118 ref64 ref115 ref63 ref116 ref66 ref65 ref114 ref60 ref122 ref62 ref120 ref61 ref121 RoJas-Perez (ref186) Mohan (ref29) 2015; 4 Chen (ref44) ref168 ref169 ref170 Vijayanarasimhan (ref149) 2017 ref177 ref178 ref175 de Queiroz Mendes (ref179) 2020 ref176 ref174 ref171 ref172 ref180 ref181 ref188 ref189 ref187 ref184 ref185 ref182 ref183 Brockers (ref192) ref148 ref146 ref147 ref155 ref156 ref153 ref154 ref151 ref152 Bhoi (ref32) 2019 ref159 ref157 ref158 Javidnia (ref11) Howard (ref173) 2017 Cho (ref57) 2019 ref166 ref167 ref164 ref165 ref162 ref163 ref161 ref12 ref15 ref14 ref10 ref17 ref16 Wang (ref113) 2020 ref18 Cabon (ref41) 2020 Redmon (ref111) 2018 Dai (ref150) 2019 Hussain Raza (ref19) 2015 Eigen (ref20) Vaswani (ref90) 2017 Han Lee (ref87) 2019 ref2 ref1 Armeni (ref46) 2017 ref191 ref190 ref197 ref195 ref196 ref193 ref194 Vasiljevic (ref59) 2019 Guizilini (ref160) |
References_xml | – year: 2020 ident: ref41 article-title: Virtual KITTI 2 publication-title: arXiv:2001.10773 contributor: fullname: Cabon – ident: ref145 doi: 10.1109/CVPR.2016.438 – ident: ref144 doi: 10.1109/CVPR.2019.01002 – ident: ref174 doi: 10.1109/CVPR.2018.00474 – ident: ref102 doi: 10.1109/CVPR.2015.7298897 – ident: ref134 doi: 10.1109/ICCV.2019.00225 – ident: ref77 doi: 10.1007/978-3-030-01270-0_7 – ident: ref141 doi: 10.1109/ROBIO49542.2019.8961504 – ident: ref156 doi: 10.1109/ICRA.2019.8793512 – ident: ref185 doi: 10.1109/IVS.2015.7225750 – ident: ref68 doi: 10.1007/978-3-030-11009-3_20 – ident: ref78 doi: 10.24963/ijcai.2019/98 – ident: ref30 doi: 10.1049/cp.2016.1453 – volume: 2 start-page: 199 issue: 2 year: 2015 ident: ref28 article-title: Depth extraction from video: A survey publication-title: Int. J. Innov. Res. Adv. Eng. contributor: fullname: Chavan – ident: ref167 doi: 10.1109/TVCG.2019.2921336 – ident: ref158 doi: 10.1109/LRA.2020.3017478 – ident: ref96 doi: 10.1007/978-3-030-20870-7_41 – ident: ref138 doi: 10.1007/978-3-030-58529-7_34 – start-page: 1161 volume-title: Proc. NIPS ident: ref13 article-title: Learning depth from single monocular images contributor: fullname: Saxena – start-page: 767 volume-title: Proc. ECCV ident: ref143 article-title: T2Net: Synthetic-to-realistic translation for solving single-image depth estimation tasks contributor: fullname: Zheng – ident: ref81 doi: 10.1109/3DV.2018.00044 – ident: ref71 doi: 10.1109/CVPR46437.2021.00400 – year: 2019 ident: ref117 article-title: DFineNet: Ego-motion estimation and depth refinement from sparse, noisy depth input with RGB guidance publication-title: arXiv:1903.06397 contributor: fullname: Zhang – ident: ref140 doi: 10.1007/978-3-030-20893-6_19 – ident: ref98 doi: 10.1109/ICIP.2019.8803170 – ident: ref39 doi: 10.1109/CVPR.2016.350 – year: 2020 ident: ref119 article-title: Semantically-guided representation learning for self-supervised monocular depth publication-title: arXiv:2002.12319 contributor: fullname: Guizilini – ident: ref152 doi: 10.1109/ICCV.2015.336 – ident: ref129 doi: 10.1109/CVPR.2018.00043 – year: 2019 ident: ref32 article-title: Monocular depth estimation: A survey publication-title: arXiv:1901.09402 contributor: fullname: Bhoi – ident: ref146 doi: 10.1109/WACV.2019.00226 – ident: ref37 doi: 10.1007/978-3-642-33715-4_54 – ident: ref49 doi: 10.1109/CVPR.2017.261 – ident: ref105 doi: 10.1007/978-3-030-01267-0_4 – start-page: 503 volume-title: Proc. CoRL ident: ref160 article-title: Robust semi-supervised monocular depth estimation with reprojected distances contributor: fullname: Guizilini – ident: ref14 doi: 10.1109/3DTV.2010.5506603 – ident: ref125 doi: 10.1109/WACV45572.2020.9093334 – ident: ref31 doi: 10.29121/granthaalayah.v5.i4RACSIT.2017.3342 – ident: ref112 doi: 10.1109/TITS.2017.2750080 – ident: ref121 doi: 10.1109/CVPR42600.2020.00481 – ident: ref36 doi: 10.1109/TPAMI.2008.132 – ident: ref94 doi: 10.1109/ICCV.2019.00181 – ident: ref42 doi: 10.1109/IROS.2012.6385773 – year: 2015 ident: ref19 article-title: Depth extraction from videos using geometric context and occlusion boundaries publication-title: arXiv:1510.07317 contributor: fullname: Hussain Raza – ident: ref53 doi: 10.1109/LRA.2018.2800083 – ident: ref136 doi: 10.1007/978-3-030-58565-5_35 – ident: ref147 doi: 10.1007/978-3-030-01249-6_43 – ident: ref17 doi: 10.1109/CVPR.2014.97 – ident: ref115 doi: 10.1109/CVPR.2019.01003 – year: 2020 ident: ref179 article-title: On deep learning techniques to boost monocular depth estimation for autonomous navigation publication-title: arXiv:2010.06626 contributor: fullname: de Queiroz Mendes – ident: ref132 doi: 10.1109/ICCV.2019.00393 – ident: ref61 doi: 10.1109/IROS45743.2020.9341435 – ident: ref168 doi: 10.1109/ICRA.2017.7989590 – ident: ref118 doi: 10.1109/LRA.2019.2896963 – ident: ref163 doi: 10.1109/TPAMI.2019.2936024 – ident: ref190 doi: 10.1109/ICRA.2019.8794425 – ident: ref177 doi: 10.1109/TMM.2018.2859034 – ident: ref191 doi: 10.1109/ICRA.2014.6907325 – ident: ref155 doi: 10.1109/TIP.2020.2968751 – ident: ref109 doi: 10.1109/TCSVT.2020.2973068 – ident: ref79 doi: 10.1109/CVPR.2015.7298715 – ident: ref196 doi: 10.1109/ICCV.2019.00397 – ident: ref95 doi: 10.1109/CVPR.2019.00570 – ident: ref181 doi: 10.1007/978-3-319-23778-7_14 – year: 2020 ident: ref113 article-title: DepthNet nano: A highly compact self-normalizing neural network for monocular depth estimation publication-title: arXiv:2004.08008 contributor: fullname: Wang – ident: ref58 doi: 10.1109/CVPR.2019.00099 – ident: ref12 doi: 10.1109/TPAMI.2002.1033214 – ident: ref75 doi: 10.1109/ICCV.2015.401 – year: 1970 ident: ref73 article-title: Shape from shading: A method for obtaining the shape of a smooth opaque object from one view contributor: fullname: Horn – ident: ref26 doi: 10.1109/WACV.2019.00116 – ident: ref82 doi: 10.1109/CVPR.2016.90 – year: 2017 ident: ref90 article-title: Attention is all you need publication-title: arXiv:1706.03762 contributor: fullname: Vaswani – ident: ref187 doi: 10.1016/j.robot.2015.11.011 – year: 2018 ident: ref111 article-title: YOLOv3: An incremental improvement publication-title: arXiv:1804.02767 contributor: fullname: Redmon – ident: ref189 doi: 10.1007/978-3-319-10605-2_54 – ident: ref99 doi: 10.1109/ITSC.2019.8917177 – ident: ref40 doi: 10.1109/CVPR.2016.470 – year: 2019 ident: ref87 article-title: From big to small: Multi-scale local planar guidance for monocular depth estimation publication-title: arXiv:1907.10326 contributor: fullname: Han Lee – ident: ref164 doi: 10.1007/s10846-020-01205-0 – ident: ref195 doi: 10.1109/TITS.2014.2308977 – ident: ref110 doi: 10.1109/CVPR.2015.7298965 – ident: ref127 doi: 10.1109/CVPR42600.2020.00069 – ident: ref162 doi: 10.1109/ICASSP.2019.8683235 – ident: ref23 doi: 10.1109/TCSVT.2017.2740321 – ident: ref126 doi: 10.1109/IROS45743.2020.9341448 – ident: ref170 doi: 10.1109/ICRA.2019.8794404 – ident: ref10 doi: 10.1109/CVPR.2016.584 – volume: 4 start-page: 235 issue: 11 year: 2015 ident: ref29 article-title: A review on depth estimation for computer vision applications publication-title: Positions contributor: fullname: Mohan – ident: ref169 doi: 10.1109/3DV.2018.00017 – ident: ref65 doi: 10.1109/CVPR.2017.699 – ident: ref60 doi: 10.1109/CVPRW.2019.00081 – ident: ref72 doi: 10.7551/mitpress/3877.003.0009 – ident: ref18 doi: 10.1109/TPAMI.2014.2316835 – ident: ref139 doi: 10.1109/IROS45743.2020.9340802 – year: 2019 ident: ref57 article-title: A large RGB-D dataset for semi-supervised monocular depth estimation publication-title: arXiv:1904.10230 contributor: fullname: Cho – ident: ref194 doi: 10.1109/ICRA.2012.6224734 – ident: ref178 doi: 10.1007/978-3-030-01237-3_50 – ident: ref159 doi: 10.1109/ICAR46387.2019.8981652 – ident: ref3 doi: 10.1007/978-3-030-01252-6_2 – ident: ref89 doi: 10.1109/CVPR.2018.00214 – ident: ref92 doi: 10.1109/CVPR42600.2020.00054 – ident: ref154 doi: 10.1109/CVPR.2018.00212 – ident: ref9 doi: 10.1007/978-3-319-14249-4_71 – ident: ref63 doi: 10.1109/TPAMI.2015.2505283 – volume: 2 start-page: 2 issue: 3 year: 1978 ident: ref74 article-title: Recovering intrinsic scene characteristics publication-title: Comput. Vis. Syst. contributor: fullname: Barrow – ident: ref184 doi: 10.1109/ICRA40945.2020.9197481 – ident: ref76 doi: 10.1109/ICRA.2018.8460184 – ident: ref120 doi: 10.1109/CVPR42600.2020.00256 – ident: ref22 doi: 10.1109/3DV.2016.32 – start-page: 730 volume-title: Proc. NIPS ident: ref44 article-title: Single-image depth perception in the wild contributor: fullname: Chen – ident: ref104 doi: 10.1109/IVS.2018.8500683 – ident: ref15 doi: 10.1109/CVPR.2010.5539823 – ident: ref83 doi: 10.1109/CVPR.2017.243 – ident: ref56 doi: 10.1007/s13042-018-0891-9 – year: 2017 ident: ref46 article-title: Joint 2D-3D-Semantic data for indoor scene understanding publication-title: arXiv:1702.01105 contributor: fullname: Armeni – year: 2019 ident: ref123 article-title: Unsupervised scale-consistent depth and ego-motion learning from monocular video publication-title: arXiv:1908.10553 contributor: fullname: Bian – ident: ref197 doi: 10.1109/ICCV.2019.00227 – ident: ref48 doi: 10.1109/3DV.2017.00081 – ident: ref54 doi: 10.1007/978-3-030-11012-3_4 – ident: ref124 doi: 10.1109/CVPR.2019.00343 – ident: ref100 doi: 10.1109/TMM.2019.2941776 – ident: ref131 doi: 10.1109/ICIP.2019.8803544 – ident: ref148 doi: 10.1109/TRO.2017.2705103 – ident: ref70 doi: 10.1109/IROS.2018.8594243 – ident: ref180 doi: 10.1145/1102351.1102426 – ident: ref84 doi: 10.1109/CVPR.2018.00745 – ident: ref193 doi: 10.1109/ICRA.2012.6224636 – ident: ref2 doi: 10.1109/TITS.2019.2955598 – ident: ref47 doi: 10.1109/CVPR.2017.272 – year: 2017 ident: ref149 article-title: SfM-net: Learning of structure and motion from video publication-title: arXiv:1704.07804 contributor: fullname: Vijayanarasimhan – year: 2017 ident: ref173 article-title: MobileNets: Efficient convolutional neural networks for mobile vision applications publication-title: arXiv:1704.04861 contributor: fullname: Howard – volume: 5 start-page: 1760 year: 2019 ident: ref34 article-title: Literature review on depth estimation using a single image publication-title: Int. J. Advance Res., Ideas Innov. Technol. contributor: fullname: Vyshna – ident: ref142 doi: 10.1109/CVPR.2018.00296 – year: 2019 ident: ref137 article-title: Edge-guided occlusion fading reduction for a light-weighted self-supervised monocular depth estimation publication-title: arXiv:1911.11705 contributor: fullname: Peng – ident: ref175 doi: 10.1109/ACCESS.2020.3000773 – ident: ref106 doi: 10.1109/CVPR.2018.00037 – ident: ref153 doi: 10.1007/978-3-030-01228-1_3 – ident: ref182 doi: 10.1109/ICRA.2017.7989752 – ident: ref80 doi: 10.1007/978-3-030-01225-0_3 – ident: ref43 doi: 10.1109/CVPR.2015.7298655 – ident: ref69 doi: 10.1109/IROS.2018.8593814 – start-page: 243 volume-title: Proc. IMAVCC ident: ref186 article-title: Real-time landing zone detection for UAVs using single aerial images contributor: fullname: RoJas-Perez – year: 2019 ident: ref150 article-title: Self-supervised object motion and depth estimation from video publication-title: arXiv:1912.04250 contributor: fullname: Dai – ident: ref108 doi: 10.1109/ICIP.2019.8803100 – year: 2020 ident: ref91 article-title: An image is worth 16×16 words: Transformers for image recognition at scale publication-title: arXiv:2010.11929 contributor: fullname: Dosovitskiy – year: 2019 ident: ref33 article-title: A survey on deep learning architectures for image-based depth reconstruction publication-title: arXiv:1906.06113 contributor: fullname: Laga – ident: ref1 doi: 10.1109/CVPR.2017.695 – start-page: 505 volume-title: Proc. ISER ident: ref192 article-title: Vision-based obstacle avoidance for micro air vehicles using an egocylindrical depth map contributor: fullname: Brockers – ident: ref24 doi: 10.1109/CVPR.2017.25 – ident: ref188 doi: 10.1109/ROBIO.2015.7418959 – year: 2018 ident: ref86 article-title: High quality monocular depth estimation via transfer learning publication-title: arXiv:1812.11941 contributor: fullname: Alhashim – ident: ref103 doi: 10.1109/ICRA.2017.7989537 – ident: ref114 doi: 10.1109/CVPR.2018.00594 – ident: ref45 doi: 10.1109/WACV.2016.7477636 – ident: ref97 doi: 10.1007/s13042-020-01251-y – start-page: 2366 volume-title: Proc. NIPS ident: ref20 article-title: Depth map prediction from a single image using a multi-scale deep network contributor: fullname: Eigen – ident: ref130 doi: 10.1609/aaai.v33i01.33018001 – ident: ref64 doi: 10.1007/978-3-319-46484-8_45 – ident: ref151 doi: 10.1109/CVPR42600.2020.00917 – ident: ref8 doi: 10.1007/11861898_48 – ident: ref50 doi: 10.1109/ICCV.2017.292 – ident: ref62 doi: 10.1109/CVPR42600.2020.00097 – ident: ref35 doi: 10.1109/DSC50466.2020.00028 – ident: ref161 doi: 10.1109/CVPR42600.2020.00339 – ident: ref21 doi: 10.1109/ICCV.2015.304 – ident: ref101 doi: 10.1109/ICASSP.2018.8462433 – ident: ref157 doi: 10.1016/j.isprsjprs.2020.06.004 – ident: ref107 doi: 10.1109/ICRA.2019.8794220 – ident: ref116 doi: 10.1109/ICRA.2019.8793637 – ident: ref27 doi: 10.1109/ICRA.2019.8794182 – ident: ref4 doi: 10.1109/ICARCV.2016.7838832 – ident: ref38 doi: 10.1177/0278364913491297 – ident: ref88 doi: 10.1109/ICCV.2019.00578 – ident: ref172 doi: 10.1109/ACCESS.2020.2983149 – ident: ref6 doi: 10.1109/ICCP.2015.7312626 – ident: ref171 doi: 10.1109/ICCV.2019.01012 – ident: ref176 doi: 10.1109/ICCV.2017.625 – start-page: 2453 volume-title: Proc. IEEE Int. Conf. Comput. Vis. Workshops (ICCVW) ident: ref11 article-title: Accurate depth map estimation from small motions contributor: fullname: Javidnia – ident: ref55 doi: 10.1007/978-3-030-01231-1_28 – ident: ref52 doi: 10.1109/CVPR.2018.00218 – ident: ref166 doi: 10.1109/TPAMI.2019.2940948 – ident: ref135 doi: 10.1109/3DV.2019.00081 – year: 2019 ident: ref59 article-title: DIODE: A dense indoor and outdoor DEpth dataset publication-title: arXiv:1908.00463 contributor: fullname: Vasiljevic – ident: ref5 doi: 10.1109/CVPR.2011.5995480 – ident: ref128 doi: 10.1109/3DV.2018.00073 – ident: ref67 doi: 10.1109/CVPR.2017.238 – ident: ref133 doi: 10.1109/CVPR.2019.01252 – ident: ref16 doi: 10.1109/CVPR.2014.19 – ident: ref25 doi: 10.1016/j.patcog.2018.05.029 – ident: ref122 doi: 10.1007/978-3-030-01252-6_30 – ident: ref93 doi: 10.1109/CVPRW.2018.00066 – ident: ref183 doi: 10.1109/CCIOT48581.2019.8980350 – ident: ref7 doi: 10.1109/IVS.2014.6856419 – ident: ref51 doi: 10.1109/CVPR.2017.28 – ident: ref85 doi: 10.1109/IROS.2016.7759632 – ident: ref66 doi: 10.1109/CVPR.2017.700 – ident: ref165 doi: 10.1109/ICCVW.2017.108 |
SSID | ssj0014511 |
Score | 2.5722253 |
Snippet | As an essential component for many autonomous driving and robotic activities such as ego-motion estimation, obstacle avoidance and scene understanding,... |
SourceID | proquest crossref ieee |
SourceType | Aggregation Database Publisher |
StartPage | 16940 |
SubjectTerms | Cameras Computer vision depth prediction Estimation Feature extraction Monocular depth estimation Motion simulation Obstacle avoidance Performance evaluation Robotics Robots Scene analysis single image depth estimation Structure from motion survey Task analysis Three-dimensional displays |
Title | Towards Real-Time Monocular Depth Estimation for Robotics: A Survey |
URI | https://ieeexplore.ieee.org/document/9750985 https://www.proquest.com/docview/2723901470 |
Volume | 23 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3NT8IwFH8BTnrwC40omh48GYtdWdfOG0EImuBBIfG2rF0bExNGYJjoX2_bDeLXwdsOXdO8177f7_V9FOBCSgvqcaaxZQcZDiMhcUpNgGWUcUEU5VHqCpzHD9FoGt4_s-caXG1qYbTWPvlMd9ynj-VnuVq5q7Lr2MGbYHWoC0LLWq1NxMD12fK9UWmIY8LWEcyAxNeTu8mT9QQptQ5q5CD0Gwb5R1V-WWIPL8NdGK8XVmaVvHZWheyojx89G_-78j3YqXgm6pUbYx9qenYA21-6DzahP_Eps0v0aMkidrUgyJ7w3Cemols9L17QwBqAsrYRWXKLHnOZu7bON6iHnlaLN_1-CNPhYNIf4epNBawssBeYW3sWMG5YFKdBajQ1OmPMcKNjITLLlhSjKgoy6o6nMJY8xinpKq0ok4ZbvngEjVk-08eArG8Yka6bgrMwJUJyKi19yhQ1KRWatOByLeVkXrbOSLzLQeLEqSRxKkkqlbSg6aS2GVgJrAXttV6S6nAtE8qpu6kJOTn5-69T2HJzlzl3bWgUi5U-s9yhkOd-03wCKom_zg |
link.rule.ids | 315,786,790,802,27955,27956,55107 |
linkProvider | IEEE |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV07T8MwED5BGYCBN6I8PTAhXBwTxwkb4qHyKAMEiS2KHVtISE1FUyT49ZydtOI1sGVIHOvOvu8738MA-0ohqCeFocgOChpGsaI5twFVUSFjprmMclfg3LuLuo_h9ZN4moLDSS2MMcYnn5mOe_Sx_KLUI3dUdpQ4eIvFNMwgzjNZV2tNYgau05bvjspDmjAxjmEGLDlKr9IH9AU5Rxc1ciD6DYX8tSq_bLEHmMtF6I2nVueVvHRGlerojx9dG_879yVYaJgmOa2XxjJMmf4KzH_pP7gKZ6lPmh2Se6SL1FWDENzjpU9NJedmUD2TCzQBdXUjQXpL7ktVusbOJ-SUPIxe38z7GjxeXqRnXdrcqkA1QntFJVq0QEgroiQPcmu4NYUQVlqTxHGBfEkLrqOg4G6DxhbpY5KzY200F8pKZIzr0OqXfbMBBL3DiB27IaQIcxYryRUSqEJzm_PYsDYcjKWcDermGZl3OliSOZVkTiVZo5I2rDqpTV5sBNaG7bFesmZ7DTMuuTurCSXb_PurPZjtpr3b7Pbq7mYL5tx_6gy8bWhVryOzg0yiUrt-AX0CdaTDIg |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Towards+Real-Time+Monocular+Depth+Estimation+for+Robotics%3A+A+Survey&rft.jtitle=IEEE+transactions+on+intelligent+transportation+systems&rft.au=Dong%2C+Xingshuai&rft.au=Garratt%2C+Matthew+A.&rft.au=Anavatti%2C+Sreenatha+G.&rft.au=Abbass%2C+Hussein+A.&rft.date=2022-10-01&rft.issn=1524-9050&rft.eissn=1558-0016&rft.volume=23&rft.issue=10&rft.spage=16940&rft.epage=16961&rft_id=info:doi/10.1109%2FTITS.2022.3160741&rft.externalDBID=n%2Fa&rft.externalDocID=10_1109_TITS_2022_3160741 |
thumbnail_l | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=1524-9050&client=summon |
thumbnail_m | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=1524-9050&client=summon |
thumbnail_s | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=1524-9050&client=summon |