Continuous Semi-autonomous Prosthesis Control Using a Depth Sensor on the Hand
Modern myoelectric prostheses can perform multiple functions (e.g., several grasp types and wrist rotation) but their intuitive control by the user is still an open challenge. It has been recently demonstrated that semi-autonomous control can allow the subjects to operate complex prostheses effectiv...
Saved in:
Published in | Frontiers in neurorobotics Vol. 16; p. 814973 |
---|---|
Main Authors | , |
Format | Journal Article |
Language | English |
Published |
Switzerland
Frontiers Research Foundation
25.03.2022
Frontiers Media S.A |
Subjects | |
Online Access | Get full text |
ISSN | 1662-5218 1662-5218 |
DOI | 10.3389/fnbot.2022.814973 |
Cover
Loading…
Abstract | Modern myoelectric prostheses can perform multiple functions (e.g., several grasp types and wrist rotation) but their intuitive control by the user is still an open challenge. It has been recently demonstrated that semi-autonomous control can allow the subjects to operate complex prostheses effectively; however, this approach often requires placing sensors on the user. The present study proposes a system for semi-autonomous control of a myoelectric prosthesis that requires a single depth sensor placed on the dorsal side of the hand. The system automatically pre-shapes the hand (grasp type, size, and wrist rotation) and allows the user to grasp objects of different shapes, sizes and orientations, placed individually or within cluttered scenes. The system “reacts” to the side from which the object is approached, and enables the user to target not only the whole object but also an object part. Another unique aspect of the system is that it relies on online interaction between the user and the prosthesis; the system reacts continuously on the targets that are in its focus, while the user interprets the movement of the prosthesis to adjust aiming. Experimental assessment was conducted in ten able-bodied participants to evaluate the feasibility and the impact of training on prosthesis-user interaction. The subjects used the system to grasp a set of objects individually (Phase I) and in cluttered scenarios (Phase II), while the time to accomplish the task (TAT) was used as the performance metric. In both phases, the TAT improved significantly across blocks. Some targets (objects and/or their parts) were more challenging, requiring thus significantly more time to handle, but all objects and scenes were successfully accomplished by all subjects. The assessment therefore demonstrated that the system is indeed robust and effective, and that the subjects could successfully learn how to aim with the system after a brief training. This is an important step toward the development of a self-contained semi-autonomous system convenient for clinical applications. |
---|---|
AbstractList | Modern myoelectric prostheses can perform multiple functions (e.g., several grasp types and wrist rotation) but their intuitive control by the user is still an open challenge. It has been recently demonstrated that semi-autonomous control can allow the subjects to operate complex prostheses effectively; however, this approach often requires placing sensors on the user. The present study proposes a system for semi-autonomous control of a myoelectric prosthesis that requires a single depth sensor placed on the dorsal side of the hand. The system automatically pre-shapes the hand (grasp type, size, and wrist rotation) and allows the user to grasp objects of different shapes, sizes and orientations, placed individually or within cluttered scenes. The system “reacts” to the side from which the object is approached, and enables the user to target not only the whole object but also an object part. Another unique aspect of the system is that it relies on online interaction between the user and the prosthesis; the system reacts continuously on the targets that are in its focus, while the user interprets the movement of the prosthesis to adjust aiming. Experimental assessment was conducted in ten able-bodied participants to evaluate the feasibility and the impact of training on prosthesis-user interaction. The subjects used the system to grasp a set of objects individually (Phase I) and in cluttered scenarios (Phase II), while the time to accomplish the task (TAT) was used as the performance metric. In both phases, the TAT improved significantly across blocks. Some targets (objects and/or their parts) were more challenging, requiring thus significantly more time to handle, but all objects and scenes were successfully accomplished by all subjects. The assessment therefore demonstrated that the system is indeed robust and effective, and that the subjects could successfully learn how to aim with the system after a brief training. This is an important step toward the development of a self-contained semi-autonomous system convenient for clinical applications. Modern myoelectric prostheses can perform multiple functions (e.g., several grasp types and wrist rotation) but their intuitive control by the user is still an open challenge. It has been recently demonstrated that semi-autonomous control can allow the subjects to operate complex prostheses effectively; however, this approach often requires placing sensors on the user. The present study proposes a system for semi-autonomous control of a myoelectric prosthesis that requires a single depth sensor placed on the dorsal side of the hand. The system automatically pre-shapes the hand (grasp type, size, and wrist rotation) and allows the user to grasp objects of different shapes, sizes and orientations, placed individually or within cluttered scenes. The system "reacts" to the side from which the object is approached, and enables the user to target not only the whole object but also an object part. Another unique aspect of the system is that it relies on online interaction between the user and the prosthesis; the system reacts continuously on the targets that are in its focus, while the user interprets the movement of the prosthesis to adjust aiming. Experimental assessment was conducted in ten able-bodied participants to evaluate the feasibility and the impact of training on prosthesis-user interaction. The subjects used the system to grasp a set of objects individually (Phase I) and in cluttered scenarios (Phase II), while the time to accomplish the task (TAT) was used as the performance metric. In both phases, the TAT improved significantly across blocks. Some targets (objects and/or their parts) were more challenging, requiring thus significantly more time to handle, but all objects and scenes were successfully accomplished by all subjects. The assessment therefore demonstrated that the system is indeed robust and effective, and that the subjects could successfully learn how to aim with the system after a brief training. This is an important step toward the development of a self-contained semi-autonomous system convenient for clinical applications.Modern myoelectric prostheses can perform multiple functions (e.g., several grasp types and wrist rotation) but their intuitive control by the user is still an open challenge. It has been recently demonstrated that semi-autonomous control can allow the subjects to operate complex prostheses effectively; however, this approach often requires placing sensors on the user. The present study proposes a system for semi-autonomous control of a myoelectric prosthesis that requires a single depth sensor placed on the dorsal side of the hand. The system automatically pre-shapes the hand (grasp type, size, and wrist rotation) and allows the user to grasp objects of different shapes, sizes and orientations, placed individually or within cluttered scenes. The system "reacts" to the side from which the object is approached, and enables the user to target not only the whole object but also an object part. Another unique aspect of the system is that it relies on online interaction between the user and the prosthesis; the system reacts continuously on the targets that are in its focus, while the user interprets the movement of the prosthesis to adjust aiming. Experimental assessment was conducted in ten able-bodied participants to evaluate the feasibility and the impact of training on prosthesis-user interaction. The subjects used the system to grasp a set of objects individually (Phase I) and in cluttered scenarios (Phase II), while the time to accomplish the task (TAT) was used as the performance metric. In both phases, the TAT improved significantly across blocks. Some targets (objects and/or their parts) were more challenging, requiring thus significantly more time to handle, but all objects and scenes were successfully accomplished by all subjects. The assessment therefore demonstrated that the system is indeed robust and effective, and that the subjects could successfully learn how to aim with the system after a brief training. This is an important step toward the development of a self-contained semi-autonomous system convenient for clinical applications. Modern myoelectric prostheses can perform multiple functions (e.g., several grasp types and wrist rotation) but their intuitive control by the user is still an open challenge. It has been recently demonstrated that semi-autonomous control can allow the subjects to operate a complex prostheses effectively; however, this approach often requires placing the sensors on the user. The present study proposes a system for semi-autonomous control of a myoelectric prosthesis that requires a single depth sensor placed on the dorsal side of the hand. The system automatically pre-shapes the hand (grasp type, size and wrist rotation) and allows the user to grasp objects of different shapes, sizes and orientations, placed individually or within cluttered scenes. The system “reacts” to the side from which the object is approached, and enables the user to target not only the whole object but also an object part. Another unique aspect of the system is that it relies on online interaction between user and the prosthesis; the system reacts continuously on the targets that are in its focus, while the user interprets the movement of the prosthesis to adjust the aiming. Experimental assessment was conducted in ten able-bodied participants to evaluate the feasibility and the impact of training on prosthesis-user interaction. The subjects used the system to grasp a set of objects individually (Phase I) and in cluttered scenarios (Phase II), while the time to accomplish the task (TAT) was used as the performance metric. In both phases, the TAT improved significantly across blocks. Some targets (objects and/or their parts) were more challenging, requiring thus significantly more time to handle, but all objects and scenes were successfully accomplished by all subjects. The assessment therefore demonstrated that the system is indeed robust and effective, and that the subjects could successfully learn how to aim with the system after a brief training. This is an important step towards the development of a self-contained semi-autonomous system convenient for clinical applications. |
Author | Dosen, Strahinja Castro, Miguel Nobre |
AuthorAffiliation | Neurorehabilitation Systems, Department of Health Science and Technology, Aalborg University , Aalborg , Denmark |
AuthorAffiliation_xml | – name: Neurorehabilitation Systems, Department of Health Science and Technology, Aalborg University , Aalborg , Denmark |
Author_xml | – sequence: 1 givenname: Miguel Nobre surname: Castro fullname: Castro, Miguel Nobre – sequence: 2 givenname: Strahinja surname: Dosen fullname: Dosen, Strahinja |
BackLink | https://www.ncbi.nlm.nih.gov/pubmed/35401136$$D View this record in MEDLINE/PubMed |
BookMark | eNp1kl1rFDEUhoNU7If-AG9kwBtvZs33ZG4EWVtbKCpor0OSSXazzCRrkin47810W2kLXuXrfR9OznlPwVGIwQLwFsEVIaL_6IKOZYUhxiuBaN-RF-AEcY5bhpE4erQ_Bqc57yDkmDPxChwTRiFChJ-Ab-sYig9znHPz006-VXOJIU7L-UeKuWxt9rlZVCmOzU32YdOo5ovdl201hBxTE0NTZc2lCsNr8NKpMds39-sZuLk4_7W-bK-_f71af75uDe1JaTXVrudi6KBj0GAqlGKk67FzBg4KC4gGwjWhmilt4WAoQaRDlA3OYei4IGfg6sAdotrJffKTSn9kVF7eXcS0kSoVb0YrafVjJzDGxlFNqTKaQqwIFwj2XOnK-nRg7Wc92cHY-lU1PoE-fQl-KzfxVope1J53FfDhHpDi79nmIiefjR1HFWzto8Sc9pgRRGGVvn8m3cU5hdqqRUWgQIwuv3v3uKJ_pTyMrQq6g8DUEeVknTS-qOKXKSk_SgTlEhB5FxC5BEQeAlKd6JnzAf5_z18dSL78 |
CitedBy_id | crossref_primary_10_1002_rcs_2617 crossref_primary_10_1038_s41597_023_02313_w crossref_primary_10_3390_robotics12060152 crossref_primary_10_1109_LRA_2024_3398563 crossref_primary_10_1109_TMRB_2023_3292419 crossref_primary_10_1126_scirobotics_adl0085 crossref_primary_10_1109_MSP_2024_3401403 crossref_primary_10_1146_annurev_bioeng_110222_095816 crossref_primary_10_1109_JSEN_2023_3308615 crossref_primary_10_1109_TMRB_2024_3377530 crossref_primary_10_3390_biomimetics8030328 crossref_primary_10_1109_TNSRE_2023_3295060 crossref_primary_10_3390_biomimetics8020250 |
Cites_doi | 10.1109/TRO.2020.3047013 10.3109/17483107.2011.635405 10.3390/s20216097 10.1145/358669.358692 10.1016/j.bspc.2007.07.009 10.1109/TNSRE.2014.2305111 10.1109/ACCESS.2018.2791583 10.1109/ACCESS.2021.3109733 10.3389/fnins.2020.00345 10.1109/TCYB.2020.2996960. 10.1109/TNSRE.2012.2196711 10.1109/BIOROB.2018.8487622 10.1126/scirobotics.abb0467 10.3390/s19204596 10.1038/srep36571 10.2147/MDER.S91102 10.1007/s00221-018-5441-x 10.1109/TNSRE.2020.3007625 10.1186/1743-0003-7-42 10.1088/1741-2560/12/6/066022 10.1080/09638288.2020.1866684. 10.1088/1741-2552/aa6802 10.1080/03093640600994581 10.1126/scirobotics.aat3630 10.1088/1741-2560/11/4/046001 10.1109/TNSRE.2014.2305520 10.3389/fnbot.2016.00009 10.1109/MSP.2012.2203480 10.1109/TNSRE.2014.2305097 10.1109/ICRA.2011.5980567 10.1038/s41551-021-00732-x. |
ContentType | Journal Article |
Copyright | Copyright © 2022 Castro and Dosen. 2022. This work is licensed under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. Copyright © 2022 Castro and Dosen. 2022 Castro and Dosen |
Copyright_xml | – notice: Copyright © 2022 Castro and Dosen. – notice: 2022. This work is licensed under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. – notice: Copyright © 2022 Castro and Dosen. 2022 Castro and Dosen |
DBID | AAYXX CITATION NPM 3V. 7XB 88I 8FE 8FH 8FK ABUWG AFKRA AZQEC BBNVY BENPR BHPHI CCPQU DWQXO GNUQQ HCIFZ LK8 M2P M7P PHGZM PHGZT PIMPY PKEHL PQEST PQGLB PQQKQ PQUKI PRINS Q9U 7X8 5PM DOA |
DOI | 10.3389/fnbot.2022.814973 |
DatabaseName | CrossRef PubMed ProQuest Central (Corporate) ProQuest Central (purchase pre-March 2016) Science Database (Alumni Edition) ProQuest SciTech Collection ProQuest Natural Science Collection ProQuest Central (Alumni) (purchase pre-March 2016) ProQuest Central (Alumni) ProQuest Central UK/Ireland ProQuest Central Essentials Biological Science Collection ProQuest Central Database Suite (ProQuest) Natural Science Collection ProQuest One ProQuest Central ProQuest Central Student SciTech Collection (ProQuest) Biological Sciences Science Database (ProQuest) Biological Science Database (ProQuest) ProQuest Central Premium ProQuest One Academic (New) Publicly Available Content Database ProQuest One Academic Middle East (New) ProQuest One Academic Eastern Edition (DO NOT USE) ProQuest One Applied & Life Sciences ProQuest One Academic ProQuest One Academic UKI Edition ProQuest Central China ProQuest Central Basic MEDLINE - Academic PubMed Central (Full Participant titles) DOAJ: Directory of Open Access Journal (DOAJ) |
DatabaseTitle | CrossRef PubMed Publicly Available Content Database ProQuest Central Student ProQuest One Academic Middle East (New) ProQuest Central Essentials ProQuest Central (Alumni Edition) SciTech Premium Collection ProQuest One Community College ProQuest Natural Science Collection ProQuest Central China ProQuest Central ProQuest One Applied & Life Sciences Natural Science Collection ProQuest Central Korea Biological Science Collection ProQuest Central (New) ProQuest Science Journals (Alumni Edition) ProQuest Biological Science Collection ProQuest Central Basic ProQuest Science Journals ProQuest One Academic Eastern Edition Biological Science Database ProQuest SciTech Collection ProQuest One Academic UKI Edition ProQuest One Academic ProQuest One Academic (New) ProQuest Central (Alumni) MEDLINE - Academic |
DatabaseTitleList | MEDLINE - Academic PubMed CrossRef Publicly Available Content Database |
Database_xml | – sequence: 1 dbid: DOA name: DOAJ Directory of Open Access Journals url: https://www.doaj.org/ sourceTypes: Open Website – sequence: 2 dbid: NPM name: PubMed url: https://proxy.k.utb.cz/login?url=http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?db=PubMed sourceTypes: Index Database – sequence: 3 dbid: BENPR name: ProQuest Central url: https://www.proquest.com/central sourceTypes: Aggregation Database |
DeliveryMethod | fulltext_linktorsrc |
Discipline | Engineering |
EISSN | 1662-5218 |
ExternalDocumentID | oai_doaj_org_article_40dc2f8222cf4b44acb402a3681096ab PMC8989737 35401136 10_3389_fnbot_2022_814973 |
Genre | Journal Article |
GeographicLocations | Germany |
GeographicLocations_xml | – name: Germany |
GrantInformation_xml | – fundername: ; grantid: 8022-00243A |
GroupedDBID | --- 29H 2WC 53G 5GY 5VS 8FE 8FH 9T4 AAFWJ AAKPC AAYXX ABUWG ACGFS ACXDI ADBBV ADDVE ADMLS ADRAZ AEGXH AENEX AFKRA AFPKN ALMA_UNASSIGNED_HOLDINGS AOIJS ARCSS AZQEC BAWUL BBNVY BCNDV BENPR BHPHI BPHCQ CITATION CS3 DIK E3Z F5P GROUPED_DOAJ GX1 HCIFZ HYE KQ8 LK8 M2P M48 M7P M~E O5R O5S OK1 OVT PGMZT PIMPY PQQKQ PROAC RNS RPM TR2 88I C1A CCPQU DWQXO GNUQQ IPNFZ NPM PHGZT RIG 3V. 7XB 8FK PHGZM PKEHL PQEST PQGLB PQUKI PRINS Q9U 7X8 5PM |
ID | FETCH-LOGICAL-c493t-b4bf968d70f50c248aa53792ffc0da2801d36b34b5abe0dc43137145dff20f683 |
IEDL.DBID | M48 |
ISSN | 1662-5218 |
IngestDate | Wed Aug 27 01:00:11 EDT 2025 Thu Aug 21 14:13:04 EDT 2025 Fri Jul 11 06:58:13 EDT 2025 Mon Jun 30 09:41:06 EDT 2025 Thu Apr 03 07:00:47 EDT 2025 Tue Jul 01 02:32:21 EDT 2025 Thu Apr 24 23:03:40 EDT 2025 |
IsDoiOpenAccess | true |
IsOpenAccess | true |
IsPeerReviewed | true |
IsScholarly | true |
Keywords | grasping myoelectric hand prosthesis point cloud processing object segmentation semi-autonomous control computer vision |
Language | English |
License | Copyright © 2022 Castro and Dosen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-c493t-b4bf968d70f50c248aa53792ffc0da2801d36b34b5abe0dc43137145dff20f683 |
Notes | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 14 content type line 23 This article was submitted to Control Using a Depth Sensor on the Hand, a section of the journal Frontiers in Neurorobotics Reviewed by: Enzo Mastinu, Chalmers University of Technology, Sweden; Toshihiro Kawase, Tokyo Medical and Dental University, Japan Edited by: Feihu Zhang, Northwestern Polytechnical University, China |
OpenAccessLink | http://journals.scholarsportal.info/openUrl.xqy?doi=10.3389/fnbot.2022.814973 |
PMID | 35401136 |
PQID | 2643081548 |
PQPubID | 4424403 |
ParticipantIDs | doaj_primary_oai_doaj_org_article_40dc2f8222cf4b44acb402a3681096ab pubmedcentral_primary_oai_pubmedcentral_nih_gov_8989737 proquest_miscellaneous_2649253140 proquest_journals_2643081548 pubmed_primary_35401136 crossref_citationtrail_10_3389_fnbot_2022_814973 crossref_primary_10_3389_fnbot_2022_814973 |
ProviderPackageCode | CITATION AAYXX |
PublicationCentury | 2000 |
PublicationDate | 2022-03-25 |
PublicationDateYYYYMMDD | 2022-03-25 |
PublicationDate_xml | – month: 03 year: 2022 text: 2022-03-25 day: 25 |
PublicationDecade | 2020 |
PublicationPlace | Switzerland |
PublicationPlace_xml | – name: Switzerland – name: Lausanne |
PublicationTitle | Frontiers in neurorobotics |
PublicationTitleAlternate | Front Neurorobot |
PublicationYear | 2022 |
Publisher | Frontiers Research Foundation Frontiers Media S.A |
Publisher_xml | – name: Frontiers Research Foundation – name: Frontiers Media S.A |
References | Došen (B6) 2010; 7 Ragusa (B27) 2021; 9 Farina (B8) 2021 Østlie (B25) 2012; 7 Ortiz-Catalan (B24) 2014; 22 Farina (B7) 2014; 22 Parajuli (B26) 2019; 19 Shi (B31) 2020; 28 Geethanjali (B12) 2016; 9 Markovic (B21) 2015; 12 Weiner (B34) 2018 Biddiss (B3) 2007; 31 Zhong (B36) 2020 Jiang (B18) 2012; 29 Laffranchi (B19) 2020; 5 Salminger (B29) 2020 Ghazaei (B15) 2019 Hahne (B17) 2018; 3 Rusu (B28) 2011 Criswell (B5) 2010 Gardner (B11) 2020; 20 Choudhry (B4) 2018 Asghari Oskoei (B1) 2007; 2 Atzori (B2) 2016; 10 Fischler (B9) 1981; 24 Mouchoux (B23) 2021; 37 Stein (B32) 2014 Fougner (B10) 2012; 20 Stephens-Fripp (B33) 2018; 6 Yang (B35) 2019; 237 Geng (B13) 2016; 6 Markovic (B20) 2014; 11 Maymo (B22) 2018 Ghazaei (B14) 2017; 14 Sensinger (B30) 2020; 14 Hahne (B16) 2014; 22 |
References_xml | – volume: 37 start-page: 1298 year: 2021 ident: B23 article-title: Artificial perception and semiautonomous control in myoelectric hand prostheses increases performance and decreases effort publication-title: IEEE Trans. Robot doi: 10.1109/TRO.2020.3047013 – start-page: 3328 volume-title: 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems year: 2018 ident: B34 article-title: “The kit prosthetic hand: design and control,” – volume: 7 start-page: 294 year: 2012 ident: B25 article-title: Prosthesis rejection in acquired major upper-limb amputees: a population-based survey publication-title: Disabil. Rehabil doi: 10.3109/17483107.2011.635405 – volume: 20 start-page: 6097 year: 2020 ident: B11 article-title: A multimodal intention detection sensor suite for shared autonomy of upper-limb robotic prostheses publication-title: Sensors doi: 10.3390/s20216097 – volume: 24 start-page: 381 year: 1981 ident: B9 article-title: Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography publication-title: Commun. ACM doi: 10.1145/358669.358692 – volume: 2 start-page: 275 year: 2007 ident: B1 article-title: Myoelectric control systems—a survey publication-title: Biomed. Signal Process. Control doi: 10.1016/j.bspc.2007.07.009 – volume: 22 start-page: 797 year: 2014 ident: B7 article-title: The extraction of neural information from the surface EMG for the control of upper-limb prostheses: emerging avenues and challenges publication-title: IEEE Trans. Neural Syst. Rehabil. Eng doi: 10.1109/TNSRE.2014.2305111 – volume: 6 start-page: 6878 year: 2018 ident: B33 article-title: A review of non-invasive sensory feedback methods for transradial prosthetic hands publication-title: IEEE Access doi: 10.1109/ACCESS.2018.2791583 – volume: 9 start-page: 123178 year: 2021 ident: B27 article-title: Hardware-aware affordance detection for application in portable embedded systems publication-title: IEEE Access doi: 10.1109/ACCESS.2021.3109733 – volume: 14 start-page: 345 year: 2020 ident: B30 article-title: A review of sensory feedback in upper-limb prostheses from the perspective of human motor control publication-title: Front. Neurosci doi: 10.3389/fnins.2020.00345 – start-page: 1 year: 2020 ident: B36 article-title: Reliable vision-based grasping target recognition for upper limb prostheses publication-title: IEEE Trans. Cybern doi: 10.1109/TCYB.2020.2996960. – volume: 20 start-page: 663 year: 2012 ident: B10 article-title: Control of upper limb prostheses: Terminology and proportional myoelectric controla review publication-title: IEEE Trans. Neural Syst. Rehabil. Eng doi: 10.1109/TNSRE.2012.2196711 – start-page: 207 volume-title: 2018 7th IEEE International Conference on Biomedical Robotics and Biomechatronics year: 2018 ident: B22 article-title: “Fastorient: lightweight computer vision for wrist control in assistive robotic grasping,” doi: 10.1109/BIOROB.2018.8487622 – volume: 5 year: 2020 ident: B19 article-title: The Hannes hand prosthesis replicates the key biological properties of the human hand publication-title: Sci. Robot doi: 10.1126/scirobotics.abb0467 – year: 2018 ident: B4 publication-title: Smart Arm – volume: 19 start-page: 4596 year: 2019 ident: B26 article-title: Real-time EMG based pattern recognition control for hand prostheses: a review on existing methods, challenges and future implementation publication-title: Sensors doi: 10.3390/s19204596 – volume-title: Cram's Introduction to Surface Electromyography year: 2010 ident: B5 – volume: 6 start-page: 6 year: 2016 ident: B13 article-title: Gesture recognition by instantaneous surface EMG images publication-title: Sci. Rep doi: 10.1038/srep36571 – volume: 9 start-page: 247 year: 2016 ident: B12 article-title: Myoelectric control of prosthetic hands: state-of-the-art review publication-title: Med. Devices doi: 10.2147/MDER.S91102 – volume: 237 start-page: 291 year: 2019 ident: B35 article-title: Improving the functionality, robustness, and adaptability of myoelectric control for dexterous motion restoration publication-title: Exp. Brain Res doi: 10.1007/s00221-018-5441-x – volume: 28 start-page: 2090 year: 2020 ident: B31 article-title: Computer vision-based grasp pattern recognition with application to myoelectric control of dexterous hand prosthesis publication-title: IEEE Trans. Neural Syst. Rehabil. Eng doi: 10.1109/TNSRE.2020.3007625 – volume: 7 start-page: 42 year: 2010 ident: B6 article-title: Cognitive vision system for control of dexterous prosthetic hands: experimental evaluation publication-title: J. NeuroEng. Rehabil doi: 10.1186/1743-0003-7-42 – volume: 12 start-page: 066022 year: 2015 ident: B21 article-title: Sensor fusion and computer vision for context-aware control of a multi degree-of-freedom prosthesis publication-title: J. Neural Eng doi: 10.1088/1741-2560/12/6/066022 – start-page: 1 year: 2020 ident: B29 article-title: Current rates of prosthetic usage in upper-limb amputeeshave innovations had an impact on device acceptance? publication-title: Disabil. Rehabil doi: 10.1080/09638288.2020.1866684. – year: 2019 ident: B15 article-title: Grasp type estimation for myoelectric prostheses using point cloud feature learning publication-title: arXiv[Preprint]. – volume: 14 start-page: 036025 year: 2017 ident: B14 article-title: Deep learning-based artificial vision for grasp classification in myoelectric hands publication-title: J. Neural Eng doi: 10.1088/1741-2552/aa6802 – volume: 31 start-page: 236 year: 2007 ident: B3 article-title: Upper limb prosthesis use and abandonment: a survey of the last 25 years publication-title: Prosthet. Orthot. Int doi: 10.1080/03093640600994581 – volume: 3 year: 2018 ident: B17 article-title: Simultaneous control of multiple functions of bionic hand prostheses: performance and robustness in end users publication-title: Sci. Robot doi: 10.1126/scirobotics.aat3630 – volume: 11 start-page: 046001 year: 2014 ident: B20 article-title: Stereovision and augmented reality for closed-loop control of grasping in hand prostheses publication-title: J. Neural Eng doi: 10.1088/1741-2560/11/4/046001 – volume: 22 start-page: 269 year: 2014 ident: B16 article-title: Linear and nonlinear regression techniques for simultaneous and proportional myoelectric control publication-title: IEEE Trans. Neural Syst. Rehabil. Eng doi: 10.1109/TNSRE.2014.2305520 – volume: 10 start-page: 9 year: 2016 ident: B2 article-title: Deep learning with convolutional neural networks applied to electromyography data: a resource for the classification of movements for prosthetic hands publication-title: Front. Neurorobot doi: 10.3389/fnbot.2016.00009 – volume: 29 start-page: 152 year: 2012 ident: B18 article-title: Myoelectric control of artificial limbs—is there a need to change focus? [in the spotlight] publication-title: IEEE Sign. Process. Mag doi: 10.1109/MSP.2012.2203480 – volume: 22 start-page: 756 year: 2014 ident: B24 article-title: Real-time and simultaneous control of artificial limbs based on pattern recognition algorithms publication-title: IEEE Trans. Neural Syst. Rehabil. Eng doi: 10.1109/TNSRE.2014.2305097 – volume-title: IEEE International Conference on Robotics and Automation (ICRA) year: 2011 ident: B28 article-title: “3D is here: Point Cloud Library (PCL),” doi: 10.1109/ICRA.2011.5980567 – year: 2021 ident: B8 article-title: Toward higher-performance bionic limbs for wider clinical use publication-title: Nat. Biomed. Eng doi: 10.1038/s41551-021-00732-x. – start-page: 304 volume-title: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition year: 2014 ident: B32 article-title: “Object partitioning using local convexity,” |
SSID | ssj0062658 |
Score | 2.3451276 |
Snippet | Modern myoelectric prostheses can perform multiple functions (e.g., several grasp types and wrist rotation) but their intuitive control by the user is still an... |
SourceID | doaj pubmedcentral proquest pubmed crossref |
SourceType | Open Website Open Access Repository Aggregation Database Index Database Enrichment Source |
StartPage | 814973 |
SubjectTerms | Cameras Classification computer vision grasping Hand myoelectric hand prosthesis Neural networks Neuroscience object segmentation point cloud processing Prostheses Prosthetics Robotics semi-autonomous control Sensors User needs Wrist |
SummonAdditionalLinks | – databaseName: DOAJ: Directory of Open Access Journal (DOAJ) dbid: DOA link: http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwrV1Na9wwEB1CTsmhNGnaOk2KAj0V3BhJtuVjPlkKCYU0kJvQJ7vQakPW-_87Y3mX3VLaS4-2ZSy_kTRvrPEbgE8198FVBsMSIWIpQxVKJZUoPSnHhKAiH7Zi7u6byaP8-lQ_bZT6opywLA-cgTuXlXc8khtzUVopjbMY8hhBOlpdYyytvujzVsFUXoORpdcq72FiCNadx2TnlDjJ-ReFIUErtrzQINb_J4b5e6Lkhue5fQ2vRsrILnJXD2AnpEPY3xASfAP3JDI1S0sM49lD-DkrzbKnvxXo-Bv91zENi9mCXeW8dDbkCTDDrsNzP8Ub0mL-wuaJYTM2MckfwePtzferSTlWSiid7ERfWmlj1yjfVrGuHJfKmFq0HY_RVd5w9EJeNFZIWxsbEFBkDaTUV_uItoiNEm9hN81TeA8MGWCDrFFIj2gr64xtuqazrRX4EIzeCqhWyGk3yohTNYsfGsMJAlsPYGsCW2ewC_i8vuU5a2j8rfElmWPdkOSvhxM4KPQ4KPS_BkUBJytj6nFOLjRSP4EECEO0As7Wl3E20RaJSQFNQm06jsuSrAp4l22_7gl9IaMKOAW0W6Niq6vbV9JsOih2U43OVrTH_-PdPsAewUV5cLw-gd3-ZRlOkRj19uMwB34BCVAL_w priority: 102 providerName: Directory of Open Access Journals – databaseName: ProQuest Central Database Suite (ProQuest) dbid: BENPR link: http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwfV1Lb9QwEB5Be4ED4k2gICNxQgqNbCdxToiWViskVhVQqbfIT3al4iyb7P9nJvEuXYR6TOIozow98409_gbgXcmdt4XGsESIkEtf-FxJJXJHzDHeq8DHrZiv82p2Kb9clVdpwa1PaZVbmzgaatdZWiM_Rsct0H0hwP64-p1T1SjaXU0lNO7CIZpghSP88ORsfvFta4sRrZdq2svEUKw5DtF0lEDJ-QeFoUEt9rzRSNr_P6T5b8LkDQ90_hAeJOjIPk26fgR3fHwM928QCj6BOZFNLeMGw3n23f9a5noz0KkFur6g8x0L3y97djrlp7MxX4Bp9tmvhgW-EPtuzbrIsBmb6eiewuX52Y_TWZ4qJuRWNmLIjTShqZSri1AWlkuldSnqhodgC6c5eiMnKiOkKbXxhbOIHoixr3QBdRIqJZ7BQeyifwEMkWCF6FFIJ41UxmpTNVVjaiPwIxjFZVBsJdfaRCdOVS2uWwwrSNjtKOyWhN1Ows7g_e6V1cSlcVvjE1LHriHRYI83uvXPNs2qVuI_8EAYxwbsptTWYDysBZGsNZU2GRxtldmmudm3f0dSBm93j3FW0VaJjh5VQm0ajuZJFhk8n3S_6wmtlFElnAzqvVGx19X9J3G5GJm7qVZnLeqXt3frFdwjQVCmGy-P4GBYb_xrhD6DeZPG9x8COQT7 priority: 102 providerName: ProQuest |
Title | Continuous Semi-autonomous Prosthesis Control Using a Depth Sensor on the Hand |
URI | https://www.ncbi.nlm.nih.gov/pubmed/35401136 https://www.proquest.com/docview/2643081548 https://www.proquest.com/docview/2649253140 https://pubmed.ncbi.nlm.nih.gov/PMC8989737 https://doaj.org/article/40dc2f8222cf4b44acb402a3681096ab |
Volume | 16 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV3fa9swED669mV7GO1-uu2CBnsauHMkWbYfxli7dmHQULYF8mYkW2oCndwlDnT__e5sxzQljL0EHMuRfCflvs86fwfwLualLSKNtEQIF0ob2TCVqQhLUo6xNnW82Yq5HKvRRH6bxtMdWJe36gy43ErtqJ7UZHFzcvf7zydc8B-JcWK8_eC8qSgtkvOTFAF_Ih7BHgamhCo5XMp-UwGhe1Ouc6gU8a9h2m5ybv-JjTDVqPlvg6APMynvhaaLfXjaYUr2uZ0EB7Bj_TN4ck9p8DmMSYVq7lfI89kP-2se6lVNrzPQ8RW9-DGzy_mSnbWJ66xJJGCafbG39Qwv8MtqwSrPsBkbaV--gMnF-c-zUdiVUggLmYk6NNK4TKVlErk4KrhMtY5FknHniqjUHMNUKZQR0sTa2KgsEFaQlF9cOnSWU6l4Cbu-8vY1MISICmGlkKU0MjWFNipTmUmMwE6Q3gUQrS2XF53OOJW7uMmRb5Cx88bYORk7b40dwPv-kttWZONfjU_JHX1D0sduvqgW13m33HKJ98AdgZ_C4TClLgwSZS1IfS1T2gRwvHZmvp5zOWJDgQgJOVwAb_vTuNxoD0V7iy6hNhnH_y0ZBfCq9X0_EnqERiVyAkg2ZsXGUDfP-PmskfSmIp6JSA7_o98jeEzWoDw4Hh_Dbr1Y2TcIjGozgL3T8_HV90HzYAE_v06Hg2YJ_AVNEQ7c |
linkProvider | Scholars Portal |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwtV1Lb9QwEB6V7QE4IN4EChgJLkihke28DgjRl7a0XVXQSr2ltmOzK7XOstkV4k_xG5nJY-ki1FuPSZxkMjOeRzz-BuBtzEtrIoVpiRAulDayYSYzEZaEHGNt5nizFHM0Soan8stZfLYGv_u9MFRW2dvExlCXlaF_5JvouAW6LwywP01_hNQ1ilZX-xYarVoc2F8_MWWrP-7voHzfcb63e7I9DLuuAqGRuZiHWmqXJ1mZRi6ODJeZUrFIc-6ciUrF0WKXItFC6lhpG5UGPSyh2sWlQ7pdkgl87i1YlwJDhQGsb-2Ojr_2th-zgzhr104x9cs3ndcVFWxy_iHDVCQVK96vaRLwv8j23wLNKx5v7z7c60JV9rnVrQewZv1DuHsFwPARjAjcauIX1aJm3-zlJFSLOe2SoONj2k8ytvWkZtttPTxr6hOYYjt2Oh_jDb6uZqzyDIexofLlYzi9EV4-gYGvvH0GDCPPBKNVIUupZaaN0kme5DrVAl-CWWMAUc-5wnTw5dRF46LANIaYXTTMLojZRcvsAN4vb5m22B3XDd4icSwHEux2c6KafS-6WVxI_AbuKKYyDsmUymjMv5UgULc8UTqAjV6YRWcL6uKv5gbwZnkZZzEtzShvUSQ0JudoDmUUwNNW9ktK6M8cdd4JIF3RihVSV6_4ybhBCqfeoKlIn19P1mu4PTw5OiwO90cHL-AOMYWq7Hi8AYP5bGFfYtg11686XWdwftPT6w9VPEHm |
linkToPdf | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwtV3db9MwED-NTkLsAfFNYICR4AUpNLKdrweE2LqqY1BVwKS9BTu2aaWRlCYV4l_jr-MuH2VFaG97TGwnzt35PuLz7wBehNzYPFAYlgjhfGkD6ycyEb4h5BhrE8ebrZiP02hyKt-fhWc78Ls_C0Nplb1ObBS1KXP6Rz5Ewy3QfKGDPXRdWsRsNH67_OFTBSnaae3LabQicmJ__cTwrXpzPEJev-R8fPTlcOJ3FQb8XKai9rXULo0SEwcuDHIuE6VCEafcuTwwiqP2NiLSQupQaRuYHK0tIdyFxuE3uCgR-NxrsBvT8dEB7B4cTWefejuAkUKYtPuoGAamQ1fokpI3OX-dYFgSiy1L2BQM-J-X-2-y5gXrN74FNzu3lb1r5ew27NjiDuxdADO8C1MCuloU63Jdsc_2-8JX65pOTND1jM6WzG21qNhhmxvPmlwFptjILus5DiiqcsXKgmE3NlGFuQenV0LL-zAoysI-BIZeaISeq5BGapnoXOkojVIda4EvwQjSg6CnXJZ3UOZUUeM8w5CGiJ01xM6I2FlLbA9ebYYsWxyPyzofEDs2HQmCu7lRrr5l3YrOJH4Dd-Rf5Q6nKVWuMRZXggDe0khpD_Z7ZmadXqiyv1LswfNNM65o2qZRhUWWUJ-Uo2qUgQcPWt5vZkJ_6agKjwfxllRsTXW7pVjMG9RwqhMai_jR5dN6BtdxWWUfjqcnj-EG0YQS7ni4D4N6tbZP0AOr9dNO1Bl8verV9QdDgUYk |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Continuous+Semi-autonomous+Prosthesis+Control+Using+a+Depth+Sensor+on+the+Hand&rft.jtitle=Frontiers+in+neurorobotics&rft.au=Castro%2C+Miguel+Nobre&rft.au=Dosen%2C+Strahinja&rft.date=2022-03-25&rft.issn=1662-5218&rft.eissn=1662-5218&rft.volume=16&rft.spage=814973&rft_id=info:doi/10.3389%2Ffnbot.2022.814973&rft.externalDBID=NO_FULL_TEXT |
thumbnail_l | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=1662-5218&client=summon |
thumbnail_m | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=1662-5218&client=summon |
thumbnail_s | http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=1662-5218&client=summon |