We propose a robust gesture-based communication pipeline for divers to instruct an Autonomous Underwater Vehicle (AUV) to assist them in performing high-risk tasks and helping in case of emergency. A gesture communication language (CADDIAN) is developed, based on consolidated and standardized diver gestures, including an alphabet, syntax and semantics, ensuring a logical consistency. A hierarchical classification approach is introduced for hand gesture recognition based on stereo imagery and multi-descriptor aggregation to specifically cope with underwater image artifacts, e.g. light backscatter or color attenuation. Once the classification task is finished, a syntax check is performed to filter out invalid command sequences sent by the diver or generated by errors in the classifier. Throughout this process, the diver receives constant feedback from an underwater tablet to acknowledge or abort the mission at any time. The objective is to prevent the AUV from executing unnecessary, infeasible or potentially harmful motions. Experimental results under different environmental conditions in archaeological exploration and bridge inspection applications show that the system performs well in the field.

Robust Gesture-Based Communication for Underwater Human-Robot Interaction in the context of Search and Rescue Diver Missions

D Chiarella;
2018

Abstract

We propose a robust gesture-based communication pipeline for divers to instruct an Autonomous Underwater Vehicle (AUV) to assist them in performing high-risk tasks and helping in case of emergency. A gesture communication language (CADDIAN) is developed, based on consolidated and standardized diver gestures, including an alphabet, syntax and semantics, ensuring a logical consistency. A hierarchical classification approach is introduced for hand gesture recognition based on stereo imagery and multi-descriptor aggregation to specifically cope with underwater image artifacts, e.g. light backscatter or color attenuation. Once the classification task is finished, a syntax check is performed to filter out invalid command sequences sent by the diver or generated by errors in the classifier. Throughout this process, the diver receives constant feedback from an underwater tablet to acknowledge or abort the mission at any time. The objective is to prevent the AUV from executing unnecessary, infeasible or potentially harmful motions. Experimental results under different environmental conditions in archaeological exploration and bridge inspection applications show that the system performs well in the field.
Campo DC Valore Lingua
dc.authority.orgunit Istituto di linguistica computazionale "Antonio Zampolli" - ILC -
dc.authority.people Gomez Chavez A it
dc.authority.people C A Mueller it
dc.authority.people T Doernbach it
dc.authority.people D Chiarella it
dc.authority.people A Birk it
dc.authority.project Cognitive autonomous diving buddy -
dc.collection.id.s 71c7200a-7c5f-4e83-8d57-d3d2ba88f40d *
dc.collection.name 04.01 Contributo in Atti di convegno *
dc.date.accessioned 2024/02/21 05:01:36 -
dc.date.available 2024/02/21 05:01:36 -
dc.date.issued 2018 -
dc.description.abstracteng We propose a robust gesture-based communication pipeline for divers to instruct an Autonomous Underwater Vehicle (AUV) to assist them in performing high-risk tasks and helping in case of emergency. A gesture communication language (CADDIAN) is developed, based on consolidated and standardized diver gestures, including an alphabet, syntax and semantics, ensuring a logical consistency. A hierarchical classification approach is introduced for hand gesture recognition based on stereo imagery and multi-descriptor aggregation to specifically cope with underwater image artifacts, e.g. light backscatter or color attenuation. Once the classification task is finished, a syntax check is performed to filter out invalid command sequences sent by the diver or generated by errors in the classifier. Throughout this process, the diver receives constant feedback from an underwater tablet to acknowledge or abort the mission at any time. The objective is to prevent the AUV from executing unnecessary, infeasible or potentially harmful motions. Experimental results under different environmental conditions in archaeological exploration and bridge inspection applications show that the system performs well in the field. -
dc.description.affiliations Jacobs University Bremen, Germany;Jacobs University Bremen, Germany;Jacobs University Bremen, Germany;CNR-ILC;Jacobs University Bremen, Germany -
dc.description.allpeople Gomez Chavez, A; A Mueller, C; Doernbach, T; Chiarella, D; Birk, A -
dc.description.allpeopleoriginal Gomez Chavez, A., C. A. Mueller, T. Doernbach, D. Chiarella, A. Birk -
dc.description.fulltext none en
dc.description.numberofauthors 5 -
dc.identifier.uri https://hdl.handle.net/20.500.14243/345332 -
dc.identifier.url https://arxiv.org/pdf/1810.07122.pdf -
dc.language.iso eng -
dc.relation.conferencedate 1-5/10/2018 -
dc.relation.conferencename 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems -
dc.relation.conferenceplace Madrid, Spain -
dc.relation.numberofpages 4 -
dc.relation.projectAcronym CADDY -
dc.relation.projectAwardNumber 611373 -
dc.relation.projectAwardTitle Cognitive autonomous diving buddy -
dc.relation.projectFunderName - en
dc.relation.projectFundingStream FP7 -
dc.subject.keywords Underwater Robotics -
dc.subject.keywords gesture-based communication -
dc.subject.keywords underwater human-robot interaction -
dc.subject.singlekeyword Underwater Robotics *
dc.subject.singlekeyword gesture-based communication *
dc.subject.singlekeyword underwater human-robot interaction *
dc.title Robust Gesture-Based Communication for Underwater Human-Robot Interaction in the context of Search and Rescue Diver Missions en
dc.type.driver info:eu-repo/semantics/conferenceObject -
dc.type.full 04 Contributo in convegno::04.01 Contributo in Atti di convegno it
dc.type.miur 273 -
dc.type.referee Sì, ma tipo non specificato -
dc.ugov.descaux1 393268 -
iris.orcid.lastModifiedDate 2024/04/04 12:03:43 *
iris.orcid.lastModifiedMillisecond 1712225023944 *
iris.sitodocente.maxattempts 4 -
Appare nelle tipologie: 04.01 Contributo in Atti di convegno
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.14243/345332
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
social impact