2021
DOI: 10.3389/fnbot.2020.610139
|View full text |Cite
|
Sign up to set email alerts
|

Intention Understanding in Human–Robot Interaction Based on Visual-NLP Semantics

Abstract: With the rapid development of robotic and AI technology in recent years, human–robot interaction has made great advancement, making practical social impact. Verbal commands are one of the most direct and frequently used means for human–robot interaction. Currently, such technology can enable robots to execute pre-defined tasks based on simple and direct and explicit language instructions, e.g., certain keywords must be used and detected. However, that is not the natural way for human to communicate. In this pa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
11
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 18 publications
(11 citation statements)
references
References 25 publications
(26 reference statements)
0
11
0
Order By: Relevance
“…21 tactile sensor arrays are assembled on the inner surface of the ABS shell of the robot mentioned in [106] Realsense D435 Camera NVIDIA Jetson TX2 Figure 4: A new task-based framework that enables robots to understand human intentions using visual-NLP semantic information [109]: it includes a language semantics module to extract keywords no matter if the command directive is explicit or not, a visual object recognition module to identify multiple objects located to the front of the robot, and a similarity computation algorithm for inferring the intention based on a given task (i.e., selecting some desired item out of multiple objects on a table and giving it to a particular user among several human participants). Result of the similarity computation is then translated into structured robot control language RCL (grasp object to place) to be comprehended by robots.…”
Section: App (Instances) On the Touch Screenmentioning
confidence: 99%
See 1 more Smart Citation
“…21 tactile sensor arrays are assembled on the inner surface of the ABS shell of the robot mentioned in [106] Realsense D435 Camera NVIDIA Jetson TX2 Figure 4: A new task-based framework that enables robots to understand human intentions using visual-NLP semantic information [109]: it includes a language semantics module to extract keywords no matter if the command directive is explicit or not, a visual object recognition module to identify multiple objects located to the front of the robot, and a similarity computation algorithm for inferring the intention based on a given task (i.e., selecting some desired item out of multiple objects on a table and giving it to a particular user among several human participants). Result of the similarity computation is then translated into structured robot control language RCL (grasp object to place) to be comprehended by robots.…”
Section: App (Instances) On the Touch Screenmentioning
confidence: 99%
“…An appropriated facial expression out of 20 choices (shyness, thinking (turning eyes), etc.) will be displayed to the child with body gestures and speech output sometimes Attention-based image captioning with gaze-following [110] A robot takes a video as input, then builds a 3D dense map based on SLAM, and estimates the gaze simultaneously Attention prediction is done by attention heat map adding occlusion and salience detection Te robot verbally describes the region focused on by the child based on attention heat mapsIntention understanding based on visual-NLP semantics and responding to the child accordingly[109] …”
mentioning
confidence: 99%
“…Li et al [23] use natural language processing to infer human-given commands for robots, by using keyword extraction, visual object recognition, and similarity computation. Its main intent is to use visual semantic information to allow a robot deduce task intents, avoiding simple keywords that map predefined tasks explicitly.…”
Section: Related Workmentioning
confidence: 99%
“…Third, data may be scarce when a customized dataset needs to be created for a specific task. For example, when developing a robot to assist human workers, developers should create new data suitable for the working environment [10], [11]. Therefore, the data scarcity problem is an important topic to be addressed.…”
Section: Introductionmentioning
confidence: 99%