2022
DOI: 10.1145/3534613
|View full text |Cite
|
Sign up to set email alerts
|

EarCommand

Abstract: Intelligent speech interfaces have been developing vastly to support the growing demands for convenient control and interaction with wearable/earable and portable devices. To avoid privacy leakage during speech interactions and strengthen the resistance to ambient noise, silent speech interfaces have been widely explored to enable people's interaction with mobile/wearable devices without audible sounds. However, most existing silent speech solutions require either restricted background illuminations or hand in… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 14 publications
(4 citation statements)
references
References 55 publications
0
4
0
Order By: Relevance
“…User State. The sensors that could be integrated within future HMDs would empower an AR system to have a rich, instant understanding of user's state, such as activities (IMU [86,219], camera [80,128,194,201], microphone [103,218,229,230]), cognitive load (eye tracking [71,104,238], EEG [20,224]), attention (eye tracking [56,99,204,231], IMU [123], EEG [213]), emotion (facial tracking [233,236], EEG [202,216]) and potential intent (the fusion of multiple sensors and low-level intelligence [14,111,211]). Depending on a user's state, the design of explanations could be different.…”
Section: Key Factorsmentioning
confidence: 99%
“…User State. The sensors that could be integrated within future HMDs would empower an AR system to have a rich, instant understanding of user's state, such as activities (IMU [86,219], camera [80,128,194,201], microphone [103,218,229,230]), cognitive load (eye tracking [71,104,238], EEG [20,224]), attention (eye tracking [56,99,204,231], IMU [123], EEG [213]), emotion (facial tracking [233,236], EEG [202,216]) and potential intent (the fusion of multiple sensors and low-level intelligence [14,111,211]). Depending on a user's state, the design of explanations could be different.…”
Section: Key Factorsmentioning
confidence: 99%
“…Contactless approaches are mainly explored through camera-based visual signals, [9][10][11][12][13][14][15] ultrasound signals. [16][17][18][19][20][21][22][23] Camerabased visual solutions require external video tracking devices, and users must remain within the camera's line of sight. Despite efforts to develop compact shoulder-mounted devices 9 to enhance portability, visual solutions still face challenges in terms of lighting conditions and angles between users and cameras, thereby limiting their practicality.…”
Section: Introductionmentioning
confidence: 99%
“…As a more portable and user-friendly ultrasound-based solution, the speaker of the cell phone was used to emit ultrasound signals, and the microphone was employed to capture reflected signals from the lips. 16,[18][19][20][21][22] This method is not hands-free and is susceptible to multipath interferences caused by bodily movements and surrounding objects.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation