2022
DOI: 10.3390/electronics11203335
|View full text |Cite
|
Sign up to set email alerts
|

Deep Learning Reader for Visually Impaired

Abstract: Recent advances in machine and deep learning algorithms and enhanced computational capabilities have revolutionized healthcare and medicine. Nowadays, research on assistive technology has benefited from such advances in creating visual substitution for visual impairment. Several obstacles exist for people with visual impairment in reading printed text which is normally substituted with a pattern-based display known as Braille. Over the past decade, more wearable and embedded assistive devices and solutions wer… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
13
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 20 publications
(13 citation statements)
references
References 61 publications
(63 reference statements)
0
13
0
Order By: Relevance
“…LITERATURE SURVEY A remedy for the issue for visually impaired to help with their difficulties in reading texts in non-ideal conditions that may include changing lighting conditions, the different font styles used in the document and the direction of the characters. The proposed system [1], developed using the MATLAB platform, serves as an affordable substitute, making it accessible to a broader user base. Its core functionalities rely on two key components: Optical Character Recognition (OCR) and Text-to-Speech technology.…”
Section: IImentioning
confidence: 99%
“…LITERATURE SURVEY A remedy for the issue for visually impaired to help with their difficulties in reading texts in non-ideal conditions that may include changing lighting conditions, the different font styles used in the document and the direction of the characters. The proposed system [1], developed using the MATLAB platform, serves as an affordable substitute, making it accessible to a broader user base. Its core functionalities rely on two key components: Optical Character Recognition (OCR) and Text-to-Speech technology.…”
Section: IImentioning
confidence: 99%
“…The GED map size of the block is 25 pixels [10]. The depth map was built based on a disparity map by calculating the camera baseline and focal length [13,18,19]. Then it was divided by x for every pixel of the disparity map.…”
Section: Grid-edge-depth Map Buildingmentioning
confidence: 99%
“…For instance, verbal description of a scene acquired in real time can help visually impaired and blind people to determine their location, navigate their surroundings, access visual information, feel safer, and increase their situational awareness [11], [12]. Indeed, there are some recent works looking at the use of image captioning to help the visually impaired and blind [13], [14]. These studies focus on the English language, because benchmark image captioning datasets with human-generated captions (e.g., Microsoft COCO Captions dataset [15]) are generally created in this high-resource language.…”
Section: Introductionmentioning
confidence: 99%