Enabling seamless access to digital graphical contents for visually impaired individuals via semantic-aware processing

Zheshen Wang, Xinyu Xu, Baoxin Li

Research output: Contribution to journalArticle

3 Citations (Scopus)

Abstract

Vision is one of the main sources through which people obtain information from the world, but unfortunately, visually-impaired people are partially or completely deprived of this type of information. With the help of computer technologies, people with visual impairment can independently access digital textual information by using text-to-speech and text-to-Braille software. However, in general, there still exists a major barrier for people who are blind to access the graphical information independently in real-time without the help of sighted people. In this paper, we propose a novel multi-level and multi-modal approach aiming at addressing this challenging and practical problem, with the key idea being semantic-aware visual-to-tactile conversion through semantic image categorization and segmentation, and semantic-driven image simplification. An end-to-end prototype system was built based on the approach. We present the details of the approach and the system, report sample experimental results with realistic data, and compare our approach with current typical practice.

Original languageEnglish (US)
Article number18019
JournalEurasip Journal on Image and Video Processing
Volume2007
DOIs
StatePublished - 2007

Fingerprint

Semantics
Processing

ASJC Scopus subject areas

  • Electrical and Electronic Engineering
  • Signal Processing
  • Information Systems

Cite this

@article{f06adedbfaf549db9953c6ede533b774,
title = "Enabling seamless access to digital graphical contents for visually impaired individuals via semantic-aware processing",
abstract = "Vision is one of the main sources through which people obtain information from the world, but unfortunately, visually-impaired people are partially or completely deprived of this type of information. With the help of computer technologies, people with visual impairment can independently access digital textual information by using text-to-speech and text-to-Braille software. However, in general, there still exists a major barrier for people who are blind to access the graphical information independently in real-time without the help of sighted people. In this paper, we propose a novel multi-level and multi-modal approach aiming at addressing this challenging and practical problem, with the key idea being semantic-aware visual-to-tactile conversion through semantic image categorization and segmentation, and semantic-driven image simplification. An end-to-end prototype system was built based on the approach. We present the details of the approach and the system, report sample experimental results with realistic data, and compare our approach with current typical practice.",
author = "Zheshen Wang and Xinyu Xu and Baoxin Li",
year = "2007",
doi = "10.1155/2007/18019",
language = "English (US)",
volume = "2007",
journal = "Eurasip Journal on Image and Video Processing",
issn = "1687-5176",
publisher = "Springer Publishing Company",

}

TY - JOUR

T1 - Enabling seamless access to digital graphical contents for visually impaired individuals via semantic-aware processing

AU - Wang, Zheshen

AU - Xu, Xinyu

AU - Li, Baoxin

PY - 2007

Y1 - 2007

N2 - Vision is one of the main sources through which people obtain information from the world, but unfortunately, visually-impaired people are partially or completely deprived of this type of information. With the help of computer technologies, people with visual impairment can independently access digital textual information by using text-to-speech and text-to-Braille software. However, in general, there still exists a major barrier for people who are blind to access the graphical information independently in real-time without the help of sighted people. In this paper, we propose a novel multi-level and multi-modal approach aiming at addressing this challenging and practical problem, with the key idea being semantic-aware visual-to-tactile conversion through semantic image categorization and segmentation, and semantic-driven image simplification. An end-to-end prototype system was built based on the approach. We present the details of the approach and the system, report sample experimental results with realistic data, and compare our approach with current typical practice.

AB - Vision is one of the main sources through which people obtain information from the world, but unfortunately, visually-impaired people are partially or completely deprived of this type of information. With the help of computer technologies, people with visual impairment can independently access digital textual information by using text-to-speech and text-to-Braille software. However, in general, there still exists a major barrier for people who are blind to access the graphical information independently in real-time without the help of sighted people. In this paper, we propose a novel multi-level and multi-modal approach aiming at addressing this challenging and practical problem, with the key idea being semantic-aware visual-to-tactile conversion through semantic image categorization and segmentation, and semantic-driven image simplification. An end-to-end prototype system was built based on the approach. We present the details of the approach and the system, report sample experimental results with realistic data, and compare our approach with current typical practice.

UR - http://www.scopus.com/inward/record.url?scp=38749126793&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=38749126793&partnerID=8YFLogxK

U2 - 10.1155/2007/18019

DO - 10.1155/2007/18019

M3 - Article

AN - SCOPUS:38749126793

VL - 2007

JO - Eurasip Journal on Image and Video Processing

JF - Eurasip Journal on Image and Video Processing

SN - 1687-5176

M1 - 18019

ER -