- •Preface
- •Contents
- •1 Disability and Assistive Technology Systems
- •Learning Objectives
- •1.1 The Social Context of Disability
- •1.2 Assistive Technology Outcomes: Quality of Life
- •1.2.1 Some General Issues
- •1.2.2 Definition and Measurement of Quality of Life
- •1.2.3 Health Related Quality of Life Measurement
- •1.2.4 Assistive Technology Quality of Life Procedures
- •1.2.5 Summary and Conclusions
- •1.3 Modelling Assistive Technology Systems
- •1.3.1 Modelling Approaches: A Review
- •1.3.2 Modelling Human Activities
- •1.4 The Comprehensive Assistive Technology (CAT) Model
- •1.4.1 Justification of the Choice of Model
- •1.4.2 The Structure of the CAT Model
- •1.5 Using the Comprehensive Assistive Technology Model
- •1.5.1 Using the Activity Attribute of the CAT Model to Determine Gaps in Assistive Technology Provision
- •1.5.2 Conceptual Structure of Assistive Technology Systems
- •1.5.3 Investigating Assistive Technology Systems
- •1.5.4 Analysis of Assistive Technology Systems
- •1.5.5 Synthesis of Assistive Technology Systems
- •1.6 Chapter Summary
- •Questions
- •Projects
- •References
- •2 Perception, the Eye and Assistive Technology Issues
- •Learning Objectives
- •2.1 Perception
- •2.1.1 Introduction
- •2.1.2 Common Laws and Properties of the Different Senses
- •2.1.3 Multisensory Perception
- •2.1.4 Multisensory Perception in the Superior Colliculus
- •2.1.5 Studies of Multisensory Perception
- •2.2 The Visual System
- •2.2.1 Introduction
- •2.2.2 The Lens
- •2.2.3 The Iris and Pupil
- •2.2.4 Intraocular Pressure
- •2.2.5 Extraocular Muscles
- •2.2.6 Eyelids and Tears
- •2.3 Visual Processing in the Retina, Laternal Geniculate Nucleus and the Brain
- •2.3.1 Nerve Cells
- •2.3.2 The Retina
- •2.3.3 The Optic Nerve, Optic Tract and Optic Radiation
- •2.3.4 The Lateral Geniculate Body or Nucleus
- •2.3.5 The Primary Visual or Striate Cortex
- •2.3.6 The Extrastriate Visual Cortex and the Superior Colliculus
- •2.3.7 Visual Pathways
- •2.4 Vision in Action
- •2.4.1 Image Formation
- •2.4.2 Accommodation
- •2.4.3 Response to Light
- •2.4.4 Colour Vision
- •2.4.5 Binocular Vision and Stereopsis
- •2.5 Visual Impairment and Assistive Technology
- •2.5.1 Demographics of Visual Impairment
- •2.5.2 Illustrations of Some Types of Visual Impairment
- •2.5.3 Further Types of Visual Impairment
- •2.5.4 Colour Blindness
- •2.5.5 Corrective Lenses
- •2.6 Chapter Summary
- •Questions
- •Projects
- •References
- •3 Sight Measurement
- •Learning Objectives
- •3.1 Introduction
- •3.2 Visual Acuity
- •3.2.1 Using the Chart
- •3.2.2 Variations in Measuring Visual Acuity
- •3.3 Field of Vision Tests
- •3.3.1 The Normal Visual Field
- •3.3.2 The Tangent Screen
- •3.3.3 Kinetic Perimetry
- •3.3.4 Static Perimetry
- •3.4 Pressure Measurement
- •3.5 Biometry
- •3.6 Ocular Examination
- •3.7 Optical Coherence Tomography
- •3.7.1 Echo Delay
- •3.7.2 Low Coherence Interferometry
- •3.7.3 An OCT Scanner
- •3.8 Ocular Electrophysiology
- •3.8.1 The Electrooculogram (EOG)
- •3.8.2 The Electroretinogram (ERG)
- •3.8.3 The Pattern Electroretinogram
- •3.8.4 The Visual Evoked Cortical Potential
- •3.8.5 Multifocal Electrophysiology
- •3.9 Chapter Summary
- •Glossary
- •Questions
- •Projects
- •4 Haptics as a Substitute for Vision
- •Learning Objectives
- •4.1 Introduction
- •4.1.1 Physiological Basis
- •4.1.2 Passive Touch, Active Touch and Haptics
- •4.1.3 Exploratory Procedures
- •4.2 Vision and Haptics Compared
- •4.3 The Capacity of Bare Fingers in Real Environments
- •4.3.1 Visually Impaired People’s Use of Haptics Without any Technical Aid
- •4.3.2 Speech Perceived by Hard-of-hearing People Using Bare Hands
- •4.3.3 Natural Capacity of Touch and Evaluation of Technical Aids
- •4.4 Haptic Low-tech Aids
- •4.4.1 The Long Cane
- •4.4.2 The Guide Dog
- •4.4.3 Braille
- •4.4.4 Embossed Pictures
- •4.4.5 The Main Lesson from Low-tech Aids
- •4.5 Matrices of Point Stimuli
- •4.5.1 Aids for Orientation and Mobility
- •4.5.2 Aids for Reading Text
- •4.5.3 Aids for Reading Pictures
- •4.6 Computer-based Aids for Graphical Information
- •4.6.1 Aids for Graphical User Interfaces
- •4.6.2 Tactile Computer Mouse
- •4.7 Haptic Displays
- •4.7.1 Information Available via a Haptic Display
- •4.7.2 What Information Can Be Obtained with the Reduced Information?
- •4.7.3 Haptic Displays as Aids for the Visually Impaired
- •4.8 Chapter Summary
- •4.9 Concluding Remarks
- •Questions
- •Projects
- •References
- •5 Mobility: An Overview
- •Learning Objectives
- •5.1 Introduction
- •5.2 The Travel Activity
- •5.2.1 Understanding Mobility
- •5.2.2 Assistive Technology Systems for the Travel Process
- •5.3 The Historical Development of Travel Aids for Visually Impaired and Blind People
- •5.4 Obstacle Avoidance AT: Guide Dogs and Robotic Guide Walkers
- •5.4.1 Guide Dogs
- •5.4.2 Robotic Guides and Walkers
- •5.5 Obstacle Avoidance AT: Canes
- •5.5.1 Long Canes
- •5.5.2 Technology Canes
- •5.6 Other Mobility Assistive Technology Approaches
- •5.6.1 Clear-path Indicators
- •5.6.2 Obstacle and Object Location Detectors
- •5.6.3 The vOICe System
- •5.7 Orientation Assistive Technology Systems
- •5.7.1 Global Positioning System Orientation Technology
- •5.7.2 Other Technology Options for Orientation Systems
- •5.8 Accessible Environments
- •5.9 Chapter Summary
- •Questions
- •Projects
- •References
- •6 Mobility AT: The Batcane (UltraCane)
- •Learning Objectives
- •6.1 Mobility Background and Introduction
- •6.2 Principles of Ultrasonics
- •6.2.1 Ultrasonic Waves
- •6.2.2 Attenuation and Reflection Interactions
- •6.2.3 Transducer Geometry
- •6.3 Bats and Signal Processing
- •6.3.1 Principles of Bat Sonar
- •6.3.2 Echolocation Call Structures
- •6.3.3 Signal Processing Capabilities
- •6.3.4 Applicability of Bat Echolocation to Sonar System Design
- •6.4 Design and Construction Issues
- •6.4.1 Outline Requirement Specification
- •6.4.2 Ultrasonic Spatial Sensor Subsystem
- •6.4.3 Trial Prototype Spatial Sensor Arrangement
- •6.4.4 Tactile User Interface Subsystem
- •6.4.5 Cognitive Mapping
- •6.4.6 Embedded Processing Control Requirements
- •6.5 Concept Phase and Engineering Prototype Phase Trials
- •6.6 Case Study in Commercialisation
- •6.7 Chapter Summary
- •Questions
- •Projects
- •References
- •7 Navigation AT: Context-aware Computing
- •Learning objectives
- •7.1 Defining the Orientation/Navigation Problem
- •7.1.1 Orientation, Mobility and Navigation
- •7.1.2 Traditional Mobility Aids
- •7.1.3 Limitations of Traditional Aids
- •7.2 Cognitive Maps
- •7.2.1 Learning and Acquiring Spatial Information
- •7.2.2 Factors that Influence How Knowledge Is Acquired
- •7.2.3 The Structure and Form of Cognitive Maps
- •7.3 Overview of Existing Technologies
- •7.3.1 Technologies for Distant Navigation
- •7.3.2 User Interface Output Technologies
- •7.4 Principles of Mobile Context-aware Computing
- •7.4.1 Adding Context to User-computer Interaction
- •7.4.2 Acquiring Useful Contextual Information
- •7.4.3 Capabilities of Context-awareness
- •7.4.4 Application of Context-aware Principles
- •7.4.5 Technological Challenges and Unresolved Usability Issues
- •7.5 Test Procedures
- •7.5.1 Human Computer Interaction (HCI)
- •7.5.2 Cognitive Mapping
- •7.5.3 Overall Approach
- •7.6 Future Positioning Technologies
- •7.7 Chapter Summary
- •7.7.1 Conclusions
- •Questions
- •Projects
- •References
- •Learning Objectives
- •8.1 Defining the Navigation Problem
- •8.1.1 What is the Importance of Location Information?
- •8.1.2 What Mobility Tools and Traditional Maps are Available for the Blind?
- •8.2 Principles of Global Positioning Systems
- •8.2.1 What is the Global Positioning System?
- •8.2.2 Accuracy of GPS: Some General Issues
- •8.2.3 Accuracy of GPS: Some Technical Issues
- •8.2.4 Frequency Spectrum of GPS, Present and Future
- •8.2.5 Other GPS Systems
- •8.3 Application of GPS Principles
- •8.4 Design Issues
- •8.5 Development Issues
- •8.5.1 Choosing an Appropriate Platform
- •8.5.2 Choosing the GPS Receiver
- •8.5.3 Creating a Packaged System
- •8.5.4 Integration vs Stand-alone
- •8.6 User Interface Design Issues
- •8.6.1 How to Present the Information
- •8.6.2 When to Present the Information
- •8.6.3 What Information to Present
- •8.7 Test Procedures and Results
- •8.8 Case Study in Commercialisation
- •8.8.1 Understanding the Value of the Technology
- •8.8.2 Limitations of the Technology
- •8.8.3 Ongoing Development
- •8.9 Chapter Summary
- •Questions
- •Projects
- •References
- •9 Electronic Travel Aids: An Assessment
- •Learning Objectives
- •9.1 Introduction
- •9.2 Why Do an Assessment?
- •9.3 Methodologies for Assessments of Electronic Travel Aids
- •9.3.1 Eliciting User Requirements
- •9.3.2 Developing a User Requirements Specification and Heuristic Evaluation
- •9.3.3 Hands-on Assessments
- •9.3.4 Methodology Used for Assessments in this Chapter
- •9.4 Modern-day Electronic Travel Aids
- •9.4.1 The Distinction Between Mobility and Navigation Aids
- •9.4.2 The Distinction Between Primary and Secondary Aids
- •9.4.3 User Requirements: Mobility and Navigation Aids
- •9.4.4 Mobility Aids
- •9.4.5 Mobility Aids: Have They Solved the Mobility Challenge?
- •9.4.6 Navigation Aids
- •9.4.7 Navigation Aids: Have They Solved the Navigation Challenge?
- •9.5 Training
- •9.6 Chapter Summary and Conclusions
- •Questions
- •Projects
- •References
- •10 Accessible Environments
- •Learning Objectives
- •10.1 Introduction
- •10.1.1 Legislative and Regulatory Framework
- •10.1.2 Accessible Environments: An Overview
- •10.1.3 Principles for the Design of Accessible Environments
- •10.2 Physical Environments: The Streetscape
- •10.2.1 Pavements and Pathways
- •10.2.2 Road Crossings
- •10.2.3 Bollards and Street Furniture
- •10.3 Physical Environments: Buildings
- •10.3.1 General Exterior Issues
- •10.3.2 General Interior Issues
- •10.3.4 Signs and Notices
- •10.3.5 Interior Building Services
- •10.4 Environmental Information and Navigation Technologies
- •10.4.1 Audio Information System: General Issues
- •10.4.2 Some Technologies for Environmental Information Systems
- •10.5 Accessible Public Transport
- •10.5.1 Accessible Public Transportation: Design Issues
- •10.6 Chapter Summary
- •Questions
- •Projects
- •References
- •11 Accessible Bus System: A Bluetooth Application
- •Learning Objectives
- •11.1 Introduction
- •11.2 Bluetooth Fundamentals
- •11.2.1 Brief History of Bluetooth
- •11.2.2 Bluetooth Power Class
- •11.2.3 Protocol Stack
- •11.2.4 Bluetooth Profile
- •11.2.5 Piconet
- •11.3 Design Issues
- •11.3.1 System Architecture
- •11.3.2 Hardware Requirements
- •11.3.3 Software Requirements
- •11.4 Developmental Issues
- •11.4.1 Bluetooth Server
- •11.4.2 Bluetooth Client (Mobile Device)
- •11.4.3 User Interface
- •11.5 Commercialisation Issues
- •11.6 Chapter Summary
- •Questions
- •Projects
- •References
- •12 Accessible Information: An Overview
- •Learning Objectives
- •12.1 Introduction
- •12.2 Low Vision Aids
- •12.2.1 Basic Principles
- •12.3 Low Vision Assistive Technology Systems
- •12.3.1 Large Print
- •12.3.2 Closed Circuit Television Systems
- •12.3.3 Video Magnifiers
- •12.3.4 Telescopic Assistive Systems
- •12.4 Audio-transcription of Printed Information
- •12.4.1 Stand-alone Reading Systems
- •12.4.2 Read IT Project
- •12.5 Tactile Access to Information
- •12.5.1 Braille
- •12.5.2 Moon
- •12.5.3 Braille Devices
- •12.6 Accessible Computer Systems
- •12.6.1 Input Devices
- •12.6.2 Output Devices
- •12.6.3 Computer-based Reading Systems
- •12.6.4 Accessible Portable Computers
- •12.7 Accessible Internet
- •12.7.1 World Wide Web Guidelines
- •12.7.2 Guidelines for Web Authoring Tools
- •12.7.3 Accessible Adobe Portable Document Format (PDF) Documents
- •12.7.4 Bobby Approval
- •12.8 Telecommunications
- •12.8.1 Voice Dialling General Principles
- •12.8.2 Talking Caller ID
- •12.8.3 Mobile Telephones
- •12.9 Chapter Summary
- •Questions
- •Projects
- •References
- •13 Screen Readers and Screen Magnifiers
- •Learning Objectives
- •13.1 Introduction
- •13.2 Overview of Chapter
- •13.3 Interacting with a Graphical User Interface
- •13.4 Screen Magnifiers
- •13.4.1 Overview
- •13.4.2 Magnification Modes
- •13.4.3 Other Interface Considerations
- •13.4.4 The Architecture and Implementation of Screen Magnifiers
- •13.5 Screen Readers
- •13.5.1 Overview
- •13.5.2 The Architecture and Implementation of a Screen Reader
- •13.5.3 Using a Braille Display
- •13.5.4 User Interface Issues
- •13.6 Hybrid Screen Reader Magnifiers
- •13.7 Self-magnifying Applications
- •13.8 Self-voicing Applications
- •13.9 Application Adaptors
- •13.10 Chapter Summary
- •Questions
- •Projects
- •References
- •14 Speech, Text and Braille Conversion Technology
- •Learning Objectives
- •14.1 Introduction
- •14.1.1 Introducing Mode Conversion
- •14.1.2 Outline of the Chapter
- •14.2 Prerequisites for Speech and Text Conversion Technology
- •14.2.1 The Spectral Structure of Speech
- •14.2.2 The Hierarchical Structure of Spoken Language
- •14.2.3 Prosody
- •14.3 Speech-to-text Conversion
- •14.3.1 Principles of Pattern Recognition
- •14.3.2 Principles of Speech Recognition
- •14.3.3 Equipment and Applications
- •14.4 Text-to-speech Conversion
- •14.4.1 Principles of Speech Production
- •14.4.2 Principles of Acoustical Synthesis
- •14.4.3 Equipment and Applications
- •14.5 Braille Conversion
- •14.5.1 Introduction
- •14.5.2 Text-to-Braille Conversion
- •14.5.3 Braille-to-text Conversion
- •14.6 Commercial Equipment and Applications
- •14.6.1 Speech vs Braille
- •14.6.2 Speech Output in Devices for Daily Life
- •14.6.3 Portable Text-based Devices
- •14.6.4 Access to Computers
- •14.6.5 Reading Machines
- •14.6.6 Access to Telecommunication Devices
- •14.7 Discussion and the Future Outlook
- •14.7.1 End-user Studies
- •14.7.2 Discussion and Issues Arising
- •14.7.3 Future Developments
- •Questions
- •Projects
- •References
- •15 Accessing Books and Documents
- •Learning Objectives
- •15.1 Introduction: The Challenge of Accessing the Printed Page
- •15.2 Basics of Optical Character Recognition Technology
- •15.2.1 Details of Optical Character Recognition Technology
- •15.2.2 Practical Issues with Optical Character Recognition Technology
- •15.3 Reading Systems
- •15.4 DAISY Technology
- •15.4.1 DAISY Full Audio Books
- •15.4.2 DAISY Full Text Books
- •15.4.3 DAISY and Other Formats
- •15.5 Players
- •15.6 Accessing Textbooks
- •15.7 Accessing Newspapers
- •15.8 Future Technology Developments
- •15.9 Chapter Summary and Conclusion
- •15.9.1 Chapter Summary
- •15.9.2 Conclusion
- •Questions
- •Projects
- •References
- •Learning Objectives
- •16.1 Introduction
- •16.1.1 Print Impairments
- •16.1.2 Music Notation
- •16.2 Overview of Accessible Music
- •16.2.1 Formats
- •16.2.2 Technical Aspects
- •16.3 Some Recent Initiatives and Projects
- •16.3.2 Play 2
- •16.3.3 Dancing Dots
- •16.3.4 Toccata
- •16.4 Problems to Be Overcome
- •16.4.1 A Content Processing Layer
- •16.4.2 Standardization of Accessible Music Technology
- •16.5 Unifying Accessible Design, Technology and Musical Content
- •16.5.1 Braille Music
- •16.5.2 Talking Music
- •16.6 Conclusions
- •16.6.1 Design for All or Accessibility from Scratch
- •16.6.2 Applying Design for All in Emerging Standards
- •16.6.3 Accessibility in Emerging Technology
- •Questions
- •Projects
- •References
- •17 Assistive Technology for Daily Living
- •Learning Objectives
- •17.1 Introduction
- •17.2 Personal Care
- •17.2.1 Labelling Systems
- •17.2.2 Healthcare Monitoring
- •17.3 Time-keeping, Alarms and Alerting
- •17.3.1 Time-keeping
- •17.3.2 Alarms and Alerting
- •17.4 Food Preparation and Consumption
- •17.4.1 Talking Kitchen Scales
- •17.4.2 Talking Measuring Jug
- •17.4.3 Liquid Level Indicator
- •17.4.4 Talking Microwave Oven
- •17.4.5 Talking Kitchen and Remote Thermometers
- •17.4.6 Braille Salt and Pepper Set
- •17.5 Environmental Control and Use of Appliances
- •17.5.1 Light Probes
- •17.5.2 Colour Probes
- •17.5.3 Talking and Tactile Thermometers and Barometers
- •17.5.4 Using Appliances
- •17.6 Money, Finance and Shopping
- •17.6.1 Mechanical Money Indicators
- •17.6.2 Electronic Money Identifiers
- •17.6.3 Electronic Purse
- •17.6.4 Automatic Teller Machines (ATMs)
- •17.7 Communications and Access to Information: Other Technologies
- •17.7.1 Information Kiosks and Other Self-service Systems
- •17.7.2 Using Smart Cards
- •17.7.3 EZ Access®
- •17.8 Chapter Summary
- •Questions
- •Projects
- •References
- •Learning Objectives
- •18.1 Introduction
- •18.2 Education: Learning and Teaching
- •18.2.1 Accessing Educational Processes and Approaches
- •18.2.2 Educational Technologies, Devices and Tools
- •18.3 Employment
- •18.3.1 Professional and Person-centred
- •18.3.2 Scientific and Technical
- •18.3.3 Administrative and Secretarial
- •18.3.4 Skilled and Non-skilled (Manual) Trades
- •18.3.5 Working Outside
- •18.4 Recreational Activities
- •18.4.1 Accessing the Visual, Audio and Performing Arts
- •18.4.2 Games, Puzzles, Toys and Collecting
- •18.4.3 Holidays and Visits: Museums, Galleries and Heritage Sites
- •18.4.4 Sports and Outdoor Activities
- •18.4.5 DIY, Art and Craft Activities
- •18.5 Chapter Summary
- •Questions
- •Projects
- •References
- •Biographical Sketches of the Contributors
- •Index
528 14 Speech, Text and Braille Conversion Technology
TTS systems compared to recorded speech
The availability of large capacity cheap memory chips has made it possible to store whole spoken utterances, thereby eliminating the need for TTS synthesis under the following conditions:
•The speech corpus consists of a fixed set of utterances.
•The speech corpus is sufficiently small to have limited memory requirements for storage and limited time requirements for production by a human speaker.
•The corpus is fixed, with low probability that it will be necessary to change it in the future.
Recorded speech has a number of applications in daily life, such as public announcements at stations. The principle is known from voice recorders for storing short dictations or other audio information. These recorders are available as both small stand-alone devices and integrated into pocket PCs and other applications. Recorded speech can be of hi-fi quality or reduced quality due to degradation by the codec used to compress the signal. However, recorded speech always sounds more natural than synthetic speech. This can lead to problems in evaluating the quality of TTS systems if users are not aware of the difference between recorded and synthetic speech.
14.5 Braille Conversion
14.5.1 Introduction
Braille is used to represent text by means of tactile symbols. Since tactile text reading is discussed in Chapter 4, it will only be considered briefly here. The main idea was developed (from an unsuccessful system developed for the French army and called night writing) by the French blind teacher Louis Braille in about 1825 and aimed to replace the visible characters in text by a tactile array of up to six dots. Braille is still an important means for blind people to access written documents, though the proportion of blind people who read Braille is low. The elements of a printed text are called characters or signs, whereas the elements of a Braille transcript are called symbols.
Braille code is physically implemented in the form of small raised bumps on paper or other material, which are touched with the fingertips. The bumps can be produced using a stylus and a slate. The slate comprises an upper part with rectangular openings to ensure appropriate positioning of the symbols, and a lower part with small indentions for positioning the dots. Producing large texts in this way is both time-consuming and complicated, as a mirror writing approach must be used for punching the Braille symbols to produce raised bumps rather than depressions. Therefore, mechanical writers called Braille writers or Braillers are used. They are similar to a typewriter with six keys (one key for each dot of a Braille symbol) and a space bar, as illustrated in Figure 14.18. They are operated by Braille transcribers,
14.5 Braille Conversion |
529 |
Figure 14.18a,b. Old and new Braille writing devices: a mechanical writer, embossing on a paper tape for stenographic purposes (Firma Karl Neubert, Leipzig, around 1950), from the historic collection of the TU Dresden (photograph by R. Dietzel); b electronic Braille writer Elotype 4E, also applicable as computer-controlled Braille embosser (by courtesy of Blista Brailletec, Marburg)
who are usually certified professionals, as they have to perform the transcription process mentally, which is not carried out by an automatic (technical) system.
Automatic systems for transcription from text into Braille and Braille into text have a number of advantages, particularly with regards to speeding up the process, but require suitable input and output devices. Other than the need for appropriate hardware, the conversion seems relatively simple, as it is performed sign by sign at the text level. As will be pointed out in the following, this is not in fact totally the case.
The original approach was based on a one-to-one coding of the characters with code combinations defined for all letters, numbers, punctuation marks, and a few special symbols, such as those indicating capitalization. This baseline system, which includes a few elementary composite signs, is called Grade 1 Braille. Producing text in Grade 1 Braille requires a lot of space. One page of a Braille book contains approximately 1000 Braille symbols, whereas one page of a printed book may contain 3500 characters. Therefore a number of rules were introduced to reduce the volume of a Braille coded text by 20–30%. (This is comparable to using shortcuts such as “asap” for “as soon as possible” or the sign “4” for the word “for” in SMS or e-mails.) This has resulted in an extensive system of so-called contractions and the rules for their application. For instance, if “&” can be used for “and”, then “h&le” stands for “handle”.
The complete system is called Literary Braille Code, Grade 2 Braille or contracted Braille. Most blind people who read Braille can also read the contracted version. Unfortunately, the contractions are language specific and therefore the code for contracted Braille must be learnt for each language separately. In German speaking countries, the terms Vollschrift (full text) and Kurzschrift (short text) are used for Grade 1 and Grade 2 Braille. Different organizations coordinate the standardization in the different countries, as shown in Table 14.7. Different approaches are required in some non-European language systems. For instance, Chinese Braille symbols represent the sounds of the spoken language rather than the characters of the written language.
530 14 Speech, Text and Braille Conversion Technology
Table 14.7. Examples for organizations and resources of the Braille system for different languages
Language |
Organization |
Examples for web resourcesa |
English |
International Council on English Braille (ICEB) |
www.iceb.org |
|
Braille Authority of North America (BANA) |
www.brailleauthority.org/ |
|
Braille Authority of the United Kingdom (BAUK) |
www.brl.org/ebae/ |
|
|
www.bauk.org.uk/ |
French |
F´ed´eration des Aveugles et Handicap´es Visuels |
www.faf.asso.fr/sommaire.htm |
|
de France |
|
German |
Brailleschriftkommission der deutschsprachigen |
www.fernuni-hagen.de/ |
|
L¨ander |
ZFE/fs/download.htm |
Chinese |
China Association of the Blind |
www.braille.ch/pschin-e.htm |
|
|
www.omniglot.com/writing/ |
|
|
braille_chinese.htm |
|
|
www.hadley-school.org/Web_SIte/ |
|
|
8_d_chinese_braille_alphabet.asp |
Japanese |
Japan Kanji-Braille Association |
kantenji.jp |
|
|
www.geocities.co.jp/ |
|
|
CollegeLife-Library/7524/ |
|
|
tenji/tenyaku.html |
|
|
|
aA general overview on the codes in various languages is presented in homepages.cwi.nl/ dik/english/codes/braille.html
There are also special forms for the following purposes, amongst others:
•Musical notation (www.brl.org/music/index.html; for details see Chapter 16).
•Mathematical notation; although there is a universal mathematical text notation, this is unfortunately not the case for Braille and different systems of mathematical Braille have developed in different countries. This has a number of disadvantages, including making both the production of mathematical Braille and communication between blind mathematicians in different countries more difficult. The US version is known as The Nemeth Braille Code.
•Scientific notation, for instance for chemistry (www.brl.org/chemistry/).
•Phonetic notation (www.clauchau.free.fr/L/phonalph.html).
•Computer Braille, which uses an 8 dot system for 256 signs, analogously to the ASCII code (see www.braille.org/papers/unive/unive.html or DIN 32 982).
With regards to text-to-Braille conversion, there are clearly additional difficulties associated with the production of contracted Braille. Some remarks on this topic can be found in the first volume of this series (Hersh and Johnson 2003, pp 262– 265). The following two subsections consider hardware and software aspects of text-to-Braille and Braille-to-text conversion. It is often useful to provide both Braille and speech output and examples of dual output systems are presented in Section 14.6.
14.5 Braille Conversion |
531 |
14.5.2 Text-to-Braille Conversion
Historical development
Analogously to speech-related technologies, the development of modern reading machines has required the availability of computer technology. However, the requirement for tools for blind people to read printed texts and thus enhance their independence was recognized decades earlier. The discovery of the photoelectric effect allowed printed signs to be converted into electrical signals, which were presented in audio form in the earliest reading devices. Early devices with audio output included the following:
•The Optophone was invented in 1914 by E.E. Fournier-D’Albe, a Professor in Birmingham, UK. In the optophone the printed text characters are irradiated with light pulses and the reflected light is converted into an audible signal by means of a Selenium cell. The listener uses the different sounds to distinguish between the different text symbols. The invention was subsequently modified and applied in several different devices.
(See www.oldweb.northampton.ac.uk/aps/eng/research/optophone/ optophone2.html.)
•Another early reading machine was constructed by Rosing, a Professor in Petersburg before 1917. The text characters were scanned by an optical beam in the horizontal and vertical directions to produce a Morse-like sound.
The complexity and lack of an intuitive relationship between the text and the acoustic patterns and the lack of synthetic speech output led to the development of reading devices with a tactile output, including the following:
•W. Thorner, an eye specialist in Berlin, obtained a patent in 1916 for converting images to electro-tactile patterns. Later inventions used grids of electromagnets for the production of mechanical stimuli. This approach is still used for presenting graphical information (see Chapter 4 of this book). However, it proved to be unsuitable for reading text unless the magnetic elements were combined to give Braille symbols.
•G. Schutkowski, a teacher at a school for blind people near Berlin, was probably the first person to recognize that the most appropriate tactile output for reading machines was Braille. He invented an optical pattern matching method for reading (Schutkowski 1938). The reading mechanism was coupled with a drum which moved the corresponding Braille symbol in front of the user (Figure 14.19a). Remarkably, he also investigated the addition of speech output to his device (Schutkowski 1952) and realized that the spelling mode which he developed was only the first step towards a full text-to-speech system.
The current situation is summarized in Figure 14.20. Technical support systems for a human transcriber are presented in the upper part of the figure. The Brailler has already been mentioned. It can be replaced by a PC with driver software which changes the standard keyboard to a six-key input device. Automatic text-to-Braille conversion is illustrated in the lower part of the figure. The components of this process will now be described.
532 14 Speech, Text and Braille Conversion Technology
Figure 14.19a,b. Old and new devices for volatile Braille output: a the drum of the early prototype of a text-to-Braille machine by G. Schutkowski (contemporary press photograph from the early 1950s); b a Braille display with 40 elements (type SuperVario), mounted in front of a PC keyboard (by courtesy of BAUM Retec AG)
Input devices
The text to be converted needs to be available in machine-readable form. If this is not the case, a text file can be produced in one of the following standard ways:
•By using a PC with a conventional text processing system.
•By using a scanner to input the printed text. The scanner is connected to a standard PC which hosts the optical character recognition (OCR) software which produces the text file. Since errors cannot be completely avoided with the current state of OCR technology, the word processing software of the PC should be used to edit the resulting file and correct any errors.
The scanner is the essential input device for a reading machine. Since it performs similarly to other OCR applications, its properties will not be discussed here. They are described in more detail in Chapter 15 and the reader is referred to Bunke and Wang (1997) for additional information.
Output devices
The output device used depends on whether a durable but bulky paper version of the converted text or a quick to produce but transient (volatile) version is required.
Producing Braille symbols on paper using an automatic system requires the manual keys of the traditional Brailler to be replaced by a corresponding computer interface, called a Braille embosser. A number of different types of embossers are available and an illustration is presented in the first volume of this series (Hersh and Johnson 2003, p 265). See also Chapter 12 of this book. Some types can produce dots on both sides of the page (interpoint Braille). An electronic Brailler can be used as a computer output device also (Figure 14.18b).
Figure 14.20. Data flow in text-to-Braille processing. The diagram shows how the work of human Braille transcribers (above) is automated by suited hardware and software components. Additionally, the inclusion of text-to-speech components in the system is illustrated
533 Conversion Braille 5.14
534 14 Speech, Text and Braille Conversion Technology
Braille displays are used to provide the user with refreshable (volatile) Braille output. They consist of cells which each produce a touchable Braille symbol by using a piezoelectric drive to raise a number of small pins. Displays consist of lines of 20, 40, or 80 cells, as shown in Figure 14.19b. Braille displays were introduced by the company F.H. Papenmeyer in 1975. Modern Braille cells have 8 dots and can therefore be used for computer as well as literary Braille.
Conversion software
Text-to-Braille conversion software is the main component of Figure 14.20. It needs to carry out the following different groups of tasks:
•Mapping the text characters to Braille symbols. This is the central module of the conversion process. It should be noted that this mapping is not one-to-one mapping to the use of a large number of contractions and special rules, which are language-dependent as it was explained already.
•Text analysis. The use of contractions is governed by a number of specific rules. These include shortcuts for prefixes and suffixes, preference rules for ambiguous sequences and contractions which cannot be applied over syllable boundaries. Therefore, the text-to-Braille mapping module needs to “know” the morphological structure of the words and this is the task of a special module. Morphological analysis is a typical AI problem, which is apparently relatively easy to carry out for a person, but difficult for an automatic system. The accuracy of the text-to- Braille conversion depends on the morphological analysis. However, there are frequently morphological ambiguities, which can only be resolved only at the semantic level. Morphological analysis is also a component of the text analysis module of a TTS system, as shown in Figure 14.15. In both cases, certain types of errors are difficult to avoid and lead to prosodic errors in TTS and transcription errors in text-to-Braille conversion.
•Producing the correct layout from the Braille symbols. There are clearly defined layout rules for books and other documents in Braille in order to facilitate navigation and structure recognition by blind readers. The conversion software must also consider these layout rules, particularly if an embosser is used to produce a paper document. The layout rules for English Braille documents are summarized in the document “Braille Formats: Principles of Print to Braille Transcription 1997” (www.brl.org/formats/).
As shown in Figure 14.20, the conversion software works on a file-to-file base. This has two main advantages and one drawback:
•Using a text file for input is convenient for storing and re-using the data and is also required to enable removal of any errors introduced by the scanner OCR software using word processing software. This is easier than correcting the Braille output.
•The Braille file can be evaluated and corrected using special software tools. This is necessary to remove errors from the text-to-Braille conversion. These
