Open main menu
Home
Random
Recent changes
Special pages
Community portal
Preferences
About Wikipedia
Disclaimers
Incubator escapee wiki
Search
User menu
Talk
Dark mode
Contributions
Create account
Log in
Editing
Screen reader
(section)
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
== Types == === Command-line (text) === In early [[operating system]]s, such as [[MS-DOS]], which employed [[command-line interface]]s (<abbr>CLI</abbr>s), the screen display consisted of [[Character (computing)|characters]] [[Memory-mapped I/O|mapping]] directly to a [[screen buffer]] in [[Computer memory|memory]] and a [[Cursor (computers)|cursor]] position. Input was by keyboard. All this information could therefore be obtained from the system either by [[hooking]] the flow of information around the system and reading the screen buffer or by using a standard hardware output socket<ref>{{cite web | title=Talking Terminals. BYTE, September 1982 | url=http://www.edstoffel.com/david/talkingterminals.html | access-date=September 7, 2006 |archive-url = https://web.archive.org/web/20060625225004/http://www.edstoffel.com/david/talkingterminals.html <!-- Bot retrieved archive --> |archive-date = June 25, 2006}}</ref> and communicating the results to the user. In the 1980s, the Research Centre for the Education of the Visually Handicapped (<abbr>RCEVH</abbr>) at the [[University of Birmingham]] developed a Screen Reader for the [[BBC Micro]] and <!--abbr?-->NEC Portable.<ref>Paul Blenkhorn, "The <abbr>RCEVH</abbr> project on micro-computer systems and computer assisted learning", British Journal of Visual Impairment, 4/3, 101-103 (1986). [http://www.visugate.biz/bjvi/1986/autumn1986.html#RCEVH Free HTML version at Visugate].</ref><ref>{{cite web | title=Access to personal computers using speech synthesis. RNIB New Beacon No.76, May 1992 | date=March 3, 2014| url=http://www.rnib.org.uk/information-everyday-living-using-technology-beginners-guides/beginners-guide-assistive-technology}}</ref> === Graphical === ==== Off-screen models ==== With the arrival of [[graphical user interface]]s (<abbr>GUI</abbr>s), the situation became more complicated. A <abbr>GUI</abbr> has characters and graphics drawn on the screen at particular positions, and therefore there is no purely textual representation of the graphical contents of the display. Screen readers were therefore forced to employ new low-level techniques, gathering messages from the [[operating system]] and using these to build up an "off-screen model", a representation of the display in which the required text content is stored.<ref>According to "[ftp://service.boulder.ibm.com/sns/sr-os2/sr2doc/guitalk.txt Making the <abbr>GUI</abbr> Talk]{{dead link|date=May 2025|bot=medic}}{{cbignore|bot=medic}}" (by Richard Schwerdtfeger, ''BYTE'' December 1991, p. 118-128), the first screen reader to build an off-screen model was outSPOKEN.</ref> For example, the operating system might send messages to draw a command button and its caption. These messages are intercepted and used to construct the off-screen model. The user can switch between controls (such as buttons) available on the screen and the captions and control contents will be read aloud and/or shown on a [[refreshable braille display]]. Screen readers can also communicate information on menus, controls, and other visual constructs to permit blind users to interact with these constructs. However, maintaining an off-screen model is a significant technical challenge; hooking the low-level messages and maintaining an accurate model are both difficult tasks.{{citation needed|date=January 2015}} ==== Accessibility APIs ==== Operating system and application designers have attempted to address these problems by providing ways for screen readers to access the display contents without having to maintain an off-screen model. These involve the provision of alternative and accessible representations of what is being displayed on the screen accessed through an <abbr>[[Application programming interface|API]]</abbr>. Existing <abbr>API</abbr>s include: * Android Accessibility Framework<ref>[https://developer.android.com/training/accessibility/index.html Implementing Accessibility on Android].</ref> * Apple Accessibility <abbr>API</abbr><ref>[https://developer.apple.com/documentation/Accessibility/Reference/AccessibilityLowlevel/index.html Apple Accessibility <abbr>API</abbr>].</ref> * [[AT-SPI]] * [[IAccessible2]] * [[Microsoft Active Accessibility]] (<abbr>MSAA</abbr>) * [[Microsoft UI Automation]] * [[Java Access Bridge]]<ref>{{cite web|url=http://java.sun.com/products/accessbridge/|title=Oracle Technology Network for Java Developers β Oracle Technology Network β Oracle}}</ref> Screen readers can query the operating system or application for what is currently being displayed and receive updates when the display changes. For example, a screen reader can be told that the current focus is on a button and the button caption to be communicated to the user. This approach is considerably easier for the developers of screen readers, but fails when applications do not comply with the accessibility <abbr>API</abbr>: for example, [[Microsoft Word]] does not comply with the <abbr>MSAA</abbr> <abbr>API</abbr>, so screen readers must still maintain an off-screen model for Word or find another way to access its contents.{{citation needed|date=January 2015}} One approach is to use available operating system messages and application object models to supplement accessibility <abbr>API</abbr>s. Screen readers can be assumed to be able to access all display content that is not intrinsically inaccessible. Web browsers, word processors, icons and windows and email programs are just some of the applications used successfully by screen reader users. However, according to some users,{{Who|date=January 2015}} using a screen reader is considerably more difficult than using a GUI, and many applications have specific problems resulting from the nature of the application (e.g. animations) or failure to comply with accessibility standards for the platform (e.g. Microsoft Word and Active Accessibility).{{citation needed|date=January 2015}} === Self-voicing programs and applications === Some programs and applications have voicing technology built in alongside their primary functionality. These programs are termed [[self-voicing]] and can be a form of [[assistive technology]] if they are designed to remove the need to use a screen reader.{{citation needed|date=January 2015}} === Cloud-based === Some telephone services allow users to interact with the internet remotely. For example, TeleTender can read web pages over the phone and does not require special programs or devices on the user side.{{citation needed|date=January 2015}} [[Virtual assistant]]s can sometimes read out written documents (textual web content, <abbr>[[Portable Document Format|PDF]]</abbr> documents, e-mails etc.) The best-known examples are Apple's [[Siri]], [[Google Assistant]], and [[Amazon Alexa]]. === Web-based === A relatively new development in the field is web-based applications like Spoken-Web that act as web portals, managing content like news updates, weather, science and business articles for visually-impaired or blind computer users.{{citation needed|date=January 2015}} Other examples are ReadSpeaker or [[BrowseAloud]] that add [[text-to-speech]] functionality to web content.{{citation needed|date=January 2015}} The primary audience for such applications is those who have difficulty reading because of learning disabilities or language barriers.{{citation needed|date=January 2015}} Although functionality remains limited compared to equivalent desktop applications, the major benefit is to increase the accessibility of said websites when viewed on public machines where users do not have permission to install custom software, giving people greater "freedom to roam".{{citation needed|date=January 2015}} This functionality depends on the quality of the software but also on a logical structure of the text. Use of headings, punctuation, presence of alternate attributes for images, etc. is crucial for a good vocalization. Also a web site may have a nice look because of the use of appropriate two dimensional positioning with CSS but its standard linearization, for example, by suppressing any CSS and Javascript in the browser may not be comprehensible.{{citation needed|date=January 2015}}
Edit summary
(Briefly describe your changes)
By publishing changes, you agree to the
Terms of Use
, and you irrevocably agree to release your contribution under the
CC BY-SA 4.0 License
and the
GFDL
. You agree that a hyperlink or URL is sufficient attribution under the Creative Commons license.
Cancel
Editing help
(opens in new window)