This post responds to concerns in ComputingEd post ‘Kindles versus Smart phones: Age matters, testing matters’. A UGa study and commentary focus on news reading as screen-dependant and vision-only. I suggest considering the print-disabled TTS-dependant ecosystem to expand understanding of human reading and assistive device capabilities.
Reading experiments might be broadened to include pure TTS, i.e. no screens. But first, what criteria matter: reading rate, absorption level; device comfort, simulated print experience, distribution costs and convenience,..?
For the record, I just read this article by RSS, then switched to my Newstand, downloaded NYTimes and other papers from Bookshare.org, cooperating with NFB Newsline, and news companies I gratefully thank. Papers are delivered wirelessly in XML-based DAISY format, retrieved and read on a Linux-powered mobile device (Levelstar Icon), spoken in an old-style “robotic voice”. For delivery efficiency and cost, this cannot be beat and I think I absorb selective news reading better than ever. But how is experience of print-disabled news readers factored into comparisons like this article?
This will soon be relevant if Kindle, iPod/iTouch, etc. TTS reading is fully enabled and adopted by some readers from proprietary delivery systems, like Amazon. For proper evaluation, it will be necessary to compare eReading by TTS on mainstream devices to that provided by evolved readers like APH book port, Humanware Victor Reader Stream, PlexTalk Pocket, Levelstar Icon, and (my favorite) GW Micro booksense. Also important is the media format, currently favored as DAISY on these devices. And finally is the provision of media, currently limited legally to print-disabled readers, as by NFB (National Federation of Blind) and non-profit Bookshare.org. In other words, there’s another ecosystem of reading open only to print-disabled that might benefit those attracted to eReading.
Oh, my, here’s the “universal design” mantra again. ‘Reading news by screen’ is, of course, more limited than ‘reading by print or audio”. It’s possible than for some reading criteria the screen-free mode or open XML-based format and its reading devices and experienced reader population may beat mainstream strategies!
Could these experiments be performed? Certainly, most universities have students who currently, or could, offer their experience with equipment provided through Disability Services. Fact quizzes and comprehension tests might raise questions about how our reading brains work and how well our reading devices and formats help or hinder. What research is in progress? Is there a CS agenda for this social and economic ecosystem? Why do people think reading is a vision-only activity? Ok, comics, photos, and crosswords are a bit challenging, but plain old print is so well handled by TTS. Let’s open our eyes and ears and fingers to a fuller range of capabilities. I would love to be a test subject for eReading experiments.