Around 1971, Ray Tomlinson developed the idea of networked electronic mail, which was hugely attractive to me because it replaced uncertain voice calls with the clarity of text. The development of the Internet was undertaken in the context of heavy use of email.
The rise of video conferencing has actually been a huge challenge for me as it reintroduces some of the uncertainty of voice calling and I look forward to real-time, automatic captioning to overcome the limitations that medium poses for me.
What message do you have for people creating technology today and how they should think about accessibility?
It must be thought through during the design phase of any product. Accessibility and ease of use go hand in hand. Many people experience temporary disability (broken arm, leg, finger, blocked ears…) and appreciate the value of accessibility features from that experience. There is no excuse for making products that are not accessible.
Appreciation for plain text and written communication is part of the Unix philosophy on which the Internet was built. Unix is “the geek Gilgamesh epic; it’s a tale told in plain text.”
Authors and writers of all stripes can learn a lot about creating and managing words from computer programmers, beginning with an appreciation for the simple, durable efficiencies of plain text. Anybody running Unix, Linux, or BSD already knows all about text, because it’s the third prong of the Unix Tools Philosophy:
Write programs that do one thing and do it well;
Write programs that work together;
Write programs to handle text streams, because that is a universal interface.
The geeks who made Unix nearly 40 years ago made plain text the universal interface because they believed in economy, simplicity, and reliability.
If Unix is the geek Gilgamesh epic, it’s a tale told in plain text.
I usually publish to my WordPress blogs via Ulysses on both macOS and iOS. Occasionally, I check out the state of the WordPress app and wordpress.com’sCalypso web interface on iOS. A litmus flow for me is image captioning.
WordPress iOS App
In the screenshot below, three freshly uploaded images are shown in the editor. When I tap an image, I see Edit in the resulting menu. I’m on the right track, yet already confused. I’m not sure which image I’m operating on. The selection indicator is too subtle, requiring me to get closer to the screen.
If the cursor is located after the selected image, the editor scrolls down to center the cursor upon exiting the Media Options flow. The image I was editing is now partially offscreen, increasing ambiguity.
This ambiguity creates anxiety. Which image am I editing? The “Media Settings” page offers no context.
My go-to anxiety flow in the face of such ambiguity is to go back and reorient. If I “Cancel” to return to the editor in order to establish context, the editor offers more ambiguity instead of reassurance by scrolling down as mentioned above. Even without the scrolling, the blue selection indicator requires me to squint. Selection visually collides with the cursor (which is image height when on the same line as an image), increasing ambiguity further.
I write descriptive captions in the interest of accessibility. I need to see the image as I do this. Here’s what image captioning looks like in Ulysses on macOS.
And here’s what it looks like in Ulysses on iOS. A little scrolling back and forth between the image and description field is needed, but at least they’re on the same page. I’ll gladly scroll if it means getting images large enough for my eyes.
In both of those interfaces, the image is available for reference while captioning. Contrast them with the WP iOS app. The “Caption” screen consists of a single text input field. The image is not displayed. No information about the image is displayed. This means I’ll have to flip back and forth between the WP app and my camera roll app to write a caption.
If I want to consult the image from within the WP app instead of flipping to a different app, the journey is: two taps back to the editor, a bad scroll interaction depending on cursor location, peer over my bifocals at images and selection indicators, and then another three taps back to the Caption field. I’d have to do this over and over to transcribe the screenshots in this post. I started this post in the iOS app and quickly tired.
Calypso on iOS Safari
The iOS app’s caption flow does not work for me. So how about captioning flow on the mobile web? Alas, Calypso on iOS Safari is buggy, erratic, and frustrating to the point that I usually give up on it and go get the laptop. Sometimes, though, I can complete an editing session. In this shot, I’m adding a caption as part of image insertion flow. The image thumbnail is on the small side for me. I need big images when writing captions, especially for screenshots. Otherwise, I have to find the image in an interface that gives me a better view and then correlate back and forth.
In the following screenshots, I’m adding a caption to an image after it has been inserted. First, delicately dismiss the Cut Copy bar without dismissing the inline image toolbar hiding behind it. This is fussy and awkward.
And, then, tap the caption button, wonder why it didn’t do anything, scroll down, realize a caption input unfurled below the fold, and start adding a caption.
There’s the possibility of good flow friendly to presbyopia beneath the unfortunately numerous interaction bugs. Though, even with the interaction bugs, at least I don’t have to caption an image I can’t see. There are many times I wish I could use the mobile web interface, but the scroll bleed, vscroll loss, keyboard flyup, lock ups, crashes, requests for more memory, and general unpredictability exclude it from consideration.
Neither interface meets my needs for captioning flow. I need images to be present on the same screen as the fields that describe them. I need access to image views large enough for my presbyopic eyes to transcribe text from screenshots. I need caption fields with enough room to comfortably compose detailed image descriptions.