Skip to content →

Month: April 2022

Variations on an Automatic Image Description

Reading through Twitter today, the following tweet showed up on the timeline of one of the people I follow as a retweet.

DocšŸ• – @DocAtCDI: A truck loaded with thousands of copies of Roget’s Thesaurus spilled its load leaving New York

Witnesses were stunned, startled, aghast, stupefied, confused, shocked, rattled, paralyzed, dazed, bewildered, surprised, dumbfounded, flabbergasted, confounded, astonished, and numbed.

I found the tweet amusing and was going to retweet but noticed it had a picture without any alt text. This lead me to be curious what was in the picture. From the tweet text, I’m assuming some form of vehicles on a road with a bunch of books scattered about is most likely.

I suspect most reading this know that iOS has the ability to automatically describe pictures. This functionality started in iOS 14. When using VoiceOver you can have a short machine-generated description of pictures such as the one attached to the tweet here.

Newer versions of iOS extended this functionality to include a feature called Explore Image. That allows you to use VoiceOver to step through individual objects recognized in the image. It can be accessed with a rotor option when focussed on the image. Here is where the experience gets a bit interesting.

My go to Twitter app on the iPhone is Twitterific. The accessibility of the app has been outstanding for years and the maker has been highly responsive if issues to creep in.

I’ve also been exploring another highly accessible Twitter app named Spring. So far I’ve had a great experience with this app as well.

As one would expect, both Twitterific and Spring offer the ability to view images included with tweets. When images are viewed in either app, the VoiceOver automatic image description and Explore Image functionality work. Differences in the same picture viewed in two different apps using the same automatic image description and exploration technology are plainly obvious though.

First off, the automatic description when viewing the image in Twitterific says:

an illustration of vehicles on a road X. VETERAN’S PLUMRNO. Rall

That same image viewed in Spring yields the following automatic description:

a group of cars driving on a highway ETERAN ‘S PLUMPING

Both descriptions mention that the picture deals with vehicles on a road in some fashion. and include what I’d suspect is the text of a sign on a van or truck in the picture from a plumbing company. Again the descriptions come from Apple, not the individual apps.

A picky point but cars do not drive, people drive them. I might not know what is in the photo for certain but I am quite confident it isn’t a bunch of Teslas with the self-driving mode engaged.

It is also interesting how the image description when using Spring is a bit more detailed. It uses the terms highway and cars, whereas the Twitterific version is more generic in nature. The detail about cars when using Spring is even more interesting when using the Explore Image feature to review the individual objects in the picture.

Again, the newest versions of iOS added a feature called Explore Image to VoiceOver. Focus an image, change the VoiceOver rotor to Actions and one of the choices will be Explore Image. This opens a pop-over experience with individual objects from the picture. You can use VoiceOver previous and next commands to move from object to object and have them highlighted visually in the picture.

Here are the objects from the picture in the tweet I mentioned when explored with Twitterrific:

  • Automobile near left edge
  • Automobile Centered
  • Automobile near right edge

Recall how the automatic description for Spring talked about cars driving on a highway? One can only wonder where the cars went and where the train came from when using the Explore Image feature. Here is what is reported when exploring the image in Spring.

  • Van near bottom-left edge
  • Van near right edge
  • Van near bottom-left edge
  • Train near top edge

Automatic image descriptions are another helpful tool for shaping the accessibility landscape. They’ll be even more impactful if the technology continues to advance to reduce the variability of something as simple as viewing an image in a different program seems to introduce and the accuracy and detail of what is described improves.

Leave a Comment

What Happens After the Accessibility Agreements End?

It is unfortunate that it takes landmark agreements far too often for progress on accessibility in the first place. Several years ago Major League Baseball was heralded for their agreement and commitment to accessibility. Progress was indeed made and I know I was one who commented on the agreement and progress.

The agreement closing date has long past and so the answer to the question about what happens afterword in their case is, ā€œUnlabeled image Unlabeled image /121.svg.ā€ If you don’t like that answer, how about ā€œUnlabeled image Unlabeled image /140.svg?ā€ Both of those answers are taken from www.mlb.com/scores as of this writing, and are how two teams are indicated from the live scoreboard when accessed with multiple screen readers.

They are just two of the more than 370 accessibility failures reported by multiple accessibility auditing tools run on the page. Of course, accessibility tools are just part of the story and I’m a big fan of user testing for accessibility. In this case I’m serving that role and just wanted to look up the score of a ballgame. I guess when I sort out who is represented by 121 or 140, I’ll know the score.

This is not a criticism of the efforts that lead to MLB’s agreement. It is, however, a serious question around what it is going to take to make accessibility matter long after the agreements end because I can assure you representing a baseball team as 121, 140 or any other random nonsense is not accessibility the way I want to experience it.

Leave a Comment