INCLUSIVE DESIGN

Twitter reminds us about alt-text, but how good are we at it?

Several humble observations on how people describe photos, artworks, and memes

Slava Shestopalov 🇺🇦
Design Bridges
Published in
8 min readOct 22, 2022

--

Look at this:

White text on black background says: “Dr. Serena Auñón-Chancellor examines her eye using a fundoscope aboard the International Space Station. She is wearing a charcoal-grey shirt, smartwatch, and silver necklace. The fundoscope is light grey and J-shaped, with blue and black accents. Behind her a hallway in the ISS with blue handles and white walls.”
Image description example #1.

What did you imagine after reading the text? How clear and detailed was it? How did you visualize the fundoscope?

Done? Good. We’ll return to that in a minute.

And what about this picture?

White text on black background says: “Drawing of a jewellery brooch.”
Image description example #2.

What brooch did you envision? What were its size, shape, and color? How exactly was it drawn in your imagination?

These are real examples of alt-text from Twitter.

Twitter has been quite vigorous recently in raising accessibility awareness among its audience and proactively reminding people to add descriptions to tweets with images.

Screenshot of the Twitter website with the “New tweet” popup overlaid with the “Don’t forget to make image accessible” popup. The “New tweet” popup shows a photo of a Berlin demonstration against unprovoked Russian aggression towards Ukraine. People are wearing blue-and-yellow flags on theot holders and carry anti-war posters. The “Don’t forget to make image accessible” popup includes the explanation why alt-text is necessary and has two buttons: “Add description” and “Not this time.”
The image description reminder I see every time I’m trying to tweet an image.

So, let’s get back to the descriptions from the beginning of this article and see that fundoscope and the brooch.

Here is an actual picture presented by the first description:

A photo by NASA with the description: “Dr. Serena Auñón-Chancellor examines her eye using a fundoscope aboard the International Space Station. She is wearing a charcoal-grey shirt, smartwatch, and silver necklace. The fundoscope is light grey and J-shaped, with blue and black accents. Behind her a hallway in the ISS with blue handles and white walls.”
A photo from NASA’s tweet about Dr. Auñón-Chancellor on the International Space Station.

Quite expected, isn’t it?

And which of the four brooches in the picture below, in your opinion, matches the second description?

Four jewellery brooch drawings. The first one is a pencil drawing of a six-pointed ornate brooch on a yellowish paper. The second one is a watercolor drawing of a golden brooch depicting seven dancers in red enamel dresses on white paper. The third one is a colored realistic triptich drawing of a shell-shaped golden brooch from the front, rear, and side. The fourth picture is a pencil drawing a diamond-shaped brooch and two different earrings on a yellowish paper with handwritten text.
Which of these brooches corresponds to the description above?

Hmm… Any of them?

Now, the correct answer. It’s the golden brooch drawn on white paper; it depicts seven dancers in red enamel dresses and is adorned with many small gemstones.

The original description in the tweet is quite generic, so if you google “drawing of a jewelry brooch,” it will match most of the search results. Here is the screenshot, by the way:

Screenshot of a Twitter post with a watercolor drawing of a golden brooch depicting seven dancers in red enamel dresses on white paper. The popup “Image description” says: “Drawing of a jewellery brooch.”
The Design Museum’s tweet with a photo of a jewelry brooch with a generic image description.

But why are we talking about image descriptions, or “alt-text,” in the first place? Let’s quickly recap.

Image description helps blind people and people with low vision to consume visual content. They browse the internet using screenreaders — programs that transform text content into speech. Besides, image descriptions are helpful for people with a poor internet connection or outdated devices: they can guess what’s depicted in a picture until it fully loads or if it doesn’t load at all.

The need for accessibility features is just a matter of time. We all will gain some disabilities as we grow old.

Content accessibility is hard to automate or implement once and for all. Each piece of non-textual content — pictures, audio, and video — should have a well-composed alternative so that people who cannot rely on one of their senses can still consume information.

Facebook and Instagram employ AI to auto-generate descriptions, but it’s of little help. The generic sentence “May be an image of multiple persons outdoors” fits half of all the photos on social media. So, as of 2022, the best description is the one you compose yourself.

Here is a vivid example from my Insta:

Screenshot of an Instagram page with switched off images. Alt-text on a grey rectangle says: “Post by Slava Shestopalov on October 14, 2022 tagging @demchenkoksu. May be an image of 1 person, monument and outdoors.”
Example of an Instagram post with the auto-generated description that says, “May be an image of 1 person, monument and outdoors.”

The actual photo doesn’t contain any recognizable face; there is just a crowd of people and a well-known building, the Brandenburg Gate in Berlin, not just a random monument. And, of course, AI cannot “see” a large artwork with six sunflowers projected on the gate’s facade during the Festival of Lights (not to mention that this year’s festival is devoted to the ideas of peace, unity, freedom, and resistance against the Russian military aggression and genocide of Ukrainians).

Screenshot of instagram post of the Brandenburg Gate in Berlin at night. A crowd of people stands in front of the gate. An artwork with six sunflowers is projected on the gate so that sunflower occupy the gate columns. The Quadriga statue at the top is bright illuminated.
The actual photo described by Instagram.

Now let’s take a look at some inspiring examples of image descriptions.

In my opinion, NASA’s account on Twitter is a great role model in terms of composing image descriptions. They literally describe what’s depicted and show tons of creativity in making it exciting.

Screenshot of NASA’s tweet with the description (cited partially): “Two large, very bright galaxies dominate the center of the image. The elliptical galaxy at left is extremely bright at its circular core, with dimmer white light extending to its transparent circular edges. At right is a bright spiral galaxy. It also has a bright white core, but has red and light purple spiral arms that start at the center and turn clockwise going outward.”
NASA’s tweet with an image of two galaxies and a detailed image description.

Generic descriptions like “two galaxies on the dark background” would probably fit half of all their publications. But specifying size, shape, location, orientation, movement, and color shows respect to readers who cannot see photos for whatever reasons.

A good alt-text manages to substitute the visual perception of an image with arousing readers’ imagination.

Try this: open NASA’s account via the Twitter app, turn on the accessibility mode on your phone, close your eyes, and browse the feed without looking at it. Scroll through some posts. Then open your eyes and compare what you imagined with the actual pictures.

Or you can also use the “Images on/off” Google Chrome plugin to expose all image descriptions on a webpage.

Screenshot of NASA’s tweet with the description: “Image of galaxy IC 5332 as taken by the Webb telescope’s MIRI instrument, resembling gray cobwebs in the shape of a spiral. These “cobwebs” are patterns of gas spread throughout the galaxy. The core of the galaxy glows a dark blue. Stars, seen as tiny blue dots, are scattered throughout the image. There are also sparser, larger red dots spread out among the spiral arms. The background of the image is dark.”
NASA’s tweet with an image of a “goth” galaxy hidden in the dust and a detailed image description.

Sadly, not all of Twitter’s content is equally accessible.

Look at the tweet by Adobe in the screenshot below. All of its three images have descriptions. But they are all the same, although a sighted person will definitely notice that the depicted characters are different.

Screenshot of Adobe’s tweet with three artworks depicting women in ethnic attires. The “Image description” popup on the first image says: “Illustration of a woman in traditional adornments and attire.” The image shows a dark-skin-tone woman with black long hair with red stripes. She wears an orange-crimson-striped blouse and a dark skirt. The woman is shown in motion from the side.
Adobe’s tweet with three images of women in traditional attires and the same generic image descriptions for all three pictures: example 1.
Screenshot of Adobe’s tweet with three artworks depicting women in ethnic attires. The “Image description” popup on the second image says: “Illustration of a woman in traditional adornments and attire.” The image shows a dark-skin-tone woman in a white theater mask and richly decorated dark hair. She wears ornate clothes on her shoulder and holds a flower.
Adobe’s tweet with three images of women in traditional attires and the same generic image descriptions for all three pictures: example 2.

Since we are talking about Twitter, the next logical question is: how does one describe memes?

I should admit this is the least accessible part of Twitter content. I had a hard time finding a decent meme with alt-text. Most memes don’t have any description, and among those that do, you can hardly find a funny one. The challenge with many memes is that you should describe both the template used and custom captions that make it comical.

Look at the example with pensive Macron:

Screenshot of a tweet with a meme about French president Emmanuel Macron. The image description says: “Photo collage of Macron looking anguished. Text: “with you can’t surrender unconditionally cause it’s not your country.”
A tweet featuring a meme about French President Emmanuel Macron with a concise image description.

The first sentence introduces the meme format, and the second one cites the caption. If we lived in the ideal world, where all images on the web had perfect alt-text, there wouldn’t be a need to describe Macron’s appearance. But since we are not there yet, a short sentence about his look might be a reasonable improvement, I guess. Apart from that, it’s an example of a simple and robust description.

But things get way trickier with complex memes or data visualization (infographics, charts, diagrams). The example below demonstrates both complicating factors:

Screenshot of a tweet with an infographics-style meme that shows the comparison of two characters. The information is color-coded with pink and blue colors and shows a comparison of different characteristics on the sliders, mini-charts, and human silhouettes. The “Image description” popup is open and shows three paragraphs of text describing the image.
A tweet featuring the “enemies meme” with two characters, Higgy and Maxy, and a detailed image description (three paragraphs of text).

This meme is described in three paragraphs, which gives enough details but is a bit hard to comprehend. Unfortunately, I didn’t understand it, but maybe its topic is just beyond me. Anyway, it’s a pretty good example of how hard it might be to write image descriptions for pictures with lots of details and text.

Another interesting example I found recently is a post with two pigeon paintings and a reference to the “moral alignment” meme format with a 3×3 grid and labels like “lawful good” or “chaotic evil.”

Three screenshots of a tweet with two pigeon paintings side by side. The first screenshot shows the tweet, and the other two show popups with image descriptions of each artwork. The first pigeon is sitting on the pink background in the pose as if it feels cold and wants to warm up. The second pigeon stands on the pink background and has two laser beams coming from its eyes. Image descriptions only say “Good pigeon” and “Evil pigeon” and authors copyright in the brackets.
A tweet by Leah Gardner with the “evil pigeon” and “good pigeon” paintings and generic image descriptions.

The first image description in the tweet says, “Evil pigeon (oil painting by Leah Gardner),” and the second one is, “Good pigeon (oil painting by Leah Gardner).” The author prioritized her copyright info and kept silent about the pigeon poses, pink background, or hilarious laser beams flashing from the eyes of the “evil” pigeon.

As you see, not all Twitter folks realize what they are supposed to write in the image description. I noticed that the best alt-texts usually belong to official accounts of well-known organizations with a global audience (mass media, museums, research institutes) and, of course, organizations and experts in the area of accessibility and inclusion.

So, we’ve already covered sceneries, artworks, and memes. Now, a couple of observations about portraits.

Screenshot of The Metropolitan Museum of Art tweet with an image of a speaker and the following image description: “A light-skin-toned woman with brown hair stares confidently into the camera wearing an ochre plaid jacket and collar. Across the top of the graphic reads DESIGNING TOMORROW’S MET, FRIDA ESCOBEDO.”
The Metropolitan Museum’s tweet about a talk by architect Frida Escobedo with her portrait and a detailed appearance description.

The example above is quite good since it allows us to imagine the woman. Just one remark: look at the way how the image caption is cited in the image description. It’s written in uppercase, which can be interpreted as an abbreviation by assistive technology. As a result, screenreaders can spell those words letter by letter.

Unlike minimalist avatar-style portraits, portrait paintings and candid photography contain more individuality, hence a need for more elaborate descriptions, for instance:

Screenshot of a tweet by The Metropolitan Museum of Art with an artwork photo and the image description: “Frederic Leighton’s painting “Flaming June” hung on a wall with a gold frame and featuring a light-skin-toned woman with long red hair lies in a large chair wearing a bright orange dress, fast asleep, her head collapsed on her arm and her hair draped across the top of the chair. It is a serene and warm moment.”
The Metropolitan Museum’s tweet with Frederic Leighton’s portrait painting and a detailed character appearance description.

Describing such an image, you should not only literally list its elements (woman, orange dress, chair, golden frame) but also convey the atmosphere and emotion (serene and warm moment).

And finally, alt-text for animated GIFs. After everything explored above, I don’t think the key principles will differ much. You should describe the keyframes and main actions. Notion’s team did a great job explaining their animation about the lightbox feature.

Screenshot of Notion’s tweet with an animated GIF-image and the image description that says: “This is an excerpt from a blog post called “The data model behind Notion’s flexibility,” displayed in the Notion app. The user double-clicks on the image of a diagram to open it in a lightbox. They then click the “+” button to zoom in twice, and focus on the bottom right corner of the diagram.”
Notion’s tweet about the image lightboxes feature illustrated by an animated GIF with an image description.

To sum up, Twitter is one of few social platforms that boldly reminds about alt-text and doesn’t hide this obnoxious feature behind “Advanced” or “Other” settings. As I said, disability is a matter of time; we all will have firsthand experience with it at some point in the future.

I’m happy to see people writing image descriptions more often; at the same time, there is a lack of understanding of why to do it and what is supposed to be written in the alt-text.

Recommended reading

Ugh, I hope I didn’t screw things up myself in this article. I’m not going to lie, it was not easy to compose proper descriptions for screenshots of tweets with images that already had alt-text. So here is some additional reading for you on the topic:

👋 Hi there! I enjoy two things — sharing knowledge and drinking good coffee. If you like what you are reading now, feel free to support me via the “Buy Me a Coffee” platform ☕ Thank you! ❤️

--

--

Slava Shestopalov 🇺🇦
Design Bridges

Design leader and somewhat of a travel blogger. Author of “Design Bridges” and “5 a.m. Magazine” · savelife.in.ua/en/donate-en