Ensuring Website Accessibility for the Blind: A Comprehensive Guide

As a website owner, how often do you think about how your website is perceived and used by people who are blind or have other visual disabilities? 

For sighted people, it’s often easy to overlook how the visually impaired navigate the internet, but understanding their experience is fundamental to fostering an inclusive web environment. 

This article goes into the unique challenges that users who are blind or visually impaired face when accessing websites.

By the end of this guide, you’ll not only gain a deeper understanding of the needs of visually impaired users but also feel empowered to take meaningful steps toward making your website more accessible to all.

Understanding website accessibility and its importance

Website accessibility refers to removing barriers that prevent people with all kinds of disabilities, impairments, and limitations from interacting with or accessing a website. When sites are properly designed, all users should have equal access to information. 

Ethically, accessibility is a stride towards inclusivity and equality, granting all individuals, regardless of their abilities, the right to information and participation. Economically, itextends the market reach and taps into the buying power of a wider audience, including the disabled community, which holds considerable untapped market potential. 

However, there is also the legal side. Legislations like the ADA (Americans with Disabilities Act)Section 508 in the U.S., and the European Union’s Web Accessibility Directiveensure that people with disabilities enjoy non-discriminatory access to technology. 

In the business world, having an easy-to-use website is not only required by law – it’s a valuable asset. For example, case studies have shown that inclusive design can lead to a significant increase in consumers willing to engage with a business or an institution

💡The A11y Collective offers the course ‘Web Accessibility, the business case’, which provides insight into how accessible design can positively impact your business strategies.

Introduction to Web Content Accessibility Guidelines (WCAG)

The Web Content Accessibility Guidelines (WCAG) were developed to make online content more user-friendly. These guidelines target the needs of individuals with disabilities while ultimately benefiting all users. WCAG is acknowledged as the international standard for web accessibility. 

WCAG outlines three levels of conformanceA, AA, and AAA. These levels represent varying degrees of accessibility, with A being the minimum level and AAA the most rigorous. For instance, the text contrast ratio at level AA goes beyond level A, ensuring readability against various backgrounds for those with mild visual impairments and in conditions of glaring sunlight for the average user. 

The guidelines are structured around four fundamental principles, often referred to asPOUR:

  • Perceivable: Information and user interface components must be presented in ways that users can perceive.
  • Operable: User interface components and navigation must be operable, ensuring that users can interact with all controls and interactive elements.
  • Understandable: Information and the operation of the user interface must be understandable, enabling users to comprehend the information as well as the operation of the interface.
  • Robust: Content must be robust enough to be interpreted reliably by a wide variety of user agents, including assistive technologies.

Understanding and applying WCAG principles can significantly impact the lives of people with disabilities, improving their ability to access and use web content effectively. 

Challenges faced by visually-impaired internet users

Visual impairment affects people differently, from minor vision problems and colour blindness to a complete lack of vision. This spectrum significantly influences how people interact with websites and digital content. 

For instance, someone with a moderate visual impairment may struggle with small text sizes or low-contrast colour schemes, while someone who is completely blind relies entirely on assistive technologies like screen readers to navigate the web. 

Users who are blind or visually impaired often encounter several common accessibility issues when using the internet. These include:

  • Missing alt text for images: Alternative text provides a textual description of images on websites, which is crucial for screen readers to convey what the image represents. Without alt text, visually impaired users miss out on essential information.
  • Lack of support for keyboard-only navigation: Many visually impaired users rely on keyboard navigation instead of a mouse. Websites that don’t supportkeyboard-only navigation limit the ability of these users to interact with site content effectively.
  • Inconsistent or complex page navigation and structure: A website that lacks a logical structure, including clear headings and easy-to-follow navigation paths, can be challenging to navigate for those using screen readers.
  • Inaccessibility to screen readers: Parts of websites that are inaccessible to screen readers, such as certain types of media or dynamic content, can exclude visually impaired users from accessing important information.
  • Unlabeled inputs, buttons, or links: Inputs, buttons, and links need associated labels or accessible descriptions. Without these, visually impaired users may not understand the function or purpose of these elements, leading to a confusing user experience.

Overcoming these obstacles is important when improving website accessibility for users who are blind or visually impaired. It is a step in acknowledging and valuing the diversity of internet users.

Making your website accessible for the blind: Strategies and techniques

When enhancing website accessibility for blind or visually impaired users, start with a solid understanding of accessible design principles. 

A great resource for this foundational knowledge is The A11y Collective’s course, “Accessible design, the basics.” This course provides a comprehensive overview of best practices in accessible web design. 

Ensure all key elements are screen-reader and keyboard accessible

Users who are blind or visually impaired typically rely on screen readers and keyboard-only navigation to interact with online content. 

Here is what you can do to help them: 

Keyboard accessibility

  • Proper tab order: Ensure that the tab order of the website follows a logical sequence, enabling users to navigate through elements in an intuitive manner.
  • Visible focus indicators: Implement visible focus indicators, like outlines or colour changes, to show which element is currently focused during keyboard navigation. This helps users know their location on the page.
  • Bypass or ‘Skip to Main Content’ links:Include ‘skip to content’ links at the beginning of each page. This allows users to bypass repetitive links and directly access the main content, improving navigation efficiency.

Screen reader-friendly content

  • Proper use of HTML and ARIA (Accessible Rich Internet Applications) tags and attributes that follow the hierarchy of the page. 
  • Meaningful link text that is descriptive and conveys the purpose of the link. 
  • Descriptive alt texts for images that describe their purpose and content, enabling users who cannot see the image to understand their relevance.
  • Transcriptions for audio and video content.

Focusing on these aspects to improve website accessibility for users who are blind or visually impaired results in a positive user experience for all. 

Add alt text for non-text content

For those who use screen readers, alt text serves as a guide, providing descriptions of non-text content such as images and videos. Crafting alt text with care gives users who are blind or visually impaired an understanding and appreciation of the content that others gain through sight.

Here are some tips for writing your alt text: 

  • Be concise and clear: For an image of a dog catching a frisbee, use “A brown dog jumping to catch a green frisbee” as opposed to “an image of a pet in action.”
  • Avoid redundancy: Since screen readers identify images, there’s no need to include phrases like “picture of.” Simply describe the scene or object.
  • For decorative images that serve no informational purpose, use an empty alt attribute, alt=””, to allow screen readers to bypass them without confusion.
  • Do not use alt text to stuff keywords. This practice not only hinders accessibility but can also negatively impact SEO.
  • Avoid vague descriptions that do not concretely convey the content’s context or function.

Organise page content with headings 

Well-structured headings are important for digital accessibility. They help screen reader users understand the organisation of your content, making it easy for them to navigate. 

This involves: 

  • Using semantic HTML to structure your content, such as headings (<h1> through<h6>), lists (<ul>, <ol>), and buttons (<button>). This helps screen readers interpret the page structure and convey it accurately to users.
  • Avoiding repetitive headings that lack distinct context or information which can be disorienting for those using assistive technologies.
  • Not using heading tags for text that isn’t a heading, like pull quotes or emphasised statements, to maintain the integrity of your page’s architecture.

To better understand accessible design principles, including using headings correctly, consider exploring The A11y Collective’s course on the basics of accessible design.

Reconsider your use of colour

When imagining a website, colours and other visuals may seem like a great way to catch people’s attention. However, for users with visual impairments, the situation is a bit different.

For example, oftentimes, websites would use colour to signal important information – a form field lighting up in bright red when not filled correctly or showing an active menu item only by changing the shade. However, to users who are visually impaired or fully blind, these signals are less clear – and in some cases, completely useless – so you need to find a better way to communicate this information. 

Here are some best practices you can follow: 

  • Make sure that all information conveyed by colour is also available in text. 
  • Include a text cue for coloured form control labels.
  • Provide additional visual cues (like underlines, changing the shape of buttons, etc.) beyond colour to convey information – this is especially important for users with colour vision deficiencies or colour blindness.

It’s important to remember that the majority of users with visual impairments are not fully blind and will still be able to perceive visual elements on your site. For this reason, you need to ensure that any colour you use in your website design meets accessibility guidelines for colour contrast.

 To help you with the exact contrast ratio, go through the WCAG 2.1 AA standards for colour contrast. Additionally, tools like the WebAIM Contrast Checker, the Stark plugin for SketchAdobe XD, and Figma, and the Accessible Color Palette Builder can assist in this process. 

An example of good colour contrast on the WebAIM contrast checker.
Alt text = An example of bad colour contrast on the WebAIM contrast checker.

Visually cluttered pages with too many colours can pose a challenge, not just for the visually impaired but also for users who are colourblind. Compare a web page bathed in neon greens and yellows, fighting for attention, to one where a clear hierarchy is established with a balanced colour palette, and you will see how less is often more.

In creating an interface where colour use is strategic and purposeful, it’s vital to introduce other differentiators, accounting for diverse user needs. A call-to-action button, for instance, could leverage a distinctive shape, an icon, or an additional label rather than relying solely on colour to stand out.

Use descriptive labels for links and buttons

Providing contextual descriptions for inputs like links and buttons enables users to grasp the function and relevance of an interactive element before they decide to engage with it.

Generic labels, such as ‘click here’, aren’t accessible because they fail to impart any meaningful information about what will occur post-click. So instead of that, you can put ‘read more about accessible web design’, which immediately clarifies the destination or action.

Labels serve as a critical navigation aid for screen reader users, helping them understand the purpose and outcome of each interactive element. Descriptive labels can contribute positively to a website’s SEO, offering search engines clear and contextual information, improving the site’s relevance and rankings.

Want to learn more about accessible design? 

Enrol in our “Accessible design, the basics” course to learn the fundamentals of how to design your website in a way that is accessible for all users!

Enrol in “Accessible design, the basics”

Make it possible to manually adjust the font size

It’s crucial to acknowledge the varied font size preferences and requirements of users. To enhance your website’s text accessibility for visually impaired users, the website’s code should allow for personal adjustments to font size

While browsers typically provide options to modify text size, and magnifying tools exist, coding your site to support manual changes by users will broaden your site’s accessibility.

Adhering to best practices in HTML and CSS, such as employing relative units like em or reminstead of fixed units like pixels, facilitates user-initiated text resizing. 

Stepping towards a more inclusive web

Designing for accessibility is more about compassion than compliance. Making sure that websites are welcoming to all, including those with disabilities or impairments, is both a moral imperative and a technical challenge. 

We have gone through essential concepts that can help pave the way for an accessible web, such as screen-reader compatibility and optimising keyboard navigation. These are not just checkpoints but powerful steps that break down digital barriers

Ready to develop accessible websites that resonate with everyone? The A11y Collective’s “Accessible code” course offers a solid foundation, equipping you with the skills to create web experiences as universal as the internet itself.

With each lesson, you’ll move closer to not just meeting but exceeding accessibility standards.Start now!

The post Ensuring Website Accessibility for the Blind: A Comprehensive Guide appeared first onThe A11Y Collective.

 

Using the Block Editor with a Screen Reader

Using the Block Editor with a Screen Reader

Posted on March 30, 2024 by abletec

The Gutenberg editor was introduced in 2018 and became the default editor in WordPress since version 5. Its adoption amongst blind persons, though, has been slow and reluctant for a variety of reasons. Initially, the accessibility was abysmal. Secondly, change is often not readily embraced because it’s just easier to keep doing things in ways that are familiar. Thirdly, and most importantly however, is the fact that, to date, there has been a profound lack of training regarding how to use the editor with a screen reader. The problem is, of course, that eventually the classic editor will cease being supported, and blind & low vision creators and developers will find themselves in a very difficult position. Paul Martz and Jackie McBride hope to remedy that as they present a class on using the block editor for those who use screen readers. The class will cover the reasons for the change, the new capabilities the block editor introduces, the appearance of the block editor, and finally, how to use it to create content. Paul will discuss using the editor with Voiceover on the Mack, while Jackie will discuss its use with Windows.

The class will be held on Apr. 26 at 2 PM Pacific and Mountain Standard Time, 3 PM Mountain Daylight Time, 5 PM Eastern Time, and 2100 UTC for the rest of the world. Please consult a timezone website if you’re unsure how that translates to your timezone.

The class will be held via:
Free Conference Call .com ((267) 807-9605) followed by access code 662501#.
Participation via
my free conference call .com room
is also available to those for whom telephone participation is a hardship. The class is free of charge and will be recorded for those who can’t participate or who just wish to review the content again. The presentation will last about an hour, and there will be time for questions at the end.

We hope to see you there.

 

 

 

Microsoft says this single key is the difference between an AI PC and just a PC with AI

What you need to know

  • Intel made several AI announcements at its AI PC Acceleration Program in Taipei, Taiwan this week.
  • During the event, Intel unveiled Microsoft’s definition of an AI PC, which includes some surprising requirements.
  • Microsoft’s definition of an AI PC requires a system to have a CPU, GPU, NPU, to support Copilot, and to have a Copilot key.
  • Notably, some PCs that have already shipped have a CPU, GPU, and NPU but do not have a Copilot key, meaning they do not meet Microsoft’s definition of an AI PC.

Intel shared Microsoft’s definition of an AI PC at its AI PC Acceleration Program in Taipei, Taiwan this week. The chipmaker is one of several industry leaders to push AI hardware. Intel’s Core Ultra processors feature neural processing units (NPUs) specifically optimized for AI. Those chips compete with CPUs from AMD and Qualcomm. But regardless of which company makes the chip in your PC, your device can be an AI PC in the eyes of Microsoft. Tom’s Hardware reported on the event.

Microsoft’s definition of an AI PC was made jointly with Intel, though it opens the doors for AMD, Intel, or Qualcomm devices to carry the designation. According to Microsoft, a computer must have a CPU, GPU, and NPU to be an AI PC. Additionally, supporting Microsoft Copilot and having a Copilot key are requirements. That last item has caused a bit of a stir, since it means that to be labeled an AI PC a system has to have a specific key on a keyboard.

Early this year, Microsoft announced that future PCs will feature a Copilot key. The company called it the “first significant change to the Windows PC keyboard in nearly three decades.” The new Copilot key will be mandatory for OEMs at some point in the near future, but some 2024 PCs lack the key. By strict definition, a computer with all the hardware needed and support for Copilot is not an AI PC unless it has a Copilot key, at least in the eyes of Microsoft.

What is an AI PC?

Microsoft's requirements for an AI PC
Microsoft requires a PC to have a Copilot key to be called an AI PC.(Image credit: Intel)

Apple infamously asked “what’s a computer” in an ad several years ago. In 2024, the question on people’s minds is “what’s an AI PC?”

The saga surrounding the definition of an AI PC reminds me of the famous xkcd comic about competing standards. In that comic there are 14 competing standards, which is considered ridiculous by one of the characters. The obvious solution is to create one universal standard. As you likely guessed, this results in there being 15 competing standards.

With Qualcomm, AMD, and Intel making processors built for AI and Microsoft and other companies working on software, there isn’t a clear definition of an AI PC. Even Intel has its own definition of an AI PC despite announcing a jointly created definition with Microsoft. Intel’s definition is largely the same as the one announced yesterday, but Intel does not require a system to have a Copilot button or Copilot support to be an AI PC.

So, what is an AI PC? It depends on who you ask. Microsoft wants any PC marketed as an AI PC to work with Microsoft software and have a Copilot button. By that definition the $1,600 ASUS ROG Zephyrus G14 with a CPU, GPU, and NPU is not an AI PC because it doesn’t have a Copilot key on its keyboard. Maybe ASUS can send users a key to swap onto the PC to earn the AI PC designation.

All the latest news, reviews, and guides for Windows and Xbox diehards.

Microsoft says this single key is the difference between an AI PC and just a PC with AI

What you need to know

  • Intel made several AI announcements at its AI PC Acceleration Program in Taipei, Taiwan this week.
  • During the event, Intel unveiled Microsoft’s definition of an AI PC, which includes some surprising requirements.
  • Microsoft’s definition of an AI PC requires a system to have a CPU, GPU, NPU, to support Copilot, and to have a Copilot key.
  • Notably, some PCs that have already shipped have a CPU, GPU, and NPU but do not have a Copilot key, meaning they do not meet Microsoft’s definition of an AI PC.

Intel shared Microsoft’s definition of an AI PC at its AI PC Acceleration Program in Taipei, Taiwan this week. The chipmaker is one of several industry leaders to push AI hardware. Intel’s Core Ultra processors feature neural processing units (NPUs) specifically optimized for AI. Those chips compete with CPUs from AMD and Qualcomm. But regardless of which company makes the chip in your PC, your device can be an AI PC in the eyes of Microsoft. Tom’s Hardware reported on the event.

Microsoft’s definition of an AI PC was made jointly with Intel, though it opens the doors for AMD, Intel, or Qualcomm devices to carry the designation. According to Microsoft, a computer must have a CPU, GPU, and NPU to be an AI PC. Additionally, supporting Microsoft Copilot and having a Copilot key are requirements. That last item has caused a bit of a stir, since it means that to be labeled an AI PC a system has to have a specific key on a keyboard.

Early this year, Microsoft announced that future PCs will feature a Copilot key. The company called it the “first significant change to the Windows PC keyboard in nearly three decades.” The new Copilot key will be mandatory for OEMs at some point in the near future, but some 2024 PCs lack the key. By strict definition, a computer with all the hardware needed and support for Copilot is not an AI PC unless it has a Copilot key, at least in the eyes of Microsoft.

What is an AI PC?

Microsoft's requirements for an AI PC
Microsoft requires a PC to have a Copilot key to be called an AI PC.(Image credit: Intel)

Apple infamously asked “what’s a computer” in an ad several years ago. In 2024, the question on people’s minds is “what’s an AI PC?”

The saga surrounding the definition of an AI PC reminds me of the famous xkcd comic about competing standards. In that comic there are 14 competing standards, which is considered ridiculous by one of the characters. The obvious solution is to create one universal standard. As you likely guessed, this results in there being 15 competing standards.

With Qualcomm, AMD, and Intel making processors built for AI and Microsoft and other companies working on software, there isn’t a clear definition of an AI PC. Even Intel has its own definition of an AI PC despite announcing a jointly created definition with Microsoft. Intel’s definition is largely the same as the one announced yesterday, but Intel does not require a system to have a Copilot button or Copilot support to be an AI PC.

So, what is an AI PC? It depends on who you ask. Microsoft wants any PC marketed as an AI PC to work with Microsoft software and have a Copilot button. By that definition the $1,600 ASUS ROG Zephyrus G14 with a CPU, GPU, and NPU is not an AI PC because it doesn’t have a Copilot key on its keyboard. Maybe ASUS can send users a key to swap onto the PC to earn the AI PC designation.

All the latest news, reviews, and guides for Windows and Xbox diehards.

Back Tap on iPhone: the forgotten accessibility action

Hand holding an iPhone with caption ” Back Tap on iPhone: the forgotten accessibility action

by Joe Lonergan

A lot was made of Back Tap when it came out a few years ago, but so much has happened since I think it has been passed over by some other features.

Back Tap is an Accessibility action introduced to the iPhone that allows you to double tap or treble tap the back of your iPhone to perform a shortcut.

  1. On your iPhone, go to Settings > Accessibility > Touch.
  2. Scroll down, then select Back Tap.
  3. Select either Double Tap or Triple Tap.
  4. Scroll down to the Shortcuts area, then choose a shortcut.
  5. Tap Back Tap to apply the setting.
  6. Double-tap or triple-tap the back of your phone to run the shortcut.

So why would we use it?

It is situated in the touch area of accessibility settings so we can presume it is there to help people that have trouble interacting with the touch screen.

It can be used for many shortcuts and for launching many applications, read on for some use cases.

If you have an iPhone with no home button you can set Back Tap to launch the app switcher or unlock the phone, these actions can be difficult for some users, and tapping the back of your phone to launch can make it so much easier.

You can use it to launch the magnifier app, there are many occasions throughout the day that you may need to use the magnifier app, and having an action to launch it quickly is good.

You can use Back Tap to invoke a Siri Shortcut. If you have set up some Siri shortcuts in the shortcuts app you will now see them in the list of actions available in the Back Tap section. Siri Shortcut could play your favourite radio station, tell you your current location turn on the lights, and more.

Note, You can add two actions to Back Tap, one to double tap and one to treble tap.

So, if Back Tap is something you have not checked out in a while, revisit it, and see if it could be useful for you. Let us know if you have any great ideas for the feature.

 

BlindShell Hearing Aid Compatibility Update

Dear BlindShell Community,

 

We hope this email finds you well. In response to continuous demand for hearing aid compatibility, we’re excited to share a significant development with you. 

 

Based on the ongoing requests from our valued community members, we’ve taken the initiative to collaborate with a third-party organization (Towson University Hearing &Balance Center) to conduct extensive hearing aid testing. As a result, we’re thrilled to announce that the BlindShell Classic 2 is compatible with leading hearing aid brands such as ReSound, Oticon, and Phonak!

 

Whether you’re a current user of the BlindShell Classic 2 or considering joining our community, this compatibility upgrade opens up new possibilities for enhancing your daily experiences. Now, you can seamlessly integrate your hearing aids with our device to enjoy improved audio clarity and communication.

 

To explore the full list of compatible hearing aids and learn more about integrating them with the BlindShell Classic 2, we invite you to visit our new dedicated landing page by clicking here. There, you’ll find comprehensive information and resources to help you make the most of this exciting compatibility upgrade.

 

Thank you for being a valued member of the BlindShell community. Together, we’re breaking down barriers and creating a more accessible world for everyone. 

 

Best regards,

 

BlindShell USA Team

 

 

 

 

Image removed by sender. icon

 

Free shipping 

all orders over $75

Image removed by sender. icon

 

30 days returns 

moneyback guarantee

Image removed by sender. icon

 

Monthly payments

Buy now, pay later

Image removed by sender. icon

 

USA Tech Support

Supported Hearing AidsMon-Fri 9-5Eastern

Trying Apple’s Personal Voice

Apple recently introduced Personal Voice to newer devices on various hardware in their lineup. I have had a little experience with the basic concept behind this sort of technology from my time at Microsoft where I dabbled with one of Microsoft’s Azure cognitive services to do something similar.

The basic concept behind these experiences is that you record some set of known text and then software converts that into a synthetic version of your voice. In Apple’s case it is 150 phrases ranging from just a few words to maybe at most 20 words in a single phrase.

After you finish recording, there is some processing time and then your voice is ready to use. On an iPhone 15 Pro, my voice was ready in about five hours. You are not able to do anything else with the phone while this is happening. On an M1 MacBook Air from 2020, processing took about two hours and I was able to do other tasks at the same time, such as writing this blog post.

Once your voice is created, you can use it as one of the voices available with Apple’s Live Speechfeature. This allows you to type in various apps where you would typically use your voice and have the synthetic voice used. It compliments the existing voices Apple makes available and has the added benefit of allowing you to have some relationship to your own voice used in the experience. In situations where a person may know that they are going to lose their voice ahead of time, it does offer some ability to preserve your own speech.

Multiple factors influence the quality of the end result here—Microphone, recording environment and more, just to name a few. For short phrases it likely is not noticeable but in my sample, even the pace at which I appear to have read the samples was different. There is a 21 second difference in having the voice read back the same text.

I made two voices in trying this experience. The first was recorded using the default Apple headphones on an iPhone 15 Pro. The second using an Arctis 7 headset. Both samples are my Apple Personal Voice reading my blog post onAccessibility Island.

I have also made a sample of my original voice sample of three phrases and then Apple’s Personal Voice speaking those phrases from my recording with the Arctis 7 device. The Personal Voice speaking the phrases is the result of my typing them into an edit box and asking for them to be spoken using my newly created voice. The phrases are in this recording and have the original voice sample followed immediately by the Personal Voice speaking the phrase. After all three phrases are played, the entire series is duplicated once. The phrases are:

can you call me in an hour

Did you remember to take out the trash?

Is she going to the grocery store now or in the morning?

Creating a personal voice is straight forward. On whatever device you are using, go to Settings:Accessibility:speech:Personal Voice. You’ll be prompted to record a short phrase to test your recording environment and advised of any changes you should make, such as too much background noise. You then start the process of recording 150 phrases. They do not all need to be recorded at once. When you are finished, you’ll be advised to lock your phone if doing this on an iPhone or just ensure your computer is charged if using a Mac.

When the voice is created, you can start using it with Live Speech by going to the same Speech area of Accessibility settings and going into Live speech. Turn Live Speech on and then pick from the list of voices. Your personal Voice should be listed.

If you are doing all of this with VoiceOver, Apple’s screen reader, as I did, the process of creating a voice works well with VoiceOver. You can use VoiceOver to read the phrase to be read, then activate a record button and repeat the phrase. Recording stops when you stop speaking. If you turn on a setting for continuous recording, you will advance to the next phrase automatically and can repeat the process. I did notice that sometimes VoiceOver automatically read the next phrase but not always. Focus seems to go to the Record button and I suspect there is a timing issue between the phrase being spoken and VoiceOver announcing the newly focused button.

Having created two voices, I would say it is probably a good idea to take a short break during the reading of the 150 phrases from time to time. I found myself not speaking as clearly as I wanted once in a while as well as having sort of the same singsong phrasing. Listening to my voice samples and how the voice came out, I would also say the microphone used has a big impact on the voice quality. This isn’t surprising but is made apparent to me comparing the samples of what my recordings sounded like and how that turns out when the same text is spoken by Personal Voice. I don’t think either microphone that I used would be what I would recommend for creating a voice to be used permanently.

I was curious if Apple would allow the personal voice you create to be used with VoiceOver. I didn’t expect it would be possible and that does seem to be the case.

As with pretty much anything in AI, synthetic speech is a rapidly changing technology. There are certainly higher quality voices in the arena of synthesized speech but Apple has done a good job at allowing you to tap your own voice on consumer hardware in an easy to use process. Listening to my own voice, it is clear it isn’t me and I wasn’t expecting it to be. But even on the basic hardware I used, there are characteristics of my voice present and if I were in a situation where I was going to lose my physical voice permanently, this is one option I would definitely explore further.

Generative AI: The World in the Palm of Your Hand

Promotional image for “On TECH & VISION with Dr. Cal Roberts” featuring a purple and blue gradient background, a Wi-Fi symbol, a wireframe globe in a hand, and a portrait of Dr. Cal Roberts in a suit.

When it comes to emerging technology, there’s no hotter topic than artificial intelligence. Programs like ChatGPT and Midjourney are becoming more popular and are inspiring people to explore the possibilities of what AI can achieve – including when it comes to accessible technology for people who are blind or visually impaired.

One of those people is Saqib Shaikh, an engineering manager at Microsoft. Saqib leads a team that developed an app called Seeing AI, which utilizes the latest generation of artificial intelligence, known as generative AI. Dr. Cal, spoke with Saqib about how Generative AI works, his firsthand experience using an app like Seeing AI, and how it helped improve his daily life. 

This episode also features Alice Massa, an occupational therapist at Lighthouse Guild. Alice described the many benefits of generative AI, and how it helps her clients better engage in their world.. 

Saqib and Alice also both agreed that the current state of AI is only the beginning of its potential. They shared their visions of what it could achieve in the future – and it doesn’t seem that far off.

Podcast Transcription

Female AI Voice: The image captures a close up view of a small rocky pool of water inhabited by various starfish. The pool is surrounded by rocks and the water is clear, allowing a detailed view of the starfish and the rocky bottom of the pool the starfish are in a variety of colors and sizes, including some with a reddish orange hue and others that are beige with brown patterns.

The starfish are scattered throughout the pool, some partially submerged in water, while others are fully visible. Their arms are spread out, showcasing their iconic star shape. The rocky bottom of the pool is visible through the clear water.

The rocks are small and appear to be smooth, with colors ranging from dark brown to light beige. There are also larger, irregularly shaped rocks that are reddish brown. There is a yellow sign on a rock on the right side of the image with the text: “But please don’t touch.”

Roberts: What you just heard is the audio description of a starfish tank at an aquarium. Judging from the details, you think it’s a prerecorded clip for an audio tour, but it’s not. This clip was generated on the spot by artificial intelligence or AI.

I’m Doctor Cal Roberts and this is On Tech and Vision. Today’s big idea is technology that can literally put the world in the palm of your hand. I’m talking about the latest generation of artificial intelligence called Generative AI.  It is revolutionizing how people who are blind or visually impaired interact with their environment, not just simple object identification and directions, but enabling users to experience their surroundings fully and engage on a much deeper level.

To learn more about this new evolution, I spoke with Saqib Shaikh, an engineering manager at Microsoft. Saqib has developed an app called Seeing AI and has firsthand experience using AI in his daily life.

It’s a pleasure to welcome to On Tech and Vision Saqib Sheikh. Saqib was the Pisart Award winner at Lighthouse Guild last October in October of 2023 and just wowed us with his vision for the future of AI and how AI will impact the lives of people who are blind and visually impaired. And so it’s just a great honor to welcome Saqib to On Tech and Vision. Welcome.

Shaikh: Thank you so much. It’s an absolute honor to be here, thank you very much.

Roberts: So, our audience is getting familiar with this term, AI, artificial intelligence. People hear it all the time. Now we have a new term for them, Generative AI. Explain. What’s Generative AI?

Shaikh: In many ways, it’s just the latest wave of artificial intelligence. So, artificial intelligence has been around for many decades, and what I started with Seeing AI at Microsoft maybe seven years there was a whole new wave back then of deep learning. It really improved what was possible. And we’re now at this new Generative AI wave. So, in a sense, the term doesn’t matter. You might prefer the term large language model in the press, and again this is a tech word that just means we have really big AI systems which are trained on really large computers and a huge amount of data in a way that was never thought possible before. And the results, which is really what matters is the results, are just remarkable. And this is the technology behind tools like ChatGPT and of course now Seeing AI as well.

Roberts: So compared to past generations of AI, what does the user experience?

Shaikh: We are able to get a lot more details because the AI is able to, like generative sounds, it is able to generate the descriptions or generate language in general from having observed huge amounts of data on the Internet. And so it is able to give the impression of, in the sense of like a chat bot responding like a human, and in the case of describing images, just way in more details than was ever possible before.

Roberts: So this is information as you say on the Internet. How about peoples personal experience? How does Generative AI help someone with the repetitive tasks that they have to function on a daily basis?

Shaikh: We’re still at the beginning of this, but some of the ways are, Microsoft has these tools are called copilots because they sit alongside the person, the human, and as like a copilot they can help you with everyday tasks. And again, it’s just the beginning. But I’m excited to see that they can summarize information or generate reports of interest to someone who’s blind. Maybe there’s a huge amount of data that can be analyzed, summarized, or trend spotted.

And in the case of like Seeing AI, we were talking about the real world with images. It can go from giving a one sentence overview about what’s in an image to giving you a whole paragraph describing intricate details of what’s in the image, again in a way that they never thought even a year or two ago would be possible.

Roberts: So take us through your own personal journey of how you got interested in AI.

Shaikh: So, I’ve been interested in technology since I was a kid. I sort of stumbled onto learning to code when I was in my early teens and I just loved it. I loved this idea that you could have an idea, do some thinking, some typing, and you could make something out of nothing.

And from that, it’s the transformative ability of technology to improve people’s lives. Growing up as a blind kid, I was surrounded by software to generate Braille or to print out the Braille and generate raised diagrams, and so much more. So assistive technology was what enabled me at school, and of course, beyond that as well. 

And so that took me to studying computer science at university and then I specialized in artificial intelligence in my postgraduate and eventually came to Microsoft where I’ve done a whole bunch of things, but it was about seven years ago where we had the opportunity to do a hackathon. The CEO said that there’s a one week period where you can do whatever catches your interest and for me I thought, I want to bring my interest in artificial intelligence and some of the needs I identified individually as a blind person, that’s where it started. 

And we spent a week prototyping some solutions for AI to help blind people and building on this vision of what if there was a visual assistant with you, like when you have a sighted guide who understood you and understood the world around you and told you what was going on, what was interesting to you and could answer your questions. And we’re still quite away from that vision, but we’re closer than we ever have been. 

And so that brought me to the Seeing AI journey which I’ve been on for, like I say, about seven years or so now, which feels like a remarkably long time, but there’s so much exciting things going on, it doesn’t feel that long at all.

Roberts: So for those who don’t know the Seeing AI app, explain it. What does it do and how does someone use it?

Shaikh: Yes, it’s a free mobile app which you can download from the App Store or Play Store and it’s a visual assistant. We sometimes talk about it as the talking camera you just hold up your phone, point it and it will start reading things to you. You could also identify your friends, or describe what’s in our picture or recognize products. And there’s a whole bunch of different tasks it can assist with in one’s daily life.

Most recently, some of the interesting things powered by Generative AI is, as I said, going from taking a photo, say, from your photo gallery to your reliving memories from your vacation, or even just what’s in front of you right now. You can go from saying it’s a man sitting on a chair in a room to actually give you maybe a whole paragraph describing what’s in the room, what’s on the shelf, what’s in the background, what’s through the window, even. And it’s just remarkable. 

I work on this every day. I understand technology, yet as an end user, I still am surprised and delighted by what this generation of AI is capable of telling me.

Roberts: Alice Massa is an occupational therapist. She provides therapy to people who are visually impaired, including how they incorporate vision technology into their lives. And she’s a user of vision technology herself.

Massa: My name is Alice Deluise-Massa. I’m an occupational therapist. I’ve been in practice 50 years. Most people are not always sure what occupational therapy is. My view of occupational therapy is that it’s an empowering profession that looks at people’s strengths, looks at the barriers that are preventing them from doing what they want to do, and offering strategies in order to meet their goals and to empower them to be able to do the things they choose to do.

So I am now working at the Lighthouse and I see clients for occupational therapy and what we do is we do what we call an occupational profile and what that is, is what people’s lives are like, what their strengths are, what their interests are, what they are struggling with doing, what they would like to do, and then we try to match them with strategies, techniques, compensation so that they can do the things they want to do, and that’s where the electronics came into play.

The reason I love my iPhone, and specifically the Seeing AI app, is it’s been developed with really visually impaired people in mind. Seeing AI is so intuitive for somebody who has vision loss. There are a lot of things that you do in a day that other people take for granted. You know, walking in the street, knowing when to cross the street, knowing which street you’re on, going into a store and buying regular milk instead of low fat milk. Those are all the kind of day-to-day things that people really take for granted. 

I was not able to shop alone before I had Seeing AI. Now, I go into TJ Maxx, I pick up a blouse. I can see what size it is. I can see what the price it is. I can see what it’s made of just by putting the tag in front of my phone. It’ll also read the menu at a restaurant.

Roberts: In fact, Seeing AI can do more than just read the menu at a restaurant. It can help you decide what to order. Here’s a recording from Saqib showing how it’s done.

Man: After scanning this restaurant menu in Seeing AI. What are the categories?

AI Voice: The categories in the document are antipasti, salads, pizza with sub categories, red and white specials and dessert.

Man: What Antipasto dishes are there?

AI Voice:  There are three antipasti dishes mentioned in the document. One: oven roasted asparagus with paprika aioli, red onion, tomatoes, arugula and lemon zest. Two: polenta with marinara sauce, granorino, homemade meatballs and herbs. Three: oven roasted castelvetrano olives marinated with chilies, garlic and lemon.

Man: How much does the asparagus antipasti cost?

AI Voice:  The asparagus antipasti costs $12.00.

Roberts: Now we’ll hear Alice tell us more about the incredible things Generative AI can help us do.

Massa: I have a client at the Lighthouse who really was very disinterested in doing anything. The only thing he did on his phone was answer a call from his pastor and call his pastor, and I was able to put Seeing AI on his phone and his wife said the first time in two years she saw a smile on his face because now he could read his Bible by himself.

He just literally had to put the phone in front of the Bible and it would read the page. So for him this was life changing. I call my phone my sister because my phone is the person I go to when I’m on the street. If I’m walking in Manhattan- the other day I was walking, I was meeting someone on 47th St. I wasn’t sure which block I was on. all I did was open Seeing AI short tech, hold it up to the street sign and it told me I was on West 46th St.

So those are the kinds of things that it really makes a difference.

Roberts: As we pointed out in many of our podcast episodes involving the users of technology in its development is vital. Saqib agrees.

Shaikh: I view my work in some ways as a conversation between the community and the scientists. And one of the most fun things about that, like, you know, I just really enjoy hearing from people. What are they using Seeing AI for? Of course, what could it do in the future? But often we’re surprised even by the way these people are using it today because everyone has different tasks and different requirements in their lives. And we are making a set of tools to help as many people as possible. 

A story that comes to mind is Seeing AI can now recognize currency bills, but it did not at the very beginning and we got an e-mail from a group of users who had decided they were going to use the face recognition feature to train Seeing AI to recognize all the presidents and faces on the bank notes like Lincoln or whoever, and you know you’re using that to recognize currency. And I was like, that is so innovative that, you know, our users are taking these tools and they themselves are inventing by finding the ways that it’s going to enable them in their daily lives.

Roberts: And an incredible innovation like this is only scratching the surface of  Generative AI’s capabilities. 

The term multimodal capabilities comes up a lot in discussions about Generative AI. Can you speak more on that?

Shaikh: Yes. So a lot of generative AI began with the concept that you’d have something similar to a chat bot like ChatGPT, where you’re typing text and you’re reading text. Multimodal means what if there were multiple modes of input like audio or a picture, or potentially even video. So of interest to someone who can’t see this idea that you can give Generative AI systems an image and maybe a question about the image or some instructions about what you want to know about the image and you’ll be able to get textual or audio output describing the image, or you’re answering your question that that’s the crux of the multimodal part.

Roberts: And so multimodal can be not only the sources of input, but then it could also be the means of output, so it’s not just everything has to be audio. Our listeners to On Tech and Vision have learned over the years a lot about haptics and about being able to feel data not just to always hear data.

Shaikh: Yeah, absolutely. And today, I don’t know of much work with sort of these multimodal models doing haptics, but why not? In a sense, we’ve seen systems doing music generation and other audio and you know haptics is not that different to audio.

And also generating images. I’m quite excited by this idea that what if someone who could not see can actually generate good looking images. So I think that we are at the cusp of this new feature where in everyday life but also some productivity at work and school we haven’t yet discovered all the possibilities of how this can transform the type of trust that someone who cannot see could independently carry out.

Roberts: So we care a lot about assistive technology for people who are blind. But we also recognize that there are other forms of disability that could benefit from AI, particularly people who can’t hear.

Shaikh: Absolutely. And in recent years, we’ve seen the technology for that audience come a long, long way as well. Speech recognition used to be very hit or miss. But now speech recognition, dictation or identifying who is speaking, these aspects of AI have really become quite accurate. So they can be usable in everyday life and work very fast so that you can even use them in real time. So you’re absolutely right that.

It’s not just one form of disability, and it doesn’t have to be limited to disabilities, though I have this philosophy that people with disabilities are often at the forefront of new technologies because we have the most to gain our the most invested in experimenting with this.

So, if you look at historically, there are so many innovations which were initially created for people with disabilities, whether that be the telephone or iPhone touchscreen, on screen  keyboards, text to speech and speech recognition, and flatbed scanners and talking books, and the list just goes on.

It’s this idea that people with disabilities have the most to gain, and so technologies which later become commonplace are often explored and created in the research phases for people with disabilities. And so if you take that one step further and saying, OK, if we look to the future, what are the needs of people with disabilities that are not being met, that this new wave of AI can label, that’s incredibly exciting. And I have this idea of what I call an assistive agent or an assistive companion. 

So what if  AI could understand you as human. What are your capabilities? What are your limitations at any moment in time, whether that’s due to a disability or your preferences or something else, and understand the environment, the world you’re in or the task you’re doing on the computer or whatever. And then can we use the AI to close that gap and enable everyone to do more and realize their full potential.

So right now, that’s my big vision. And yes, we’re probably still a ways away, but we’re getting closer all the time. So, it’s just this idea that we are all different, every single one of us has our own needs, whether it’s we call it disability or not.

Roberts: Alice Massa also has a vision for the future of AI.

Massa: What I would love is if it could describe an action scene. Let’s say I was at a baseball game and the guy hit the ball and the shortstop is pitching the ball at the third base. I believe that eventually it will be able to describe those kinds of seats. I think that would be ideal and the reason I say that is I have worked with some young people who really do want to socialize by going to sporting events and things like that. So, I always think of Phil Rizzuto, like listening to a ball game as it’s happening right in front of them, almost like a radio announcer.

When I was at the theater the other night, it would have been nice if I could have had some description of the theater. The movies have the auditory description, but that’s done by someone as they watch them because the movie never changes. But in real life things are not scripted, so therefore AI would be able to react to what’s actually happening rather than just be a scripted description of the action.

One of the places in reality, other than in theater entertainment is walking in the street. Last night we were walking on Times Square at night. Talk about having vision problems and feeling a little bit challenged, it would be wonderful if there was like a button I could wear that, as I’m walking in the street it’s sort of saying to me there’s a crowd of five people in front of you. Four people are walking towards you directly in your path. A car is coming on your left. You know, if there was some way that you could use artificial intelligence to cue you so that when you’re moving in a communal area that you would be able to navigate much more easily.

Roberts: Saqib has the same hope. I asked him to tell me more about how Generative AI could someday function as a navigational tool, and much more.

Shaikh: Personally, I imagine having this visual assistant, sometimes think of a friend sitting on my shoulder whispering in my ear. When I’m with friends and family, maybe I wouldn’t need such a thing. We just are together enjoying our company in the moment. But there are always those times when there aren’t other people around, so I really want that thing which does the equivalent to what my friends and family do where they’re telling me that, oh, there’s someone walking towards you who you know. Oh, there’s a new shop just opened up on your left. Or be careful, there’s a cleaning trolley down the corridor.

These are things which are just what a friend or colleague would do without even thinking about it, and I really want the AI to be able to take on that role when there aren’t other people around.

Roberts: So recently, legislators have wanted to talk about the safety of AI, and could AI be dangerous? And could the computer take over for people and they’ve created a large level of fear. Talk to me about this subject. What is the risks associated with AI? Do we need to be establishing curbs on the advancement of technology?

Shaikh: I’m absolutely not an expert in this particular area, but many of my colleagues at Microsoft are working on our responsible AI standards and I’m really glad that policymakers are thinking about this, because it is really important for society as a whole that we do start thinking about what are the potential harms and put the things in place so that, you know, we detect and prevent any problems before they happen. I also think as someone on a personal level, with disability that we should also just make sure that we can also continue to leverage these technologies in the ways that can benefit people. So I’m very happy that people are looking into this and it’s it’s not my area of expertise, but responsible AI and making sure that we innovate in a way that does not cause those unintended consequences is critical.

Roberts: Alice agrees. She thinks there is much more to gain from AI than to fear from it.

Massa: There is a lot of anxiety about AI. Everybody right now is thinking, Oh my God, AI is going to take over the world. But the truth of the matter is, AI really has a lot to offer for people who need new strategies, who need other strategies and that the people who need those strategies are often fearful and I just hope people can get over that fear and recognize the value.

Roberts: As we plunge into this AI driven era, the sheer magnitude of what generative AI can achieve becomes crystal clear. Is not merely interpreting texts or offering navigational support. It’s an ever evolving force that adapts and grows with an expanding data universe. Picture a future where AI not only keeps pace with our needs, but anticipates them constantly pushing the boundaries of innovation. Prepare to be blown away by the endless possibilities.

Did this episode spark ideas for you? Let us know at podcasts@lighthouseguild.org and if you liked this episode, please subscribe, rate and review us on Apple Podcasts or wherever you get your podcasts.

I’m Doctor Cal Roberts. On Tech and Vision is produced by Lighthouse Guild. For more information visit www.lighthouseguild.org on tech and vision with Doctor Cal Roberts produced at Lighthouse Guild by my colleagues Jane Schmidt and Anne Marie O’hearn. My thanks to Podfly for their production support. 

Blind Abilities – Wellness Wednesday: Taking a Look at Self-Care, What Does Self-Care Feel Like to You?

In this 28th    episode of Wellness Wednesday, Beth Gustin, LPC, Robin Ennis, LCSW, CPC, dive into self-care. Taking care of yourself and recognizing some habits that maybe you may not realize is self-care. Join Beth and Robin and get a look at how they define self-care. 

Check out all the other wednesday wellness episodes at: https://blindabilities.com/?cat=532&nbsp;

Increasing the Effectiveness of Envision Smart Glasses for the Blind

In today’s world, technology continues to revolutionize the way we live, work, and interact with our surroundings. One groundbreaking innovation that has made a significant impact is Envision smart glasses for the blind. These glasses have the power to change the lives of visually impaired individuals by providing them with increased independence and accessibility. However, like any technology, there is always room for improvement. In this article, we will explore various methods to enhance the effectiveness of Envision smart glasses for the blind.

One crucial aspect to consider is the accuracy and speed of object recognition. Envision smart glasses utilize artificial intelligence and machine learning algorithms to identify and describe objects in the wearer’s environment. By continuously training and updating these algorithms, the glasses can become more adept at recognizing a wider range of objects with greater precision. This improvement would allow users to navigate their surroundings more confidently and efficiently.

Another area of focus is the voice feedback system. Currently, Envision smart glasses provide audio cues to describe objects, people, and text, but there is room for expansion in this area. By integrating natural language processing technology, the glasses could offer more detailed descriptions, including color, size, and texture. Moreover, it would be beneficial to have the ability to answer questions or engage in dialogue, allowing blind individuals to communicate more effectively with their surroundings.

Furthermore, enhancing the accessibility features of Envision smart glasses can significantly increase their effectiveness. Many visually impaired individuals rely on Braille as their primary means of reading and writing. By integrating a Braille display into the glasses, users could read messages or menus directly from the smart glasses. This addition would eliminate the need for carrying separate Braille devices and further streamline the user experience.

Additionally, incorporating a navigation system into the Envision smart glasses would be a game-changer for blind individuals. By utilizing GPS technology, these glasses could help users navigate unfamiliar areas confidently. The glasses could provide turn-by-turn directions, alert users to potential obstacles or hazards, and even identify nearby points of interest. This feature would greatly enhance mobility and improve the overall independence of visually impaired individuals.

Lastly, collaborating with the blind community and obtaining continuous feedback is vital in improving the effectiveness of Envision smart glasses. By actively seeking input and suggestions from users, developers can gain invaluable insights into the challenges faced by visually impaired individuals in their everyday lives. This feedback can drive further innovations and improvements, ensuring that the glasses meet the specific needs and preferences of their target audience.