Skip to content →

The Idea Place Posts

A Small Refinement to iOS 18 Audio Ducking

A few months ago, I wrote about audio ducking in iOS 18. Since the initial release came out, there has been a small refinement to what I originally described.

The ducking choices themselves remain the same: off, when speaking and always. The ducking amount, however, has been moved to a separate setting by the name of Ducking Amount that you can add to the VoiceOver rotor.

To find this setting, open Settings, Accessibility, VoiceOver, Rotor Items and add Ducking Amount.

The Ducking Amount setting allows you to adjust the amount from zero to 100 percent. This is far less confusing than having this controlled by setting volume percentages of greater than 100, as it used to work.

One Comment

A Nice Enhancement for Voice Access on Windows

As a matter of choice, not necessity, I try from time to time to use the various speech and voice input systems in operating systems. My ideal scenario is still to be able to use the computer by voice entirely as well as running a screen meter. I’ve not found a reliable solution as of yet that meets my needs completely.

I know there are combinations of solutions that have made great strides in this area largely using Dragon products and screen readers but as the basis of what I use, I try to use either Voice Access on windows or Voice Control on the Mac. Both platforms also have solutions, as I expect many know, for strictly text input.

I no longer recall how long ago this was but the Voice Access product on Windows did make one change that helps with using screen readers. As a start, Voice Access produces notifications of what Voice Access has heard so that screen readers can echo this back. It is fairly basic and in need of much refinement  it’s at least a start.

I am mentioning this here because in trying voice access this week, I noticed a change that is another step in helping Improve the experience. I do not know when this change was made to be clear. It is just that I noticed it this week. I also run Insider builds of Windows so if this does not work for you, that may be why.

When you’re trying to control the computer by voice, it is common to issue commands such as click and then an item that you want to activate. The challenge becomes that if there is more than 1 item with the same name, you are usually presented some experience to disambiguate what you want to click on.

When I first tried voice access, to the best of my recollection, the experience of identifying what you wanted to activate was not usable with a screen reader. It has been enhanced a bit so that now when that list of choices comes up, the list of choices is echoed back similar to how what Voice Access heard is repeated. Again this needs extensive refinement because it is kind of like a one time listen or read and Braille experience with no way to have the list repeated, step through the list in item at a time or otherwise understand what was said.

As an example of using the feature to identify what I want to click, here was what was read when I asked for the word paste to be clicked.

click paste. Which one?
There are 2 options available. (1) Paste, (2) Paste

Here is another example when I said “click login” on the Fidelity home page.

Click login. Which one?
There are 2 options available. (1) LOG IN, (2) Open login link

It is also worth noting that these disambiguation choices if using Braille appear as flash messages. For those unfamiliar with how Braille displays and screen readers work, this means that the messages stick around for a set period of time and then disappear from the display.

. Here is one last example when I tried to activate the OK button with my voice after running a spell check on an email message. Note, I intentionally replaced the actual email address with email@provider.com.

Click ok. Which one?
There are 2 options available. (1) OK, (2) Sent – email@provider.com – Outlook – 2 running windows

The experiences I’ve described work independent of what screen reader is being used.

Again this experience overall of using the computer with a screen reader and voice on Windows as far from finished. In fact one of the key experiences for correcting words that have not been recognized correctly does not work at all with screen readers. Voice access in fact gives the following notification when you try and correct something and a screen reader is running:

Alert: This experience is not optimized for use with screen readers. Say “Cancel” to exit.

Microsoft has a document on using Voice Access in general. If they have screen reader-specific documentation, I wasn’t able to find it.

If you do try Voice Access, two important hotkeys to know are Alt+Shift+b for toggling the microphone between sleep and awake and Alt+shift+c for toggling the Microphone off and on. When sleeping, the microphone remains on to listen for certain words. See the support article or say, “what can I say,” when Voice Access is running for a full list of commands.

Leave a Comment

No Big Finish, Giving Microsoft’s Recall a Try

Microsoft released the Recall feature for users of Windows Insider builds. I gave the feature a try and made an audio recording with a roughly edited transcript. The short version of my experience is that I can get around Recall but am still trying to figure out how to read search results effectively with a screen reader, hence the title of No Big Finish.

Since this blog post is about an AI feature, I’ll offer an unedited version of Copilot’s summarization of my transcript as well. Overall the summary is reasonable but in my opinion misstates what I said about the Click To Do feature. I said I hadn’t figured out how it is supposed to work.

Perhaps because it was at the end, the summary also missed my biggest point entirely about still trying an effective way to use the search functionality of Recall and reading the meaningful text easily. Here is the Copilot summary:

The document is a transcript of a demonstration by Kelly on the Microsoft feature called “Recall.” This feature captures screenshots of your screen and allows you to search for information within those screenshots. Kelly explains that Recall was released as a test version for Microsoft Insiders on November 22nd. The demonstration focuses on the accessibility aspects of Recall, particularly how it works with screen readers like NVDA, Narrator, and JAWS.

Kelly describes the setup process, which includes signing in with Windows Hello, and demonstrates how to navigate the Recall interface using keyboard commands and screen reader feedback. The document details various functionalities of Recall, such as saving snapshots, opening Recall from the system tray, and navigating through captured images. Kelly also mentions some issues and limitations encountered during the demonstration, such as certain screen reader commands not working as expected and the need for further development to improve accessibility.

Kelly starts by explaining the initial setup, which involves a brief opt-in experience where the user grants permission for the computer to take screenshots. This process takes about 20 to 30 seconds and requires signing in with Windows Hello. Kelly notes that each time Recall is opened or used, the user is prompted to sign in with Windows Hello.

The demonstration continues with Kelly navigating the Recall interface. Kelly shows how to open Recall from the system tray and describes the various options available, such as saving snapshots, opening Recall, pausing the feature, and accessing more settings. Kelly highlights that the timeline feature in Recall displays the captured images, referred to as “memories,” and sessions, which represent the time periods during which the computer was used.

Kelly also points out some issues with the screen reader’s ability to read the titles of the snapshots automatically. Kelly demonstrates how to navigate through the list of snapshots and mentions that sometimes the titles do not read correctly on the first attempt. Kelly suggests using the home and end keys to navigate through the list and get the titles to read properly.

Additionally, Kelly explores the commands available for each snapshot, such as opening the associated application (e.g., Outlook), copying the snapshot, deleting it, and accessing more options. Kelly notes that some features, like the “Click to do” button, are not fully functional yet and require further development.

Kelly also demonstrates how to get more details about a captured image using Narrator’s command for image descriptions. However, the output is often gibberish due to the limitations of the OCR (Optical Character Recognition) technology used. Kelly suggests using other solutions like JAWS, Google’s Gemini, or Be My Eyes for better image descriptions.

Overall, the document provides an insightful look into the Recall feature, its current capabilities, and areas for improvement from an accessibility perspective. Kelly’s demonstration highlights the potential of Recall while also acknowledging the need for further development to enhance its functionality and accessibility.

Leave a Comment

Unlock Windows Efficiency with PowerToys Run

Sometimes you make a handy discovery that makes you wonder why you did not know about it long before. I just had that happen with PowerToys Run, one of the toys included in the full suite of items in the electronic toybox known as Windows PowerToys.

PowerToys are a set of utilities that allow you to customize different aspects of Windows and offer functionality that is not directly built in. You can learn more in general and find details on installation in a handy Microsoft Learn article.

I installed the most recent version of PowerToys because I had read about something new known as Advanced Paste. That is not the discovery though.

After installing PowerToys, I used one of my more common shortcuts in Windows Alt+Space to bring up the system menu for an application. That menu is where you find entries including restore, close and the one I often use, Maximize. My web browsing windows in particular often get sized quite small and with responsive design much of the content I’m expecting has disappeared so maximizing the browser window is necessary.

Imagine my surprise when instead of what I was expecting, my screen reader communicated, “query, edit.”

It turns out this is the default shortcut for the aforementioned PowerToys Run. In short, this is like having a command line to do everything from choosing between open windows on your computer to performing quick calculations, file, and web searches, browsing the Windows registry and more.

Using PowerToys run

Using PowerToys Run is fairly straight forward. Press Alt+Space, enter some text and arrow through the results. You can start your text with various text characters to direct PowerToys Run what you want to do. The full list of text to use here is detailed in a Microsoft Learn article.

Some examples I am finding I have already incorporated into my daily use include:

  • <, that is the less than symbol: Window Walker, to browse through all the open windows on your computer. Start pressing down arrow after entering the less than character, or add part of the window name and then down arrow to shorten the list.
  • $, that is the dollar sign character: Windows Settings to, as you would expect, browse through all the different settings for Windows. As with Window Walker, just start pressing down arrow after entering the dollar sign, or add some of the text from the setting you want and then press down arrow. In either case, when you reach the item you want, press enter and that settings page will open.
  • =, that’s the equals sign: Calculator for performing calculations. Just enter your calculation and the result will be displayed. If, as I do, you are using a screen reader, the fastest way I have found to read the result is to press down arrow after entering my calculation. Note that you do not need to press enter after you have input your calculation. Also, again if using a screen reader, if you are comfortable with advanced screen reading techniques such as the JAWS Touch Cursor, NVDA Object Navigation or Narrator navigation commands, the result of the calculation and more can be read with these techniques. Last, after your result is displayed, you can press enter on the result and have it put on the Windows Clipboard.
  • !!, that is two presses of the exclamation mark key: History, quickly browse through your history with PowerToys Run with presses of down arrow.

Some Notes

PowerToys Run documentation indicates you can change the default shortcut for launching from Alt+Space.

According to PowerToys documentation, pressing tab is supposed to move you through search results and any buttons or context menus that exist. I am not finding anything being communicated by multiple screen readers as of now when using Tab. I still need to figure out if this is a case of the key simply not working or items taking focus not being communicated properly.

Leave a Comment

Accessible Entertainment in the Air

Flying home from a recent vacation, I had a first for myself. I independently used the in-flight entertainment system to track my flight, listen to music, and watch a movie with audio descriptions. I even played a bit of trivia for a bit. How fun!

I’m not sure when United Airlines added accessibility to their in-flight technology, but it was available on the return flights of my trip. The system used TalkBack and a two-finger triple-tap to start the screen reader. There was a video offered to show you how to use the system. I was in the proverbial cheap seats, so I used just the touch screen option for control. Apparently, premium seats get physical buttons in the arm of the seat as well.

Aside from the map showing you flight progress and some games, I found all the other experiences worked well with TalkBack. Those that didn’t were indicated by a message talking about not being available with TalkBack. In the case of the flight map, the alternative to tracking flight distance, elevation, and such did work with TalkBack. I do wish that display had a compass option as well, but the experience just worked, so what more can you ask for when it comes to accessibility? Picking my own movie, having audio descriptions, and being able to check on my flight independently was pretty sweet!

Leave a Comment

Command Customization in Mac OS 15 with VoiceOver

One of the larger changes I’ve encountered in the Mac OS 15 betas is not immediately obvious unless you know to hunt for it. The VoiceOver Commanders item within VoiceOver Settings has been replaced with an item named Commands. But this is far more than just the renaming of the same old controls.

The various tabs for the different commanders such as Numpad and Trackpad have been replaced with check boxes to turn control on or off with each of those items as well as the Option key. Again none of this is new, just a bit of a different way to turn things on or off.

However, move past all those check boxes and you’ll find the new and powerful ability to customize VoiceOver commands. First off is a set of radio buttons to choose if you want to control VoiceOver with Built-In commands or Custom commands. Choose built-in and VoiceOver works the way it always has. Pick Custom commands and that’s where the power for customization appears.

An edit button becomes enabled and activating that you can customize more than 350 VoiceOver commands.

For each command the customization options are wide-ranging. You can capture a keyboard command, assign a trackpad gesture, assign a numpad command, assign a QuickNav gesture and an option key command.

Choices within the dialog allow you to filter the command list by the areas you’d expect including tables, text, information and the remaining VoiceOver command areas that appear when you bring up the VoiceOver command list. You can also filter based on the different command types, such as trackpad, numpad, assigned commands, unassigned commands, commands that are not able to be changed and more.

Using all of this is quite straight forward and a welcome enhancement for VoiceOver.

Leave a Comment

Using Rufus to Install Windows 11 on Unsupported Hardware

If you are not comfortable tinkering with your OS install and potential consequences, please ignore the rest of this post.

I have a range of computers I use, several of which do not meet the official hardware requirements to run Windows 11. Most notably, they either do not support what is known as Secure Boot and or do not have a TPM 2.0 chip. Every few months I see another mainstream publication write an article about how to install Windows 11 on such hardware. These all basically say the same thing, go set a few registry keys and run Windows setup. Sometimes this has worked for me and sometimes not.

Recently I tried what has proven for me to be the most reliable and easiest option here. This involves using an open-source program called Rufus to create a bootable USB drive. Note, even though you are creating a bootable drive, you do not have to boot off the drive. You simply need to run the setup off the created drive.

Rufus can be downloaded from https://rufus.ie/en/

Take note, use the download links within the table of product versions. The other download links are for different ads on the site as near as I can tell.

The author of Rufus has an extensive FAQ on Rufus at https://github.com/pbatard/rufus/wiki/FAQ.

In addition to downloading Rufus, you’ll want to download an ISO file. Windows 11 ISO files can be downloaded from https://www.microsoft.com/software-download/windows11

Locate the heading that says, “Download Windows 11 Disk Image (ISO) for x64 devices” and then download the ISO file.

Once you have these two downloads, run Rufus, ensure you’ve selected a USB drive and the ISO file from within Rufus and choose to start. Another dialog will come up with various setup options including a check box to bypass the Windows 11 hardware requirements. This is checked by default. Start the USB creation process at this point.

Once your USB drive is created, open it from within Windows and run Setup. Again, you should not need to boot off the USB drive. Finish setup and your OS will be updated.

Within Rufus as I mentioned you can select an ISO file. Although it isn’t announced as a split button, the select button here does support that functionality. Instead of pressing space or enter, press Alt+Down arrow and you can choose between select and download. If you do not have an ISO file, arrow down to download and press enter. You’ll return to the same button but this time it will be named download. Now press space or enter.

You are then given choices for Windows version, edition, and language. Each of these are separate selections where you pick from a list, choose continue and advance to the next choice. Eventually you reach the standard file save as dialog.

Note, although this process will update your computer to a released Windows 11 version, if you try and take actions such as joining the Insider program, the hardware checks for those installs will still fail. Additionally, the process of creating a bootable USB drive from a Windows Insider ISO is not supported by Rufus. You can create the drive, but it too will fail hardware checks. The Rufus FAQ clearly says insider ISOs are not supported.

As far as I can tell, if you put your machine into this sort of a state, Windows updates for security and such appear to download properly. I suspect when the next official OS release happens, that install will not work and you’ll need to update through another Rufus-created bootable USB.

Finally, although I’ve outlined using Rufus for Windows 11, it supports creation of more than just a bootable Windows USB drive. You can explore other options from within the program.

11 Comments

The Text of The Section 508 Refresh Act of 2024

In late July, Senator Bob Casey introduced the Section 508 Refresh Act of 2024. The text of the law has not been published online as far as I can find. The senator’s office sent me a PDF of the proposed legislation and I’m sharing it here with their permission.

I’m told the bill text should show up as S.4766 at https://www.congress.gov/bill/118th-congress/senate-bill/4766/text?s=8&r=16 but that there can be a backlog for legislation to show up here.

In an election year I don’t know how much traction a bill like this is going to receive but some of the accountability requirements proposed are stronger than I’ve seen before.

Section 7 of the legislation outlines the procurement process requirements. They include:

  • 30 days for an agency to report a violation of Section 508 from the time they are notified.
  • Mandatory civil penalties for delivering technology that violates Section 508 that is not less than 3 percent of the contract value.
  • A requirement that Section 508 defects be addressed within six months of being reported or the vendor faces termination of the entire contract.

There is much more in the legislation and I’m sure policy and legal experts will have more analysis in the coming months. This though is long overdue legislation in my opinion.

Any factual errors in this posting about requirements are my error. I’ve made every attempt to summarize what I’ve read correctly but please review the actual document.

If you are reading this document with a screen reader, I’ve found it reads more accurately in the full Adobe software then the PDF interpretations in Edge or Chrome.

One Comment

The Good and Bad of Accessibility in Two Minutes with the Olympics

Tuning into a bit of the Olympics this morning, within two minutes the reality of accessibility in 2024 is on display. Audio description for a channel showing multiple sports is impressive but the schedule view of the Olympics iOS app not so much. The progress is appreciated here but the gaps are still far too many.

NBC has taken a page from the NFL’s Red Zone and introduced a channel called Olympic Gold Zone. It provides whip-around coverage of events and live audio description of a two-box split screen for two sports mixing between live commentary is impressive as an example.

The Gold Zone channel is part of the coverage available on the Peacock streaming service. Scott Hanson of NFL Red Zone fame is one of the channel hosts. Coverage runs for 10 hours a day, starting at 6A central.

Downloading the Olympics iOS app and choosing schedule though, shows accessibility that would not make it to the metal round. My experience with VoiceOver was a jumble of words, untagged images and more. I had no success understanding the actual schedule.

Back to the Gold Zone, the live audio description is excellent. As you’d expect, you get details that are just not included in the standard TV broadcast. Player reactions, details about the stadiums, surroundings for events in the city and more. And all of that is mixed in with both the Gold Zone host and announcing from the sports. It will be a fun two weeks of athletic competition.

Leave a Comment