Overcast Update Adds Double Tap on Apple Watch
Over the weekend, I came across this Mastodon post in which it was announced popular indie podcast client Overcast, version 2025.6, received a “minor update” that includes not only the boilerplate big fixes and performance improvements, but also one accessibility enhancement of particular note: Double Tap on Apple Watch. The feature, which debuted with Apple Watch Series 9 and Apple Watch Ultra 2 in 2023, is available in the watchOS app to start and stop Play/Pause. Overcast is made by Marco Arment.
The update is described by Arment as “not an exciting one, but good nonetheless!”
The update grabbed my attention for a few reasons. For one thing, Double Tap is, like the iPadOS pointer feature which was redesigned in iPadOS 26 such that the pointer is a Mac-like arrow rather than a circle, one feature whose origins trace back to AssistiveTouch. As with the pointer, it’s my understanding from sources the Double Tap accessibility feature—designed to help those with limited mobility, motor-wise—was handed off to the broader software engineering team within Apple Park to massage it into something with more mainstream sensibility. That Overcast users can now use Double Tap makes playback controls more accessible not only if one’s hands are full, but also if tapping the Watch’s screen would otherwise be burdensome or untenable altogether. For another reason, the advent on Double Tap in Overcast’s Watch app is yet another strong show of solidarity with the disability community by Arment. He’s been a staunch ally of disabled people dating back to days as Instapaper’s keeper, so this latest Overcast only strengthens his resolve in this regard. I interviewed Arment back in 2018 for a story for MacStories’ coverage of the App Store turning 10. He told me over email in part he considers working on accessibility “a peer to other aspects of my design and structure, such as the colors I choose or how interfaces are laid out.” Lastly, Arment has spoken positively of Apple’s new Speech APIs on ATP and his enthusiasm portends well for Overcast getting transcripts sometime in the not-too-distant future.
Overcast’s 2025.6 update is available now on the App Store.
Entrust Product Design Chief Mark Opland Talks the European Accessibility Act, More in Interview
A couple of weeks ago, a landmark law spearheaded by the European Commission went into effect. It’s called the European Accessibility Act (EAA) and it mandates digital goods be accessible to people with disabilities. The June 28, 2025 timeframe was the cutoff date for member states in the European Union to reach compliance, but the law was officially passed back in 2019. The deadline was imposed so as to achieve synchronicity concerning digital accessibility spanning the European Union. Companies whose products shipped to customers prior to June 28 have a grace period of 5 years to comply with the EAA. Hardware has double the time, with 10 years’ grace.
A cogent primer on the EAA was posted in late January on the AccessibleEU website.
One company that’s been thinking a lot about the EAA is Entrust. A digital identify firm based in the United Kingdom, Entrust offers a suite of products and services which cater to financial institutions like banks. Entrust builds everything from ID verification technology to encryption tech and more. In an interview conducted last week via videoconference, Entrust’s vice president of product design, Mark Opland, encapsulated his company’s scope as “[offering] an enormous amount of products and services to financial institutions, but really centered around identity and security.”
When asked about why accessibility matters to him and his team, Opland explained it has been “a huge part of the way we’ve built products for a long time,” adding accessibility has been personally pertinent for the better part of 15 years. To raise awareness for accessibility, he told me, not only aligns with his value system, but helps Entrust “deliver more successful products and services into the industry.” Accessibility, Opland went on to say, isn’t viewed as a “constraint” for Entrust; rather, the company views it as an opportunity to innovate and thus better build its business. Accessibility can be, and often proves to be, “an enabler of innovation” for Entrust, Opland said.
“If we fundamentally approach design problems and product problems thinking about the largest possible user base in mind, we ultimately build products that are more successful,” he said of Entrust’s philosophy on prioritizing accessibility in its work.
As to the EAA, Opland said the legislation is a directive aimed at “[making] a wide range of products and services more accessible to people with disabilities,” adding the European Union considered the things people used day-to-day in an effort to contribute to the betterment of society and wanted to find a way to “[encourage] greater inclusion and breaking down the barriers across the European Union for all people.” The EAA, he continued, touches myriad industries and, as such, while compliance to the EAA is compulsory, the byproduct of it is what Opland characterized as enabling businesses to “tap into a much larger customer base.” He pointed to a large bank in the United Kingdom who reported its total addressable market increases by more than 10% when they build products with accessibility in mind. “For the European Accessibility Act and the European Union, it’s not only about providing access, but about building their GDP and increasing the [gross domestic product] for all their member states,” Opland said.
For Entrust’s part, Opland made crystal clear his main job as it relates to the EAA is to ensure the company enters into, and then maintains, compliance with the law. Entrust must follow the law’s legal structure and, more pointedly, “we can’t be building and shipping products anymore that are not accessible.” Opland was forthright in telling me he cares not about being the “accessibility police” and running around into people’s offices internally to enforce abiding by the EAA. Instead, he told me the company has spent lots of time leading up to last month’s deadline auditing and doing remediation. Moreover, Entrust has focused its energies on prioritizing advocacy and evangelism with the goal of what Opland said is “building a culture of continuous improvement.”
“Our goal is to make sure every team at Entrust, whether it’s Human Resources or an engineering team, is focused on making sure they’re better this quarter than they were the last quarter and better next quarter than they were this quarter,” Opland said. “That advocacy has us out of the cycle of managing accessibility from audit to audit, and seeing the job is being done when we earn our accessibility accreditation. This focus on continuous improvement means it’s top of mind for everyone in the company and has now become part of our DNA… that’s been the secret to our success over time.”
Opland acknowledged coming into compliance with any sort of law has its challenges, but in context of the EAA, the economic and social benefits can make the headaches worth it. Especially from a social justice perspective, he said “it’s been fantastic” to work with the law at Entrust because it aligns with both his personal and the company’s institutional values. The main theme that threaded my conversation with Opland is that greater accessibility vis-a-vis the EAA is two-pronged: it benefits people obviously, but it also benefits businesses. The more people one markets to, the bigger its bottom line can become. Understanding those principles takes education, and Opland told me it can be challenging unto itself to teach people how to make accessibility happen. The EAA, as a law, and accessibility standards like WCAG aren’t necessarily congruent with one another; Opland said they “aren’t always black and white… in some places they’re gray.” Entrust understands “there’s always a tradeoff between usability and security,” according to Opland, which isn’t always a question with a black-and-white answer.
“What we’ve discovered is the more you have those conversations, the more you dig in and the more you learn, the stronger and more resilient you become,” Opland said of the company’s learnings. “Accessibility is a unique challenge in that there is often quite a lot of subjectivity and just a huge spectrum in human ability. There isn’t just sort of a one-size-fits-all solution that’s going to allow me to wave a magic wand to make everything accessible. I think it’s just a constant cycle of learning and improving.”
Entrust has been working on accessibility for close to a decade, or 9 years, now. This gave Opland and team a lot of runway in terms of comfort and confidence when the company felt the looming EAA deadline. The work on compliance, he told me, had been an 18-month effort into understanding EN 301–549—which is linked to the EAA—WCAG, and the EAA itself. Companies like Entrust who are generally concerned with the aforementioned WCAG standards, Opland told me, are “in a really good position to be compliant with the law with the exception of a few slightly more specific directives.”
“If you’ve been focused on WCAG, you set yourself up really well,” Opland said. “We’ve had a pretty big head start and have been positioned pretty well to be compliant.”
Opland is optimistic the EAA will help make accessibility more top of mind and more present in products. The European Union, he told me, has set a standard because to EAA applies to anybody who wants to do business in the Union, so products and websites must meet the new regulation. In the United States, Opland pointed to the Americans with Disabilities Act as greatly improving the quality of life for the disability community, but conceded there is more work yet to be done. He hopes businesses everywhere “continue to invest” in accessibility for the people—and for their business.
As to the future, Opland and Entrust are committed to walking the righteous path.
“Our hope is we are continuing to build products and services that enable more and more people to enact with their communities, to enact with the businesses around them, [and] to have more opportunities and greater advantages in the lives they lead,” he said of Entrust’s view of its future work. “There’s something really meaningful and deep in doing so. Identity is such a great vehicle to help advance underrepresented folks in all stations of life, and accessibility is one important aspect of that. If you track back to Entrust’s mission, it perfectly aligns with our mission. It perfectly aligns with our growth as a business. We were working on [accessibility] long before the law mandated we do it. We’ll continue to invest in accessibility, whether the law continues to mandate it, so it just aligns perfectly with our mission, with our values, and with our business.”
How Type to siri trumps courtesy and convenience
MacRumors contributor Aaron Perris posted on X today Apple has started airing a new Apple Intelligence ad which highlights Type to Siri and ChatGPT. The 15-second spot, which also features iPhone 16 Pro on T-Mobile’s network, takes place in a workplace elevator. As of this writing, the video isn’t (yet?) on Apple’s official YouTube channel.
I wouldn’t typically cover the advent of a new Apple commercial, but this particular one merits an exception. During the Apple Intelligence portion of last year’s WWDC keynote, senior vice president of software engineering Craig Federighi talked up Type to Siri as a feature of convenience: it’s a mode by which people can quietly interact with the virtual assistant so as not to be disruptive of others. The reality is Type to Siri is not an all-new feature; it’s existed as an accessibility feature on iPhone and iPad since iOS 12 in 2018.
This context matters greatly in the grand scheme. It is extremely noteworthy that Apple “graduated” what was once an ostensibly esoteric, niche assistive technology and expanded upon it so as to become more mainstream. Despite Federighi’s message to the masses that Type to Siri is about courtesy and convenience, the truth is the feature’s benefits for accessibility remain bountiful. Yes, courtesy and convenience are important factors, but Type to Siri is a great feature whereby a Deaf or hard-of-hearing person or, in my case, someone with a speech delay can interact with Siri with complete fidelity without the voice component. That isn’t at all trivial or ancillary to Apple’s core messaging. The overarching point is Type to Siri illustrates yet again that accessibility oftentimes is an incubator for innovation—it’s something Apple rarely, if ever, gets lauded for by those who comprise the mainstream technology commentariat.
As I alluded in the previous sentence, Type to Siri stands not alone. The pointer feature in iPadOS began life as an AssistiveTouch feature, of which Apple’s Sarah Herrlinger told me years ago “isn’t your typical cursor.” My understanding has long been the company’s Accessibility team handed off the AssistiveTouch feature to the broader iPadOS software group so they could massage it into something meant for more mass adoption. Likewise, the Double Tap feature on Apple Watch germinated as an AssistiveTouch feature in watchOS, was then similarly made over for broader applications. Many popularized modern technologies—audiobooks, speech-to-text, et al—were invented by disabled people for their unique needs, then adopted by the able-bodied masses for their enjoyment. As Dr. Victor Pineda told me last year, the disability community is chockfull of technologists out of sheer necessity. Technology makes the world more accessible to people like Dr. Pineda (and yours truly). Last December, Apple used its precious holiday ad space to highlight the hearing aid feature on AirPods Pro. My understanding is the ad, called “Heartstrings,” was the first time the company used an accessibility feature in the holiday campaign—and for good reason. It shows the profundity of assistive technologies truly being for everyone, with earbuds everyone uses every day. It’s a rare example of people being able to have their cake and eat it too.
So yeah, Type to Siri is highly significant—especially so, again, in a TV commercial.
Max Says ‘Sinners’ will Stream in Black American Sign Language in ‘Groundbreaking’ First
In a press release published on Monday, popular video streamer Max announced the Ryan Coogler-helmed movie Sinners will stream in Black American Sign Language (BASL) alongside the original version in the United States on Independence Day this Friday, July 4th. Max touts the release is “groundbreaking” while noting it “marks the first time a streaming platform will exclusively debut a film interpreted in BASL.”
The BASL interpreting of Sinners will be done by Nakia Smith. A trailer is on YouTube.
“The release of Sinners with BASL is a major step forward in accessibility, representation, and visibility in streaming. BASL is a distinct dialect of American Sign Language (ASL) with its own dynamic history and unique grammar, signing space, rhythm, facial expressions, and cultural nuances,” Max said in its announcement. “For the first time, the Black Deaf community will have streaming access to a more immersive experience in their language. Max subscribers, who sign in ASL but are unfamiliar with this dialect, will also be able to follow along with this interpretation.”
Max, like its peers in Apple and Netflix, reaffirmed its support of disabled people.
“Accessibility within streaming is not a one-size-fits-all approach. Our goal at Max is to make these great stories accessible to all audiences in a way that is authentic to the content and the communities we serve,” Naomi Waibel, senior vice president of global product management at Warner Bros. Discovery, said in a statement included in Max’s press release. “Sinners with Black American Sign Language is an example of how culturally nuanced access can enrich the viewing experience for our audiences.”
Today’s news comes after Max announced a similar initiative this past March to stream The Last Of Us in ASL. Elsewhere, the National Hockey League, which has a TV deal with Warner Bros. Discovery-owned TNT, has aired highly successful “NHL × ASL” broadcasts in partnership with PXP. They’ve proven so successful with fans the league earned Sports Emmy nominations for its work in furthering accessibility and inclusivity.
Max announced in May it’ll mercifully rebrand itself (back) to HBO Max “this summer.”
Accessibility Amplifies Apple Music’s first decade
Apple on Monday issued a press release in which the company celebrates Apple Music’s 10th birthday by sharing some big announcements about the service. The headliner is a new three-story campus based in Culver City, which Apple says sprawls more than 15,000 square feet and houses two radio studios and a 4,000 square foot soundstage. The new campus is designed to “anchor a global network of creative hubs” in other cities such as Berlin, Nashville, New York, Paris, and Tokyo, according to Apple.
“Apple Music Radio has always been a home for storytelling and artistry, serving as a space for bold conversations and surprising moments,” Rachel Newman, co-head of Apple Music, said in a statement included with today’s announcement. “With this new studio, we are furthering our commitment to creating a space for artists to create, connect, and share their vision.”
Amongst the other news is the advent of a playlist Apple calls Replay All Time, which the company describes as “a special version of the annual Replay experience that allows listeners to see and stream the songs they’ve played the most since joining Apple Music.” Replay All Time can be found in the Home tab in the Music app, Apple said.
As with Apple Podcasts, Apple and music have been constant in my everyday digital life since getting my first-ever Apple product, the original iPhone, in 2007. Until Apple Music debuted in 2015. I spent a lot of money buying songs and albums in iTunes and synced my music via cable to myriad iPods and iPhones of various vintage. Those purchases remain available to me today, of course, along with the streaming content Apple Music provides. From an accessibility standpoint, the “all-you-can-eat” model of streaming Apple Music is great because I no longer need to fiddle with a physical cable to sync data, which involves somewhat tortuous tests of my lackluster hand-eye coordination. It’s also easier on my wallet too, since I needn’t budget money on individual albums from my favorite artists. Likewise, that the iPhone subsumed the iPod’s functionality—the Music app in what was then known as iPhone OS 1 was literally named iPod—means I have an “all-in-one” device and needn’t carry a separate iPod along with a cell phone, a setup I contemplated prior to getting the iPhone. It’s more accessible for me to carry one small object than two (admittedly small) objects, especially when you have relatively compromised muscle tone in your hands and less strength overall to accommodate weightiness. Moreover, from a software standpoint, it’s meaningful how the Music app supports accessibility features such as VoiceOver, Dynamic Type, and most recently, Music Haptics. It makes the listening experience more accessible—and thus more enjoyable—that I can follow along to the words in a song in the app’s Lyrics View on the Now Playing screen, for example. And once again, with the nature of streaming and technologies like iCloud sync, I can move from my iPhone to my iMac and beyond and have all my music ready when I’m ready to listen.
I look forward to using Apple Music into 2035 and beyond.
Google Adds Captions to Gemini Live Conversations
Abner Li reported for 9to5 Google earlier this week Google has added support for captions to Gemini Live in the company’s eponymously-named Gemini app on iOS and Android. The captions began appearing for some users earlier in June, according to Li.
“When you launch Gemini Live on Android or iOS, a rectangular captions button appears in the top-right corner. Tapping will enable a floating box that provides a transcript of Gemini’s responses. (This does not show what you’re saying in real-time, but that remains available in the full text transcript after ending the conversation.),” Li wrote. “It appears near the middle of the fullscreen interface in audio mode, and at the top when video streaming is enabled. These three lines of text cannot be moved or resized. In Gemini > Settings, there’s a new ‘Caption preferences’ item underneath the ‘Interrupt Live responses’ on/off toggle that links to system settings on Android.”
The big takeaway is, obviously, conversations with Gemini will be more accessible.
As I’ve noted before, I have the Gemini app for iOS on my iPhone’s Home Screen, as well as a widget on the Lock Screen. I really enjoy Gemini as my preferred generative AI tool, and have found it has supplanted much, if not most, of my web searches via Safari. I find Gemini to be a way more accessible (and digestible) method to get quick bursts of information collated in one place rather than manage a half-dozen browser tabs. “Trust but verify” goes the axiom, of course, so I’m well aware Gemini will (and does!) hallucinate from time to time, but I’ve been more than satisfied with its performance overall. I have a ChatGPT Plus subscription too, since notably Apple Intelligence integrates with it, but I generally like the Gemini app experience better. Perhaps that’ll change once the Jony Ive-Sam Altman partnership bears more fruit, but for now, I’m a happy Gemini user. Despite the rapidity with which AI seemingly advances nowadays, the reality is the technology still is really early in the proverbial ballgame. That Google—and OpenAI, for that matter—is clearly committing to making its respective tools accessible is a ray of hope for the inclusiveness of AI’s ever-burgeoning capabilities.
Google’s Successor to the Nest × Yale Lock Arrives
Ben Schoon reports for 9to5 Google the first-ever “Google Home Preferred” product is here and it’s the successor to the Nest × Yale lock: the $189 Yale Smart Lock with Matter.
The Matter moniker is an important detail, as it means the Yale Smart Lock can be used with Google Home, Alexa—and, pertinent to my preferred ecosystem, Apple’s HomeKit.
“This lock is very much designed to fit into Google’s ecosystem and acts as a successor to the Nest × Yale Lock,” Schoon said of Yale’s newest smart lock. “The design of the lock is meant to match the finish of Google’s Nest Doorbell lineup, with ‘Snow’ and ‘Matte Black’ finishes available today and an ‘Ash’ colorway coming later on. The accents on each are meant to match common door hardware finishes. As a backup to your app or a keycode, there’s a keyhole which was missing on the Nest × Yale Lock.”
I’m writing about this because (a) Curb Cuts is my website; but (b) because I’ve been using the legacy Nest × Yale lock for a few years now. It still works with aplomb, but admittedly part of my allegiance to sticking with it is due to the fact I’m simultaneously clutching to the OG Nest app on iOS and iPadOS for dear life. I do have Google Home on my devices too, but the UI, design-wise, is inferior to that of the old Nest app. Someday the Nest app will be put out to pasture and I’ll begrudgingly have to adopt Google Home. But today is not that day, so I’ll be riding with the Nest app until the absolute very end.
Speaking of an end, this week’s news from Schoon on the new Yale Smart Lock means damn near every device in my smart home setup—all devices running through HomeKit via the Starling Home Hub—is, while remaining perfectly serviceable in a functional sense, is otherwise “antiquated” and summarily discontinued technologically.
Nest Hello doorbell
Nest E thermostat (with accompanying room sensors)
Nest Protect smoke and carbon dioxide detector
Nest Outdoor Cams
Nest × Yale door lock
The Nest × Yale lock in particular has been a game-changer for me in terms of accessibility. It only controls the deadbolt, however, as my partner still prefers a physical key for the actual doorknob. Nonetheless, not having to fiddle with the key on both locks is far more accessible; my lackluster hand-eye coordination makes it such that it can be hard to find the keyhole, insert the key, and turn. Especially when coming home with, say, a bag of groceries, that I can use my iPhone—or better yet, Siri—to unlock the door transcends sheer convenience. It makes my house more accessible.
As to a potential upgrade, I’m intrigued by another Yale product: the Assure Lock 2 Plus. On its website, the company describes it as “the smart lock made for Apple users” as it supports Apple’s Home Key feature. Released with iOS 15 in 2021, Home Key allows users to use their iPhone or Apple Watch as their “house keys” by integrating with the Wallet app on iOS and watchOS. The reason I’m so fascinated by Home Key is, of course, accessibility; instead of tapping a button, I could simply hold my device close to the Assure Lock 2 Plus and the door would unlock. This is exactly why I adore the Auto Unlock feature in Waymo, whereby the car doors unlock as you approach. The Waymo One app does have an Unlock button, but it’s far more accessible to not have to tap it.
For the foreseeable future, though, I’ll be clinging to my OG Nest × Yale lock.
My bit Part In Apple Podcasts’ two Decade story
Apple on Thursday celebrated a big anniversary: Apple Podcasts has turned 20! To mark the milestone, the company released a staff-selected list of “20 podcasts we love.”
“Since the medium came to iTunes in 2005, our team has dedicated countless hours to helping people discover new shows. To celebrate 20 years, here are 20 favorites that best exemplify how far podcasting has come—and where it can go in the next two decades,” Apple writes in the list’s introduction. “This list is a love letter to the podcasts that left a lasting impact on us and the ones we continue to recommend again and again. They are shows with hosts that feel like friends, and shows that make us press play immediately on the latest episode to hear what happens next. These shows have measurably improved our lives and helped define this medium we know and love.”
Of those on Apple’s list, only The Daily (started in 2017) is one I listen to religiously.
Launched in 2005, Apple Podcasts predates my usage by a couple years; my first-ever Apple product was the original iPhone two years later. Over the last 18 years, however, podcasts have remained a constant in my digital life. I love them for the background noise they provide as I work on stories like this very piece, for the ways newsy shows like the aforementioned The Daily keeps me informed, and for the ways they let me indulge in nerdery on shows from friends of mine in the Apple/tech media communities. Once upon a time, I even had a podcast of my own called Accessible. The show’s website/listing is long gone from the web, but it was a fortnightly program I co-hosted with my close friend Timothy Buck during which we discussed all things accessibility in tech. We had a good run, even interviewing Apple’s accessibility boss in Sarah Herrlinger in person at one San Jose-based WWDC. I’m decidedly not an active podcaster nowadays, but have guested on my share of shows since Accessible unceremoniously ended. I do think about getting back into the game from time to time, but for now, I think I’ll focus my energies into getting Curb Cuts featured in Apple News.
(If you’d like me on your podcast to talk disability inclusion and the like, get in touch.)
From 30,000 feet, Apple has generally been a strong steward of its Podcasts platform. From an accessibility perspective, it’s certainly damn notable how the company has invested time and resources in making podcasts more accessible through transcripts. As with Music Haptics in Apple Music, it’s not at all trivial that, as I’ve espoused many times recently, Apple is taking an ostensibly exclusionary medium to, say, Deaf and hard-of-hearing people and making it eminently more inclusive vis-a-vis transcripts.
Apple’s “20 Podcasts” list follows a “100 Best Albums” list shared on Apple Music.
Thoughts On AirPods, Cameras, and Accessibility
Ryan Christoffel reports for 9to5 Mac this week a new feature in iOS 26 is a harbinger of camera-equipped AirPods Pro. Among the enhancements coming to AirPods this year is a Camera Remote feature. The functionality is similar to that on Apple Watch, whereby users can use the Watch as a shutter button; as Christoffel writes, users can “use AirPods to capture photos and video on your iPhone in situations that’s helpful… either by pressing once on the AirPods stem or pressing and holding—your choice.”
Christoffel goes on to say the Camera Remote feature on AirPods is notable because rumors suggest the next generation AirPods Pro are, again, said to include cameras. A release timeframe is unknown outside of Apple Park, but Bloomberg’s Mark Gurman has reported the refreshed earbuds feature “external cameras and artificial intelligence to understand the outside world and provide information to the user.” That iOS 26 does include the Camera Remote feature is evidence Apple is getting its proverbial ducks in a row by ensuring its software can support its forthcoming hardware whenever it comes.
There’s much to extrapolate from Christoffel’s informed speculation, not the least of which is how intrepid observers (and spelunkers) oftentimes will notice Apple setting the stage in advance of grand reveals. The size class APIs introduced at WWDC 2014 presaged the iPhone 6 and 6 Plus announcement a few months later. HealthKit, also new in 2014, was destined for Apple Watch. At WWDC 2015, iPad multitasking by way of Slide Over and Split View were announced prior to the iPad Pro’s unveiling. And in more recent times, the advent of ARKit in 2017 seemingly provided clues Apple was working on something using augmented reality; that hunch proved prescient as the company eventually revealed Vision Pro at WWDC 2023. It could even be persuasively argued the all-new Liquid Glass design language, ostensibly meant for today’s devices, was created with an eye towards still-in-development AR-powered glasses for tomorrow.
In an accessibility context, the Camera Remote feature on AirPods strikes me as fascinating. To wit, it’s entirely plausible someone who wears AirPods also doesn’t have use of their arms/hands, whether wholly or in part. While it’s possible today for said person to use voice to control playback and the like, having Camera Remote on AirPods would give a person whom I mention another avenue through which to accessibly take pictures. Likewise, that Gurman said Apple reportedly plans to imbue these future AirPods Pro with AI chops such that people could better “understand the outside world” means, for example, navigation could become easier, as might people recognition. Imagine Siri telling a Blind or low vision person, Her-style, that the person approaching them to say hello is their brother or sister or someone else in their digital rolodex. There are other possible applications, of course, but these examples are intriguing because they potentially can make the world literally more accessible to disabled people. It’s cool tech, but also genuinely useful—and empowering to boot.
Such a sentiment is a common refrain with AirPods Pro lately.
Waymo is making its way to new York city
Andrew Romero reported late last week for 9to5 Google that Waymo has teased New York City as its next place it plans to “plant autonomous ride-sharing.” As Romero caveats, however, the proverbial seeds are going to take some time to bear fruit.
According to Romero, the Alphabet-owned Waymo took to X last week to share it has officially initiated the process to get its autonomous vehicles running on the streets of New York City. The company announced it has applied for a permit with the city’s Department of Transportation as a first step, noting a “specialist” from the agency will be sitting behind the wheel of Waymo’s Jaguar SUVs. Waymo also noted it’s working with state lawmakers to amend legislation so as to legalize fully autonomous vehicles.
“We want to serve more people in more places, including New York,” Waymo said.
The NYC news comes after Waymo announced expansion to Washington DC in March.
I’ll have more Waymo news on Curb Cuts in the coming weeks, but for now, any news of expansion is great for accessibility’s sake. Last week, I took part in a panel discussion at the Accessible Futures conference during which I spoke of the immense accessibility gains rideshare services have for Blind and low vision people such as myself. Of course Lyft and Uber have relevance here, but I spoke most enthusiastically of Waymo and the positive effects it has on my life. (Video and a transcript will be posted soon, I’m told.) As Waymo attests, there’s a helluva lot of bureaucratic stuff to tend to first, but any expansion news is heartening because it means (a) Waymo is doing well, business-wise; and (b) arguably more importantly, it means greater agency and autonomy in transit for myself and others like me across the country. The more places Waymo sets up shop, the more places we non-drivers can go—with greater independence, no less.
The Waymo-to-NYC news is joined by Tesla launching its robotaxi service in Austin.
Amazon’s Recent Kindle Software Update Adds More Line, Text Spacing options for accessibility
Andrew Liszewski reports for The Verge today on Amazon releasing a Kindle-based software update to users which, according to the release notes, includes the boilerplate performance enhancements and bug fixes, but notably includes upgrades for better visual accessibility. He writes the improvements come by way of “adjusting text and line spacing, improving legibility and accessibility for many users.” The new 5.18.3 update is supported by the Kindle Scribe, Kindle Colorsoft, as well as the 11th and 12th generation of the Kindle and Kindle Paperwhite models, according to Liszewski.
“Amazon is slowly rolling it out through the Kindle’s automatic updates system,” Liszewski said of the company’s recently released upgrade. “[If] you don’t want to wait, you can download the specific update file for your e-reader, copy it over to your device, and perform a manual update using the instructions Amazon has provided.”
I have a Paperwhite from 2018 and it remains a nice piece of kit, although I haven’t used it in quite some time. While I find e-ink displays to be generally accessible and easy on my eyes, I’ve actually come to favor using Apple Books on an iPad for reading books. The brightness and sharpness of the display—especially on that of the OLED screen on the M4 iPad Pro—is far nicer to look at and even more accessible. That said, I’m deeply intrigued about the aforementioned Kindle Colorsoft from Amazon; I’d love to try it out someday and then subsequently write about my experience using a color e-ink screen.
News of the Kindle 5.18.3 update was first reported by The eBook Reader.
Apple’s New Speech Frameworks Portend More Accessible Audio for Developers and users alike
John Voorhees wrote a piece for MacStories earlier this week which delves into a new class in Apple’s Speech APIs and compares it to OpenAI’s open source speech-to-text model called Whisper. Apple’s class, named SpeechAnalyzer and its corresponding module in SpeechTranscriber, both are included in the developer betas released during last week’s WWDC festivities. As Voorhees explains, Apple’s APIs are highly performant.
Software development minutia notwithstanding, the nut of Voorhees’ story—especially pertinent to accessibility—is his conclusion that the speed with which Apple’s framework works is “a game-changer for anyone who uses voice transcription to create text from lectures, podcasts, YouTube videos, and more.” By contrast, Voorhees writes Whisper is inexpensive yet glacially slow; he calls its pokiness “frustrating” when trying to finish a YouTube project. Voorhees’ son, Finn, cobbled together a command-line utility aptly named Yap that chews up audio and video files and spits them out as SRT and TXT files. Notably for accessibility purposes, SRT is a plain-text file format commonly used for displaying captions and subtitles in videos. Using Yap, built atop Apple’s aforementioned APIs, it took Voorhees a mere 45 seconds to create an SRT file for his AppStories podcast. In MacWhisper, Voorhees reports the conversion process took 1:41 and 3:55 using OpenAI’s Large V3 Turbo and Large V2 models, respectively.
Voorhees posits Apple’s frameworks represent what he described as “a significant leap forward in transcription speed without compromising on quality” while adding he believes the new APIs will “replace Whisper as the default transcription model for transcription apps on Apple platforms.” The advent of the APIs are a noteworthy development, seconded by John Gruber. He writes today it “bodes very well for all sorts of use cases where transcription would be helpful, like third-party podcast players.”
Podcast players. I’ve extolled the virtues of having high quality transcripts in podcast apps many times before; to have them is to turn what’s ostensibly an exclusionary medium to, say, Deaf and hard-of-hearing people into something substantially more inclusive. Before WWDC, I advocated for Apple to build a “TranscriptKit” API which helped App Store developers include transcripts in apps such as Marco Arment’s Overcast. As far as I know, no such tool came to be this year—but perhaps it will come in the future presumably built with these very technologies. As it stands, however, part of the problem with including transcripts in any audio app is production—it takes a while to generate accurate transcripts in a timely fashion, as Voorhees illustrates. Apple’s new APIs seemingly are designed to solve that issue, the net effect of which could be more accessible and inclusive audio apps from Arment and others in the years to come.
Yap is available for download on GitHub.
Forbes Announces First-Ever ‘Accessibility 100’ list
In surprising news, Forbes this week released its “Accessibility 100” list. As the name implies, the publication is using its famous who’s who lists to spotlight 100 people who are doing notable work in the accessibility arena. The Accessibility 100 is a Forbes first.
“Accessibility is about far more than wheelchair ramps or live captioning,” Alan Schwarz, assistant managing editor of Forbes Media, writes in the lede of the introductory portion of the list. “The field has emerged as a bustling innovation hub, an educational imperative and—unapologetically—an untapped business opportunity.”
The list is a veritable all-star team of A-listers in the accessibility space. People from Apple’s Sarah Herrlinger to her counterpart in Google’s Eve Andersson to Microsoft’s Jenny Lay-Flurrie, amongst others, all are people I have interviewed numerous times in the past. As Schwarz notes, Forbes defined accessibility in terms of software; this explains why folks like the aforementioned Herrlinger and Andersson appear on the list.
The disability community deserves more time in the limelight like the Accessibility 100.
“[Prioritizing work on accessibility] isn’t [merely] a legal obligation or ‘We just want to do nice things for those poor people,’” Marcie Roth, executive director of the World Institute on Disability, said in a statement to Schwarz. “You’re leaving money on the table because you’re only marketing to three out of four people.”
Who knows, maybe I’ll make next year’s list should Forbes include us newspeople.
Google’s June Pixel Drop Includes Android 16 for Pixel Devices, Accessibility Improvements, More
Google last week released its Pixel Drop update for June. The monthly software upgrade includes, most notably, the official release of Android 16 for Pixel devices, VIP contacts, accessibility enhancements, and much more. The brand-new improvements were highlighted in a blog post by Austin S. Lin, technical program manager at Google.
“Your Pixel just got an upgrade with the June Pixel Drop, full of new features and updates that make your devices more helpful, personalized and accessible,” Lin wrote in the introduction to Google’s announcement. “Android 16 also starts rolling out to Pixel devices today. Here are a few new things you can do with your Pixel!”
Most pertinent to my purview are the accessibility improvements. According to Lin, users are able to use Live Search in the Magnifier app—available on the Play Store—to help them “learn more about your surroundings in real time without needing to take a picture first.” If you’re looking for a specific dish on a restaurant menu, for example, it’s now possible to type what you’re looking for and Magnifier then will highlight matches on screen, along with giving a little haptic nudge. Elsewhere, Lin wrote it’s possible to use LE audio features with hearing aids to “take calls on the go and access your hearing aid presets, as well as change your ambient volume through your Pixel phone settings.” This functionality is available on Pixel 9 and newer running Android 16, Google said.
Finally, Lin notes Google’s Expressive Captions accessibility feature has been updated so as to “capture elongated words” such as yessss—particularly useful with sporting events. Beyond the United States, Expressive Captions is available in Canada and the United Kingdom. I wrote about Expressive Captions late last year for my old Forbes column, as the feature first arrived on Android as part of the December 2024 Pixel Drop.
Google has posted a video to YouTube all about the June Pixel Drop.
The shoe is on the other foot, WWDC Edition
As I write this, it’s been exactly one week since I was on the ground at Apple Park in Cupertino to cover this year’s WWDC keynote. In the days since, the majority of the chatter online has surrounded Apple’s all-new design language called Liquid Glass. It’s been interesting to observe the discourse, mostly because a common refrain has been people complaining about things like contrast and readability. The first iOS 26 developer beta, the peanut gallery alleges, goes way too far in prioritizing the glass-like effects so as to be very detrimental towards legibility and other visual considerations.
In other words, iOS 26 Beta 1 has people talking about accessibility.
On one hand, I’m appreciative people are paying attention to the accessibility of Liquid Glass. More able-bodied people should pay attention to accessibility. I offer the caveat, however, the aforementioned beta is intended for software developers—WWDC is a developer-oriented event after all, with the public beta not slated to ship until next month—so it’s very early days in the testing cycle. There’s a whole summer yet for Apple to tweak and refine Liquid Glass before it’s officially released come September. The early returns on public opinion, at least what I’ve noticed on social media, fail to take into account that (a) Beta 1 is for developers; and (b) there’s time for improvements.
The umbrage from some people over the current state of Liquid Glass is eye-rolling enough, but what really gets my eyes rolling is how so many people who are running the beta are co-opting accessibility. These commenters are convinced Apple has turned its back on its ethos on accessibility by shipping Liquid Glass in iOS 26 in such a lackluster state by accessibility standards. I find the criticisms egregiously disingenuous; not because I devalue constructive criticism or believe Liquid Glass is bereft of problems, but because I believe people are tokenizing accessibility insofar as it suits their gripes.
Put another way, accessibility is ancillary… until it isn’t.
It’s rich. As I wrote on Mastodon last week, to see people invoke the accessibility argument is hilarious, if not somewhat infuriating, because people like me have been banging the proverbial drum on the topic over a decade now. Obviously, I’ve built my career in tech journalism by effectively creating my own beat in accessibility and assistive technologies. My lived experiences as a lifelong disabled person who copes with multiple conditions allows me to write authoritatively and expertly on the subject. I know this stuff, which is why I’m able to detect bullshit when I smell it. The majority of the time, accessibility coverage is confined to Global Accessibility Awareness Day or so-called “Accessibility Weeks.” There is no regular, daily coverage of disability at most newsrooms, especially at the old guard outlets like The New York Times, despite the fact disability deserves every ounce the limelight that gender, race, and sexuality get in contemporary social justice reporting. Similarly, the WWDC coverage I’ve seen makes no mention of the new accessibility features Apple announced last month—all of which are traditionally included in the platforms’ public releases. Of course Liquid Glass and iPadOS 26 are amongst the biggest pieces of news, but I say it’s downright journalistic malpractice to not at least add a sentence or two to one’s copy that macOS 26 Tahoe, for instance, gets the Magnifier for Mac feature, amongst others, later this year. You needn’t be an accessibility aesthete like myself to do the bare minimum in acknowledging Apple is giving users empowering new accessibility features too.
I’ve written about this phenomenon before. I don’t mean to be gatekeeper-y about accessibility, but the disability community deserves better than one-day-a-year shoutouts on blogs and podcasts, rife with “gee whiz, that’s great” platitudes. We deserve more recognition than what GAAD co-founder Joe Devon recently told me is our “364 days of global accessibility oblivion.” There’s zero reason, for example, an accessibility-focused iPhone review couldn’t run alongside the mainstream takes of journalists like my close friend and peer in Joanna Stern at The Wall Street Journal. Disabled people read the Journal and use iPhones too, so it makes perfect sense in my mind. It seemingly isn’t so to the people tasked with overseeing the WSJ’s tech desk.
(It’s me. I’m a disabled person who uses an iPhone and reads The Wall Street Journal.)
If nothing else, the handwringing over Liquid Glass and its current inaccessibility should prove enlightening to people if only because it shows that everyone can, and inevitably will, benefit from greater accessibility. As for Apple’s role vis-a-vis Liquid Glass, I will reiterate what I wrote last week by again saying Sarah Herrlinger, the company’s senior director of global accessibility policy and initiatives, indicated Liquid Glass was created to be accessible as possible and is simpatico with features such as Reduce Transparency. To suggest Apple has cratered its reputation on accessibility in the first developer beta of iOS 26 is categorically untrue and lacking common sense. My understanding has long been accessibility is on par with readying the new iPhones to ship as far as internal importance. The company’s efforts in accessibility is neither extraneous nor a lark; it’s a highly serious endeavor. Apple isn’t perfect in accessibility, of course, but to presume they purposely ignored accessibility in making Liquid Glass is to show a gross misunderstanding of a huge part of how the company thinks and works.
Let this be a lesson to journalists and users alike: accessibility fucking matters.
Apple Honors Speechify, Art of Fauna for accessibility at this year’s Apple Design Awards
At WWDC 2025 this week, Apple recognized this year’s Apple Design Award (ADA) winners. The Cupertino-based company celebrated a dozen apps and games for the honor, with one app and one game winning in six categories: Delight and Fun, Innovation, Interaction, Inclusivity, Social Impact, and Visuals and Graphics.
Pertinent to accessibility is, of course, the Inclusivity category. The winners here are Speechify and Art of Fauna. Speechify transforms written text in audio, with support of hundreds of voices and over 50 languages. The app also features robust support for iOS accessibility stalwarts in Dynamic Type and VoiceOver, and is an indispensable tool for anyone who copes with conditions such as ADHD, dyslexia, and/or low vision. Regarding Art of Fauna, made by Austria-based Klemens Strasser, the game is a conservation-themed puzzle title which incorporates wildlife imagery. The game tasks players with rearranging visual elements or pieces of descriptive text, with the game supporting accessibility by way of rich screen reader and haptic feedback support.
Apple announced the ADA finalists and winners earlier this month. The honorees spanned the globe and were chosen as winners because their work demonstrated “excellence in innovation, ingenuity, and technical achievement,” according to Apple.
“Developers continue to push the boundaries of what’s possible, creating apps and games that are not only beautifully designed but also deeply impactful,” Susan Prescott, Apple’s vice president of worldwide developer relations, said in a statement. “We’re excited to celebrate this incredible group of winners and finalists at WWDC and spotlight the innovation and craftsmanship they bring to each experience.”
On a related note, the physical trophy Apple gives to ADA winners is truly something to behold. The hardware is a dense, substantially weighty cube milled from a single piece of aluminum. I bring it up because, at last year’s WWDC, I got to see (and hold!) an ADA trophy for the first time. It happened during a briefing at Apple Park with the makers of Oko, an iPhone app which leverages artificial intelligence to make street-crossing more accessible to Blind and low vision people. The team had the cube proudly displayed on a coffee table during our discussion, after which they asked if I’d be interested in picking up the cube and feeling it. I’m sure there exists a picture of me somewhere.
The salient point? The ADA trophy has every bit the fit and finish of an iPhone or iPad.
Inside Unlimited Play’s Mission to Make sure ‘No Child is left on the sidelines’ at the playground
Take a gander at the homepage of Unlimited Play’s website and you’ll see the nonprofit organization makes an unequivocal proclamation: “At our playgrounds, no child is left on the sidelines.” Dig deeper and you’ll notice Unlimited Play’s philosophy that “no matter their physical or cognitive abilities, [children deserve] to feel welcome and experience the joy of play without barriers.” Make no mistake, adaptive playgrounds are shining examples of assistive technologies existing, quite literally, in the real world.
Unlimited Play’s origin story begins with a 3-year-old boy named Zachary Blakemore. He lived with a rare genetic condition known as Pelizaeus-Merzbacher disease, or PMD, which required him to use a wheelchair. Blakemore, like any other kid, relished visiting the playground—the problem, however, was it was unbuilt for his needs. It was inaccessible. These experiences pushed Blakemore’s mom, Natalie Mackey, to start Unlimited Play in 2003. Four years and $750,000 later, the first inclusive space, aptly named Zachary’s Playground, opened in Lake Saint Louis, Missouri on April 21, 2007.
As Mackey told me recently, she never imagined Unlimited Play becoming a thing.
“We had reporters there that day [of Zachary Playground’s opening] and they asked me if I would continue the work into the future. I said, ‘Not a chance.’ It was four years of fundraising and I was excited to let other people follow what we had done,” she said. “But we’re talking today because families like mine and cities more and more have been calling to ask we help and continue the journey of providing inclusive play for everyone. We have over 100 projects from east coast to west coast, even into Canada. Every day, we’re contacted by families looking for ways to have their children play.”
It’s been over two decades now Mackey has served as the nonprofit’s founder and chief executive officer. She explained she’s worked with "so many families” while noting she and her team are “continually learning” along the way. Mackey shared an anecdote on a current project involving a boy named Teddy, who has dwarfism. Typical playgrounds, she told me, are inaccessible to him, adding Unlimited Play’s experience of working on his project has proven enlightening. “[It’s] taught us things about that population and how to better design playgrounds for individuals with dwarfism,” Mackey said.
When asked what exactly constitutes an adaptive and inclusive playground, Mackey said there are several elements. A lot of the design work is topographical, what with using soft turf on the ground, as well as ensuring wheelchair access by way of ramps and sidewalks. Similarly, communication boards are crucial for interaction, and the use of large type and bold colors can be beneficial for those with low vision. It’s quite the challenge, Mackey went on to tell me, to incorporate all these things while also bearing in mind typically developing children likely want to play along too. Everybody, she added, has their own needs and tolerances regardless of one’s ability level(s).
Of course, Unlimited Play constructing playgrounds isn’t like Apple constructing Apple Park. Unlike Apple, Mackey and Unlimited Play has not unlimited funds with which to procure everything needed to build these playgrounds. They need help, which Mackey and team found in the folks at Little Tikes Commercial. The subsidiary, a scaled-up offshoot of the children’s products company, manufacturers playground equipment—including inclusive ones like Zachary’s Playground. According to Mackey, Unlimited Play’s partnership with Little Tikes Commercial means her organization gains the ability to train “about 120 representatives” on the importance of inclusive play and, more pointedly, the ABCs of building accessible play-based environments. After said training, Mackey told me the trainees go out and effectively act as ambassadors to Unlimited Play; this enables the nonprofit to have “many more feet on the ground” doing this evangelism. Mackey’s role involves reporting to the Little Tikes team on the feedback she’s heard from communities “then work together to create new products,” she said.
To work with a well-known entity as Little Tikes is game-changing for Unlimited Play.
“I never imagined that any big company would care,” Mackey said. “Especially when my son was little, life felt very isolating… like nobody understood what I was going through and what I was desperately wanting to provide for my child. For Little Tikes Commercial to say, ‘We see you, we understand, and we want to be part of this’ was really exciting and meaningful for me. For a small nonprofit, it helps give credibility to our mission to say we have such a successful corporation believe in us and backing Unlimited Play.”
Looking towards the future, Mackey was succinct in sharing she hopes to continue pushing forward in fulfilling her organization’s mission. Ideally, she hopes playground standards are raised on a national level so as to reflect the needs of disabled children and their families. Mackey hopes this work raises enough awareness that it doesn’t always take a small-time nonprofit like Unlimited Play to do all the heavy lifting. More parks and recreation staffers should know kids like Zachary deserve open, welcoming, and accommodating spaces in which to play—the problem is most don’t recognize the barriers present in the majority of neighborhood playgrounds across America today.
Zachary died in September 2021, but his legacy lives on in his mom and her mission.
“I would say it’s important to tell people we must be creating environments where everyone can thrive at their very best,” Mackey said when asked to distill Unlimited Play’s raison d’être. “We grow as communities and grow as a nation. My son had a hard time doing typical things, but what he brought was so much. He brought creating environments where we all feel like we belong [and] we can all be our best… we only, as communities and nations, get better by doing that. Play, I truly believe, is the one language we all speak. It’s the international language we all spoke to begin with and how we learn from each other. We should care about creating environments that make that language possible for us to understand, grow, and develop even better.”
New ‘F1’ Trailer makes Movies More Accessible
Apple’s Greg Joswiak took to X earlier today to post about the new trailer Apple released for its upcoming Apple TV+ film F1. The executive, the company’s senior vice president of worldwide marketing, describes what Apple calls a “haptic trailer” as “the coolest trailer ever”—and his boastfulness is deserved. I just watched the 2:10 preview, available in the TV app on iPhone and it is incredibly cool and, dare I say, innovative.
From an accessibility perspective, applying haptic feedback to movie trailers is a genius-level move. It brings a level of access and immersion that transcends sheer coolness, as Joswiak said. To wit, someone who’s Deaf or hard-of-hearing, or Blind or low vision, could literally feel the rumble of the cars as they race around the speedway. Likewise, even for someone with typical hearing and vision, the added sensory input makes for a richer experience because, once again, a person is able to have a tactile approximation of the cars’ horsepower. When I first read about the F1 trailer, my mind immediately recalled the similar-styled Music Haptics accessibility feature Apple introduced last year in iOS 18. I wrote about Music Haptics again recently, and the advent of Apple’s novel “haptic trailer” stands not only as yet another example of potential accessibility, but of the company’s vaunted vertical integration as well.
I hope the F1 creative team appreciates how cool this technology is. Especially for people with limited hearing, if any at all, that haptics are present in the trailer makes it such that the cars’ power can be felt if not heard. This is exactly the value proposition for Music Haptics; a Deaf person, like Troy Kotsur’s Frank Rossi in CODA, can enjoy music because, as he says in the film, it “makes my ass shake.” It isn’t merely a funny, throwaway line: haptic feedback makes ostensibly exclusionary arts like music more inclusive to the Deaf and hard-of-hearing community. I don’t know the nitty-gritty technical details on the F1 trailer, but it’d be great to see Apple someday release a public API for App Store developers to hook into their own app(s). Imagine, for instance, sitting down to watch Star Wars on Disney+ and being able to feel explosions or the rumble of lightsabers. Haptics makes the content not only more immersive—it’s accessible too.
The Brad Pitt-led F1 opens in theaters on June 27. It’ll stream on TV+ later this year.
It’s Official: Yours Truly Needs a new Mac Soon
Amidst this week’s hubbub regarding Liquid Glass and more, one tidbit of news has me feeling melancholy. The next version of macOS, named Tahoe, is the last edition to support Intel-based Macs. This is important news because, as I suspected, I’ll need to upgrade my Mac pretty soon. My trusty Retina 4K iMac from 2019, the 21.5-inch model, is sated to stay on Sequoia forevermore. You’ve served me with great honor, old friend.
According to Apple, macOS Tahoe is “the last release for Intel-based Mac computers,” adding the machines will get three years of security updates. The news coincides with the company’s plans to sunset its Rosetta 2 translation system, introduced in 2020 to help developers with transitioning their apps from Intel to Apple silicon. “Rosetta was designed to make the transition to Apple silicon easier, and we plan to make it available for the next two major macOS releases—through macOS 27—as a general-purpose tool for Intel apps to help developers complete the migration of their apps,” Apple says. “Beyond this timeframe, we will keep a subset of Rosetta functionality aimed at supporting older unmaintained gaming titles, that rely on Intel-based frameworks.”
As with my beta-testing strategy for this summer, I’m still contemplating which Mac should supplant my 4K iMac. I have lots of options; I love the all-in-one design of the iMac and its anchoring of my workspace, but am also enamored by a more modular setup headlined by something like the big and bright Pro Display XDR. On the mobile side, I’m tempted to turn the 13-inch M4 iPad Pro I got for my birthday last September into my “laptop” paired with the Magic Keyboard. I’m really excited for the forthcoming (and arguably overdue) Mac-like features coming to iPadOS 26. In terms of my main work setting, macOS, the jump to using an Apple silicon Mac full-time is tantalizing because of Apple silicon-exclusive features such as iPhone Mirroring, as well as a larger display, and other quality-of-life improvements like faster processing and stuff.
My birthday is coming up in September, so perhaps my present will be a new computer.
Apple’s ‘Liquid Glass’ won’t make the sky fall
I was on the ground yesterday at Apple Park covering this year’s WWDC keynote. The star of the show was Apple’s introduction of its all-new Liquid Glass design language. The company has a great session on Liquid Glass in the Developer app. I highly suggest watching it (and other sessions) on a big screen television through an Apple TV 4K; I did so last night on my 77-inch LG C3 OLED last night and it was a blast, if terminally nerdy.
I’m still trying to devise a plan of action for testing iOS 26, et al, throughout the summer. In full transparency, I’m one of those rare birds in tech media circles who neither run the developer betas nor the public betas of the new operating systems. I’m a terrible tester, although I usually do jump to the iOS public beta on my iPhone late in the summer closer to its official release come September. With the advent of Liquid Glass, however, I feel it’s in my best journalistic interest to prioritize testing at least one beta. As I told some Apple employees following the presentation, Liquid Glass truly is a de-facto accessibility feature unto itself. While it’s undoubtedly true Apple’s stated goals of creating cohesiveness and harmony are important to accessibility, the reality is what really matters is how Liquid Glass performs in a practical sense. Practicality entails legibility, contrast, and motion. For people with low vision—and people with 20/20 vision, for that matter—the choices Apple has made with Liquid Glass, the proverbial proof in the pudding lies in usability. All the flowery, romanticized marketing bluster regarding harmony means zippo if Liquid Glass isn’t readable. Personally, I find Liquid Glass to look damn cool and quite beautiful; nonetheless, I’m predisposed to be skeptical as a lifelong disabled person and thus was alarmed by some of what I saw in Apple’s marketing video. Fortunately, my fears were quickly allayed by high confidence in Apple’s track record in accessibility and confirmation that changes will be coming.
I connected with Sarah Herrlinger, Apple’s senior director of global accessibility policy and initiatives, for a few minutes after the keynote ended. It was decidedly not an official, on the record interview, but I can confidently report Herrlinger told me her teams worked in tandem with the design team to build Liquid Glass and make it as accessible as possible. To that end, she noted Liquid Glass works with features such as Reduce Transparency, amongst others, in increasing legibility. I’m sure I’ll have more to report in the coming weeks and months. For now, I’m willing to take Herrlinger at her word, along with the reporter’s grain of salt, that Liquid Glass is accessible unto itself.
As I write this, it’s been roughly 24 hours since Apple introduced Liquid Glass. In that time, the timelines across my myriad social media services have been insufferable. There are so many insipidly bad takes on Liquid Glass from wannabe Apple designers who are posting hot takes to feed into social media’s worst impulses. There’s absolutely room for constructive criticism—👋🏼, journalist here—but there’s also room for common sense. Apple released the first beta of its new platforms yesterday. There is a whole summer yet for the company to tweak and refine Liquid Glass. Of course Apple engineers must reach a degree of “doneness” when readying the beta builds, but they’re betas for a reason: they’re essentially unfinished. The software will evolve before being publicly released later this year. Back in 2013, iOS 7 similarly overshot on the usability vector before dialing back to the mean before its final release alongside the iPhone 5C and 5S in the fall. There’s no doubt in my mind iOS 26 and Liquid Glass will walk the very same path in 2025, so crying Chicken Little seems utterly pointless.
Finally, a cool little personal postscript to Monday’s announcement. My first WWDC was in 2013, the year iOS was first redesigned. 13 years later, I was literally at Apple Park watching iOS 26 getting its Liquid Glass redesign. As a friend said to me after the presentation ended, 13 plus 13 equals 26. All told, I think it’s nice symmetry all around.