Archives: Sessions
Session custom post type.
Envision: What happens when smart glasses really get smart?
Envision is a pioneer in the effort to connect computer vision with everyday life in the form of tech-enabled glasses that can tell a blind user about their surroundings. Using the Google glass platform, Envision found a market with blind users who value a hands-free interaction, and the experience only got better with the launch of scene description AIs in the past two years. But what’s really changed the game for Envision is generative AI, and the tantalizing possibility of a multimodal AI that’s more like an all-around personal assistant.
Immediately following this session, Karthik Mahadevan will be available to take questions live in a breakout session.
Can we enlist AI to accelerate human-led work in alt text and audio description?
To watch the recently released “All The Light We Cannot See” with audio descriptions “on” is a revelation, at least for a sighted person. The audio description uses words sparingly to augment the obvious soundscape and to call out subtle details anyone might easily miss. It’s art only a human team could produce (sorry AI proponents), but then it’s also expensive and time consuming. In that regard, producing alt text for images online or audio descriptions for video face the same challenge: how to do more and do it well. At Scribely and MAX, the human-first approach is uppermost, but they are also exploring how AI and related tech can be narrowly channeled to speed up their vital work.
Salesforce: The Office of Accessibility – four years on
Nearly four years ago, Salesforce stepped ahead in the tech and corporate world by announcing the formation of an Office of Accessibility, charged with pulling together all the strands of accessibility across the CRM giant, including workforce development, product development and design, and customer relations. Sight Tech Global touched base with the fledgling effort in 2020 and in this session we’ll hear what the accessibility team has learned after three years work to ensure every aspect of Salesforce embraces accessibility.
dot 2.0
With new partnerships to lean on and generative AI looking to make their dynamic tactile display more useful than ever, the founders of the DOT pad return to Sight Tech Global to discuss how their vision for the breakthrough device is fast evolving, including a new, faster Dot Pad that can refresh in the blink of an eye.
Why did the AI cross the street? The OKO app story
Did that mechanical voice just say it’s safe to cross the street? It’s a dilemma every blind person faces when they are about to step off the curb. What if the camera on the back of your mobile phone could assess the signals and your path to make a crossing safer? A small team of AI engineers at the startup AYES took on that challenge and created the OKO app, which uses computer-vision-based AI to “read” the signals and suggest when it’s safe to cross. How does the app work and just how safe is it?
Seeing AI meets Generative AI – The View from Microsoft
Microsoft’s Saqib Shaikh co-founded the Seeing AI app, which was one of the first deployments of AI computer vision in an app to help blind users “narrate the world around you” by using AI to describe nearby people, text and objects. Shaikh’s employer, Microsoft, is a leading investor in OpenAI, the organization that created the ground-breaking chatGPT, a type of AI called “generative” because of its ability to “generate” novel content in response to questions. Today, Seeing AI can tell you there is a chair in the room because it is “trained” to identify a chair. With chatGPT, Seeing AI might be able to answer a random question it was not specifically trained for, such as,”Is there a cat in the room?” The answers chatGPT provides can be wondrous or wildly off base “hallucinations,” in the lingo of AI specialists. Can generative AI’s quirky nature be tamed for accessibility?
Welcome to Day Two of Sight Tech Global
See you tomorrow!
The Screen Reader Interoperability Gap – A Hidden Systemic Barrier
At Sight Tech Global two years ago, we unveiled the ARIA AT initiative, which aimed to address the frustrating, damaging reality that screen readers are not interoperable on the Web, unlike their cousins for sighted users – browsers like Chrome and Safari. In other words,any developer that takes accessibility seriously has to test their code on JAWS, VoiceOver, NVDA and the rest. In this session, the people advancing the ARIA AT project are back with a refresher, progress to report, and a call to action.
