The goal of Sight Tech Global, a virtual, global event on December 2-3, 2020, is to gather the world’s top experts who are applying advanced technologies, notably AI, to the future of accessibility and assistive tech for people who are blind or visually impaired.
Today we’re excited to roll out most of the agenda. There are another half-dozen sessions and breakouts still to come, notably sessions on AI bias and civil rights. What we’ve discovered over the many weeks of research and conversation is a consistent, strong interest on the part of researchers, technologists and product and design thinkers to convene and talk over the future — its promises, challenges and even threats.
We’re delighted to have top-level talent from virtually every leading technology company, many research universities and some startups ready for fireside chats and small panel discussions with expert moderators. Some sessions will take questions from our audience as well.
When the event dates are closer, we will add dates and times to each of these sessions as well as announce additional speakers. Register today to get a free pass and please browse the first edition of the Sight Tech Global agenda below.
Seeing AI: Where does Microsoft’s blockbuster app go from here?
With ever more powerful computer and data resources available in the cloud, Microsoft’s Seeing AI mobile app is destined to become a steadily better ally for anyone with vision challenges. Co-founder Saqib Shaikh leads the engineering team that’s charting the app’s cloud-enabled future.
Saqib Shaikh, co-founder of Seeing AI, Microsoft
Moderator: Devin Coldewey, TechCrunch
The future according to OrCam
As AI-based computer vision, voice recognition and natural language processing race ahead, the engineering challenge is to design devices that can perceive the physical world and communicate that information in a timely manner. Amnon Shashua’s OrCam MyEye is the most sophisticated effort yet to merge those technologies in a seamless experience on a dedicated device.
Amnon Shashua, co-founder of OrCam and Mobileye
Moderator: Matthew Panzarino, TechCrunch
Accessibility from the wheels up: The Waymo fully autonomous taxi
If people who are blind or visually impaired find Uber and Lyft liberating, imagine how they will feel summoning a self-driving ride from an app on their mobile phones. But wait, how exactly will they locate the cars and what happens when they climb in? Presenter Clem Wright is responsible for the self-driving taxi’s accessibility, and he will be joined by leadership from two organizations closely involved in that effort: The Lighthouse for the Blind SF and the Foundation for Blind Children.
Our AI future is already here
Whether it’s Alexa, Tesla or Facebook, AI is already deeply embedded in our daily lives. Few understand that better than Dr. Kai-Fu Lee, a scientist who developed the first speaker-independent, continuous speech recognition system as a Ph.D. student at Carnegie Mellon, led Google in China and held senior roles at Microsoft and Apple. Today, Dr. Lee runs Sinovation Ventures, a $2 billion fund based in China, is president of the Sinovation’s Artificial Intelligence Institute and has 50 million followers on social media.
Dr. Kai-Fu Lee, chairman and CEO, Sinovation Ventures
Moderator: Ned Desmond, Sight Tech Global
The future of AT devices and the companies that make them
Dedicated devices versus accessible platforms? Victor Reader Stream versus iPhones and Alexa? How will AT companies take advantage of a world with cloud data and edge computational power, AI algorithms and more demanding customers than ever? Humanware, eSight and APH are already looking far into that future.
If the Jetsons had screen readers, would they be using keyboard commands?
The screen reader is arguably the most consequential digital technology ever for people who are blind or visually impaired. At the same time, screen readers depend on a dizzying array of keyboard commands, and — when it comes to reading websites in a browser — they struggle with the ugly reality of poor website accessibility. New technologies may lead the way to better outcomes.
Glen Gordon, Software fellow, Vispero; architect, JAWS
James Teh, Accessibility engineer, Mozilla; co-founder, NVDA
Léonie Watson, director, TetraLogical
Moderator: Matt King, Accessibility technical program manager, Facebook
Alexa, what is your future?
When Alexa launched six years ago, no one imagined that the voice assistant would reach into millions of daily lives and become a huge convenience for people who are blind or visually impaired. This fall, Alexa introduced personalization and conversational capabilities that are a step-change toward more human-like home companionship. Amazon’s Josh Miele and Anne Toth will discuss the impact on accessibility as Alexa becomes more capable.
Augmented reality and perception: What’s the best way to get the message across?
It’s one thing for an AI-based system to “know” when it’s time to turn left, who came through the door or how far away the couch is: It’s quite another to convey that information in a timely fashion with minimal distraction. Researchers are making use of haptics, visual augmented reality (AR), sound and language to figure out the right solutions.
Amos Miller, Product strategist, Microsoft AI and Research
Ashley Tuan, VP Medical Devices, Mojo Vision
Sile O’Modhrain, associate professor, Performing Arts Technology, University of Michigan
Moderator: Nick Giudice, professor of Spatial Informatics, University of Maine
Wayfinding: Finding the mark
Map apps on mobile phones are miraculous tools accessible via voice output, but mainstream apps don’t announce the detailed location information (which people who are blind or visually impaired really want), especially inside buildings and in public transportation settings. Efforts in the U.S. and U.K. are improving accessible navigation.
Computer vision, AI and accessibility: What’s missing from this picture?
For an AI to interpret the visual world on behalf of people who are blind or visually impaired, the AI needs to know what it’s looking at, and no less important, that it’s looking at the right thing. Mainstream computer vision databases don’t do that well — yet.
Danna Gurari, assistant professor and director of the Image and Video Computing Group, University of Texas
Cecily Morrison, principal researcher human experience & design, Microsoft
Patrick Clary, product manager, AI and accessibility, Google
Moderator: Roberto Manduchi, professor CS and Engineering, UC Santa Cruz
Keep an out for more sessions and breakouts later this month. In the meantime, registration is open. Get your pass today!
Sight Tech Global is eager to hear from potential sponsors. We’re grateful to current sponsors Amazon, Ford, Google, Microsoft, Mojo Vision, Waymo, Wells Fargo and Humanware. All sponsorship revenues go to the nonprofit Vista Center for the Blind and Visually Impaired, which has been serving the Silicon Valley area for 75 years.
Special thanks to the Sight Tech Global advisors — Tech Matters’ Jim Fruchterman, UC Santa Cruz’s Roberto Manduchi, Verizon Media’s Larry Goldberg, Facebook’s Matt King and Be My Eyes’ Will Butler — who are playing an invaluable role on this project.