UC Berkeley and UC San Francisco researchers have successfully restored speech to Ann Johnson, a woman who lost the ability to speak after a brainstem stroke 18 years ago, using an AI-powered brain-computer interface. The breakthrough technology translates brain activity into speech in real-time, offering hope for people with locked-in syndrome and potentially transforming accessibility in the workforce and beyond.
What you should know: Johnson suffered a brainstem stroke at age 30 in 2005 that left her with locked-in syndrome—a rare condition causing near-complete paralysis and loss of speech while leaving cognitive abilities intact.
- She joined the clinical trial in 2022 as the third participant in a study that began in 2020.
- The technology uses a neural implant that reads brain activity from the speech processing region and translates it into audio and visual communication through an AI model.
- Johnson can currently communicate at about 14 words per minute using eye-tracking technology, compared to normal conversational speech of 160 words per minute.
How it works: The neuroprosthesis reads brain signals when someone attempts to speak and translates them into audio and avatar movements.
- An implant rests on the brain region that processes speech, acting as a “thought decoder” that captures neural signals when Johnson tries to speak.
- The AI model translates these brain signals into text, audio, or facial animations, similar to how Siri converts voice to text.
- Researchers used a recording of Johnson’s wedding speech to recreate her voice and allowed her to choose from different avatars for visual representation.
- The system only activates when someone makes a deliberate effort to speak—it cannot read random thoughts.
Major breakthrough: The team dramatically reduced response delays from eight seconds to just one second using new streaming architecture.
- The original 2023 system required users to attempt entire sentences before translation, using sequence-to-sequence architecture.
- The March 2024 update introduced streaming architecture that translates brain activity to sound in real-time with only about a one-second delay.
- This advancement was published in Nature Neuroscience and represents a significant step toward natural conversation flow.
What they’re saying: The emotional impact of hearing her own voice again was profound for Johnson and the research team.
- “What do you think of my artificial voice?” Johnson asked during the trial. “Tell me about yourself. I am doing well today.”
- “We didn’t want to read her mind,” said Gopala Anumanchipalli, UC Berkeley assistant professor of electrical engineering and computer sciences. “We really wanted to give her the agency to do this.”
- Johnson wrote to researchers: “I want patients there to see me and to know their lives are not over now. I want to show them that disabilities don’t need to stop us or slow us down.”
Future vision: Researchers envision plug-and-play neuroprostheses becoming standard medical care within years.
- “We need to be able to have neuroprostheses be plug-and-play, so that it becomes a standard of care and not a research experiment,” Anumanchipalli explained.
- Future developments could include wireless implants, 3D photorealistic avatars, and digital clones with personalized preferences.
- Johnson hopes to become a counselor in physical rehabilitation facilities, using the technology to communicate with clients.
The bigger picture: While the target population is relatively small, researchers emphasize these patients are among the most vulnerable in terms of quality of life.
- The technology has “enormous potential to make the workforce and the world more accessible to people like Johnson,” according to the research team.
- Johnson had her implant removed in February 2024 for unrelated reasons but continues collaborating with researchers on future improvements.
- The breakthrough represents a significant advance in restoring human communication abilities through direct brain-computer interfaces.
Recent Stories
DOE fusion roadmap targets 2030s commercial deployment as AI drives $9B investment
The Department of Energy has released a new roadmap targeting commercial-scale fusion power deployment by the mid-2030s, though the plan lacks specific funding commitments and relies on scientific breakthroughs that have eluded researchers for decades. The strategy emphasizes public-private partnerships and positions AI as both a research tool and motivation for developing fusion energy to meet data centers' growing electricity demands. The big picture: The DOE's roadmap aims to "deliver the public infrastructure that supports the fusion private sector scale up in the 2030s," but acknowledges it cannot commit to specific funding levels and remains subject to Congressional appropriations. Why...
Oct 17, 2025Tying it all together: Credo’s purple cables power the $4B AI data center boom
Credo, a Silicon Valley semiconductor company specializing in data center cables and chips, has seen its stock price more than double this year to $143.61, following a 245% surge in 2024. The company's signature purple cables, which cost between $300-$500 each, have become essential infrastructure for AI data centers, positioning Credo to capitalize on the trillion-dollar AI infrastructure expansion as hyperscalers like Amazon, Microsoft, and Elon Musk's xAI rapidly build out massive computing facilities. What you should know: Credo's active electrical cables (AECs) are becoming indispensable for connecting the massive GPU clusters required for AI training and inference. The company...
Oct 17, 2025Vatican launches Latin American AI network for human development
The Vatican hosted a two-day conference bringing together 50 global experts to explore how artificial intelligence can advance peace, social justice, and human development. The event launched the Latin American AI Network for Integral Human Development and established principles for ethical AI governance that prioritize human dignity over technological advancement. What you should know: The Pontifical Academy of Social Sciences, the Vatican's research body for social issues, organized the "Digital Rerum Novarum" conference on October 16-17, combining academic research with practical AI applications. Participants included leading experts from MIT, Microsoft, Columbia University, the UN, and major European institutions. The conference...