Researchers Find Mind-Reading Artificial Intelligence Turns Thoughts Into Speech

Artificial Intelligence News

Researchers_Find-Mind-Reading-Artificial-Intelligence-Turns-Thoughts-Into-Speech Researchers Find Mind-Reading Artificial Intelligence Turns Thoughts Into SpeechIn today’s digital world where innovations straight out of the world of science fiction, a team of researchers have implemented Artificial Intelligence to turn brain signals into computer-generated speech. The deed was achieved with the five epilepsy patients’ assistance, where all had been outfitted with different sorts of brain electrodes as part of their seizure treatment. That enabled researchers to perform very sensitive brain monitoring, electrocorticography.

After that, researchers noted that the outcome presented a major leap toward the goal of brain-to-computer communication. The study author Nima Mesgarani, an associate professor with Columbia University’s Zuckerman Mind Brain Behavior Institute, in New York City, stated that previous efforts in this direction concentrated on simple computer models that were able to generate audio that sounded similar to the original speech, but not comprehensible in any way. However, the study on this utilized advanced AI to reform sounds from the brain that were much more comprehensible compared to prior one. Mesgarani added that this is a huge milestone for his team, and they weren’t sure that they could do it. Throughout the research process, brain activity was tracked while each participant listened to short stories and number lists, as read to them by four different speakers. The signal patterns from the brain that was recorded while the patients listened to the numbers, then fed into a computer algorithm blindly that means without any sign of which pattern matched which number.

Mesgarani said that his algorithm is the first to produce a sound that is actually intelligible to human listeners. And longstanding efforts to correctly decode the brain are finally coming to completion. Such breakthroughs also mean better brain-computer interfacing that would open up whole new platforms for man-machine communication. In addition, He pointed out that is the ultimate goal is to advance technologies that can decode the internal voice of a patient who can’t able to speak. In that regard, future experiments will focus on more complex words and sentence structure.