in ,

IIT Madras Researchers Use AI To Translate Brain Signals Of Speech-Impaired Persons Into English

IIT Madras Researchers Use AI To Translate Brain Signals Of Speech-Impaired Persons Into English

Indian Institute of Technology, Madras researchers have developed an artificial intelligence technology to convert brain signals of speech impaired humans into language. This ground-breaking research has opened up the field for interpreting nature’s other signals like photosynthesis or their response to external forces.

Electrical signals, brain signal or any signal, in general, are waveforms which are decoded to meaningful information using physical law or mathematical transforms such as Fourier Transform or Laplace transform. These physical laws and mathematical transforms are science-based languages discovered by renowned scientists such as Sir Isaac Newton and Jean-Baptiste Joseph Fourier.

A team of researchers led by Dr Vishal Nandigana, Assistant Professor, Fluid Systems Laboratory, Department of Mechanical Engineering at IIT Madras, is working on this area of research.

Elaborating on this Research, Dr Nandigana said, “The output result is the ionic current, which represents the flow of ions which are charged particles. These electrically driven ionic current signals are worked on to be interpreted as human language meaning speech. This would tell us what the ions are trying to communicate with us. When we succeed with this effort, we will get electrophysiological data from the neurologists to get brain signals of speech impaired humans to know what they are trying to communicate.”

Further, Dr Nandigana added, “The other major application of this field of research we see potentially is, can we interpret nature’s signals, like plant photosynthesis process or their response to external forces, mean when we collect their real data signal. The data signal also, we believe, is going to be in some wave-like pattern with spikes, humps and crusts. So the big breakthrough will be can we interpret what plants and nature is trying to communicate to us. This will help in predicting monsoons, earthquake, floods, Tsunami and other natural disasters using our Artificial Intelligence and Deep Learning algorithms. If we understand the nature signals better we can take care of it well is our objective that we want to pitch in from our laboratory.”

IIT Madras Researchers are working on how these real data signal can be decoded into human language like English and if the real data signal can be interpreted as a simple human language that all human beings can understand.

Brain signals are typically electrical signals. These are wave-like patterns with spikes, humps and crusts which can be converted into simple human language meaning speech using Artificial Intelligence and Deep Learning algorithms. This enabled the Researchers to read direct electrical signals of the brain.

They tested this concept by getting experimental electrical signals through experiments in the laboratory to get signals from nanofluidic transport inside nanopores. The nanopores were filled with saline solution and mediated using an electric field.

Source: analyticsindiamag.com

What do you think?

40 points
Upvote Downvote

Leave a Reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Artificial Intelligence Influencers To Follow in 2020 (Updated List)

Elon Musk tweeted he'll be holding a "super fun AI party/hackathon" at his house next month.

Elon Musk tweeted he’ll be holding a “super fun AI party/hackathon” at his house next month.