{"id":12549,"date":"2024-09-03T10:00:00","date_gmt":"2024-09-03T10:00:00","guid":{"rendered":"https:\/\/modernsciences.org\/staging\/4414\/?p=12549"},"modified":"2024-08-23T06:16:03","modified_gmt":"2024-08-23T06:16:03","slug":"from-thoughts-to-words-how-ai-deciphers-neural-signals-to-help-a-man-with-als-speak","status":"publish","type":"post","link":"https:\/\/modernsciences.org\/staging\/4414\/from-thoughts-to-words-how-ai-deciphers-neural-signals-to-help-a-man-with-als-speak\/","title":{"rendered":"From thoughts to words: How AI deciphers neural signals to help a man with ALS speak"},"content":{"rendered":"\n<div class=\"theconversation-article-body\">\n    <figure>\n      <img  decoding=\"async\"  src=\"data:image\/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAEAAAABAQMAAAAl21bKAAAAA1BMVEUAAP+KeNJXAAAAAXRSTlMAQObYZgAAAAlwSFlzAAAOxAAADsQBlSsOGwAAAApJREFUCNdjYAAAAAIAAeIhvDMAAAAASUVORK5CYII=\"  class=\" pk-lazyload\"  data-pk-sizes=\"auto\"  data-pk-src=\"https:\/\/images.theconversation.com\/files\/614950\/original\/file-20240821-16-x00qm4.jpg?ixlib=rb-4.1.0&#038;rect=0%2C700%2C3368%2C2317&#038;q=45&#038;auto=format&#038;w=754&#038;fit=clip\" >\n        <figcaption>\n          Casey Harrell, who has ALS, works with a brain-computer interface to turn his thoughts into words.\n          <span class=\"attribution\"><span class=\"source\">Nicholas Card<\/span><\/span>\n        <\/figcaption>\n    <\/figure>\n\n  <span><a href=\"https:\/\/theconversation.com\/profiles\/nicholas-card-1635329\" target=\"_blank\" rel=\"noopener\">Nicholas Card<\/a>, <em><a href=\"https:\/\/theconversation.com\/institutions\/university-of-california-davis-1312\" target=\"_blank\" rel=\"noopener\">University of California, Davis<\/a><\/em><\/span>\n\n  <p>Brain-computer interfaces are a groundbreaking technology that can help paralyzed people regain functions they\u2019ve lost, like moving a hand. These devices record signals from the brain and decipher the user\u2019s intended action, bypassing damaged or degraded nerves that would normally transmit those brain signals to control muscles. <\/p>\n\n<p><a href=\"https:\/\/doi.org\/10.1038\/nature04970\" target=\"_blank\" rel=\"noopener\">Since 2006<\/a>, demonstrations of brain-computer interfaces in humans have primarily focused on restoring arm and hand movements by enabling people to <a href=\"https:\/\/doi.org\/10.1088\/1741-2560\/8\/2\/025027\" target=\"_blank\" rel=\"noopener\">control computer cursors<\/a> or <a href=\"https:\/\/www.nbcnews.com\/video\/watch-obama-fist-bump-a-robotic-arm-powered-by-a-brain-chip-785470531610\" target=\"_blank\" rel=\"noopener\">robotic arms<\/a>. Recently, researchers have begun developing <a href=\"https:\/\/doi.org\/10.1038\/s41583-024-00819-9\" target=\"_blank\" rel=\"noopener\">speech brain-computer interfaces<\/a> to restore communication for people who cannot speak. <\/p>\n\n<p>As the user attempts to talk, these brain-computer interfaces record the person\u2019s unique brain signals associated with attempted muscle movements for speaking and then translate them into words. These words can then be displayed as text on a screen or spoken aloud using text-to-speech software.<\/p>\n\n<p>I\u2019m a <a href=\"https:\/\/scholar.google.com\/citations?hl=en&amp;user=Dwcp1iIAAAAJ&amp;view_op=list_works&amp;sortby=pubdate\" target=\"_blank\" rel=\"noopener\">reseacher<\/a> in the <a href=\"https:\/\/neuroprosthetics.science\/\" target=\"_blank\" rel=\"noopener\">Neuroprosthetics Lab<\/a> at the University of California, Davis, which is part of the <a href=\"https:\/\/www.braingate.org\/\" target=\"_blank\" rel=\"noopener\">BrainGate2<\/a> clinical trial. My colleagues and I recently demonstrated a speech brain-computer interface that <a href=\"https:\/\/doi.org\/10.1056\/NEJMoa2314132\" target=\"_blank\" rel=\"noopener\">deciphers the attempted speech of a man with ALS,<\/a> or amyotrophic lateral sclerosis, also known as Lou Gehrig\u2019s disease. The interface converts neural signals into text with over 97% accuracy. Key to our system is a set of artificial intelligence language models \u2013 artificial neural networks that help interpret natural ones.<\/p>\n\n<h2 id=\"recording-brain-signals\">Recording brain signals<\/h2>\n\n<p>The first step in our speech brain-computer interface is recording brain signals. There are several sources of brain signals, some of which require surgery to record. Surgically implanted recording devices can capture high-quality brain signals because they are placed closer to neurons, resulting in stronger signals with less interference. These neural recording devices include grids of electrodes placed on the brain\u2019s surface or electrodes implanted directly into brain tissue.<\/p>\n\n<p>In our study, we used electrode arrays surgically placed in the speech motor cortex, the part of the brain that controls muscles related to speech, of the participant, Casey Harrell. We recorded neural activity from 256 electrodes as Harrell attempted to speak.<\/p>\n\n<figure class=\"align-center zoomable\">\n            <a href=\"https:\/\/images.theconversation.com\/files\/614924\/original\/file-20240821-16-az47v1.png?ixlib=rb-4.1.0&amp;q=45&amp;auto=format&amp;w=1000&amp;fit=clip\" target=\"_blank\" rel=\"noopener\"><img  decoding=\"async\"  alt=\"A small square device with an array of spikes on the bottom and a bundle of wires on the top\"  src=\"data:image\/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAEAAAABAQMAAAAl21bKAAAAA1BMVEUAAP+KeNJXAAAAAXRSTlMAQObYZgAAAAlwSFlzAAAOxAAADsQBlSsOGwAAAApJREFUCNdjYAAAAAIAAeIhvDMAAAAASUVORK5CYII=\"  class=\" pk-lazyload\"  data-pk-sizes=\"auto\"  data-ls-sizes=\"(min-width: 1466px) 754px, (max-width: 599px) 100vw, (min-width: 600px) 600px, 237px\"  data-pk-src=\"https:\/\/images.theconversation.com\/files\/614924\/original\/file-20240821-16-az47v1.png?ixlib=rb-4.1.0&amp;q=45&amp;auto=format&amp;w=754&amp;fit=clip\"  data-pk-srcset=\"https:\/\/images.theconversation.com\/files\/614924\/original\/file-20240821-16-az47v1.png?ixlib=rb-4.1.0&amp;q=45&amp;auto=format&amp;w=600&amp;h=416&amp;fit=crop&amp;dpr=1 600w, https:\/\/images.theconversation.com\/files\/614924\/original\/file-20240821-16-az47v1.png?ixlib=rb-4.1.0&amp;q=30&amp;auto=format&amp;w=600&amp;h=416&amp;fit=crop&amp;dpr=2 1200w, https:\/\/images.theconversation.com\/files\/614924\/original\/file-20240821-16-az47v1.png?ixlib=rb-4.1.0&amp;q=15&amp;auto=format&amp;w=600&amp;h=416&amp;fit=crop&amp;dpr=3 1800w, https:\/\/images.theconversation.com\/files\/614924\/original\/file-20240821-16-az47v1.png?ixlib=rb-4.1.0&amp;q=45&amp;auto=format&amp;w=754&amp;h=523&amp;fit=crop&amp;dpr=1 754w, https:\/\/images.theconversation.com\/files\/614924\/original\/file-20240821-16-az47v1.png?ixlib=rb-4.1.0&amp;q=30&amp;auto=format&amp;w=754&amp;h=523&amp;fit=crop&amp;dpr=2 1508w, https:\/\/images.theconversation.com\/files\/614924\/original\/file-20240821-16-az47v1.png?ixlib=rb-4.1.0&amp;q=15&amp;auto=format&amp;w=754&amp;h=523&amp;fit=crop&amp;dpr=3 2262w\" ><\/a>\n            <figcaption>\n              <span class=\"caption\">An array of 64 electrodes that embed into brain tissue records neural signals.<\/span>\n              <span class=\"attribution\"><span class=\"source\">UC Davis Health<\/span><\/span>\n            <\/figcaption>\n          <\/figure>\n\n<h2 id=\"decoding-brain-signals\">Decoding brain signals<\/h2>\n\n<p>The next challenge is relating the complex brain signals to the words the user is trying to say.<\/p>\n\n<p>One approach is to map neural activity patterns directly to spoken words. This method requires recording brain signals corresponding to each word multiple times to identify the average relationship between neural activity and specific words. While this strategy works well for small vocabularies, as demonstrated in a <a href=\"https:\/\/doi.org\/10.1056\/NEJMoa2027540\" target=\"_blank\" rel=\"noopener\">2021 study with a 50-word vocabulary<\/a>, it becomes impractical for larger ones. Imagine asking the brain-computer interface user to try to say every word in the dictionary multiple times \u2013 it could take months, and it still wouldn\u2019t work for new words.<\/p>\n\n<p>Instead, we use an alternative strategy: mapping brain signals to phonemes, the basic units of sound that make up words. In English, there are 39 phonemes, including ch, er, oo, pl and sh, that can be combined to form any word. We can measure the neural activity associated with every phoneme multiple times just by asking the participant to read a few sentences aloud. By accurately mapping neural activity to phonemes, we can assemble them into any English word, even ones the system wasn\u2019t explicitly trained with.<\/p>\n\n<p>To map brain signals to phonemes, we use advanced machine learning models. These models are particularly well-suited for this task due to their ability to find patterns in large amounts of complex data that would be impossible for humans to discern. Think of these models as super-smart listeners that can pick out important information from noisy brain signals, much like you might focus on a conversation in a crowded room. Using these models, we were able to decipher phoneme sequences during attempted speech with over 90% accuracy.<\/p>\n\n<figure>\n            <iframe loading=\"lazy\" width=\"440\" height=\"260\" src=\"https:\/\/www.youtube.com\/embed\/fowCcdy3SFo?wmode=transparent&amp;start=0\" frameborder=\"0\" allowfullscreen=\"\"><\/iframe>\n            <figcaption><span class=\"caption\">The brain-computer interface uses a clone of Casey Harrell\u2019s voice to read aloud the text it deciphers from his neural activity.<\/span><\/figcaption>\n          <\/figure>\n\n<h2 id=\"from-phonemes-to-words\">From phonemes to words<\/h2>\n\n<p>Once we have the deciphered phoneme sequences, we need to convert them into words and sentences. This is challenging, especially if the deciphered phoneme sequence isn\u2019t perfectly accurate. To solve this puzzle, we use two complementary types of machine learning language models.<\/p>\n\n<p>The first is n-gram language models, which predict which word is most likely to follow a set of <em>n<\/em> words. We trained a 5-gram, or five-word, language model on <a href=\"https:\/\/openwebtext2.readthedocs.io\/en\/latest\/\" target=\"_blank\" rel=\"noopener\">millions of sentences<\/a> to predict the likelihood of a word based on the previous four words, capturing local context and common phrases. For example, after \u201cI am very good,\u201d it might suggest \u201ctoday\u201d as more likely than \u201cpotato\u201d. Using this model, we convert our phoneme sequences into the 100 most likely word sequences, each with an associated probability.<\/p>\n\n<p>The second is large language models, which power AI chatbots and also predict which words most likely follow others. We use large language models to refine our choices. These models, trained on vast amounts of diverse text, have a broader understanding of language structure and meaning. They help us determine which of our 100 candidate sentences makes the most sense in a wider context.<\/p>\n\n<p>By carefully balancing probabilities from the n-gram model, the large language model and our initial phoneme predictions, we can make a highly educated guess about what the brain-computer interface user is trying to say. This multistep process allows us to handle the uncertainties in phoneme decoding and produce coherent, contextually appropriate sentences.<\/p>\n\n<figure class=\"align-center zoomable\">\n            <a href=\"https:\/\/images.theconversation.com\/files\/614926\/original\/file-20240821-16-1equ7e.png?ixlib=rb-4.1.0&amp;q=45&amp;auto=format&amp;w=1000&amp;fit=clip\" target=\"_blank\" rel=\"noopener\"><img  decoding=\"async\"  alt=\"Diagram showing a man, his brain, wires and a computer screen\"  src=\"data:image\/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAEAAAABAQMAAAAl21bKAAAAA1BMVEUAAP+KeNJXAAAAAXRSTlMAQObYZgAAAAlwSFlzAAAOxAAADsQBlSsOGwAAAApJREFUCNdjYAAAAAIAAeIhvDMAAAAASUVORK5CYII=\"  class=\" pk-lazyload\"  data-pk-sizes=\"auto\"  data-ls-sizes=\"(min-width: 1466px) 754px, (max-width: 599px) 100vw, (min-width: 600px) 600px, 237px\"  data-pk-src=\"https:\/\/images.theconversation.com\/files\/614926\/original\/file-20240821-16-1equ7e.png?ixlib=rb-4.1.0&amp;q=45&amp;auto=format&amp;w=754&amp;fit=clip\"  data-pk-srcset=\"https:\/\/images.theconversation.com\/files\/614926\/original\/file-20240821-16-1equ7e.png?ixlib=rb-4.1.0&amp;q=45&amp;auto=format&amp;w=600&amp;h=464&amp;fit=crop&amp;dpr=1 600w, https:\/\/images.theconversation.com\/files\/614926\/original\/file-20240821-16-1equ7e.png?ixlib=rb-4.1.0&amp;q=30&amp;auto=format&amp;w=600&amp;h=464&amp;fit=crop&amp;dpr=2 1200w, https:\/\/images.theconversation.com\/files\/614926\/original\/file-20240821-16-1equ7e.png?ixlib=rb-4.1.0&amp;q=15&amp;auto=format&amp;w=600&amp;h=464&amp;fit=crop&amp;dpr=3 1800w, https:\/\/images.theconversation.com\/files\/614926\/original\/file-20240821-16-1equ7e.png?ixlib=rb-4.1.0&amp;q=45&amp;auto=format&amp;w=754&amp;h=583&amp;fit=crop&amp;dpr=1 754w, https:\/\/images.theconversation.com\/files\/614926\/original\/file-20240821-16-1equ7e.png?ixlib=rb-4.1.0&amp;q=30&amp;auto=format&amp;w=754&amp;h=583&amp;fit=crop&amp;dpr=2 1508w, https:\/\/images.theconversation.com\/files\/614926\/original\/file-20240821-16-1equ7e.png?ixlib=rb-4.1.0&amp;q=15&amp;auto=format&amp;w=754&amp;h=583&amp;fit=crop&amp;dpr=3 2262w\" ><\/a>\n            <figcaption>\n              <span class=\"caption\">How the UC Davis speech brain-computer interface deciphers neural activity and turns them into words.<\/span>\n              <span class=\"attribution\"><span class=\"source\">UC Davis Health<\/span><\/span>\n            <\/figcaption>\n          <\/figure>\n\n<h2 id=\"real-world-benefits\">Real-world benefits<\/h2>\n\n<p>In practice, this speech decoding strategy has been remarkably successful. We\u2019ve enabled Casey Harrell, a man with ALS, to \u201cspeak\u201d with over 97% accuracy using just his thoughts. This breakthrough allows him to easily converse with his family and friends for the first time in years, all in the comfort of his own home.<\/p>\n\n<p>Speech brain-computer interfaces represent a significant step forward in restoring communication. As we continue to refine these devices, they hold the promise of giving a voice to those who have lost the ability to speak, reconnecting them with their loved ones and the world around them. <\/p>\n\n<p>However, challenges remain, such as making the technology more accessible, portable and durable over years of use. Despite these hurdles, speech brain-computer interfaces are a powerful example of how science and technology can come together to solve complex problems and dramatically improve people\u2019s lives.<!-- Below is The Conversation's page counter tag. Please DO NOT REMOVE. --><img  loading=\"lazy\"  decoding=\"async\"  src=\"data:image\/png;base64,iVBORw0KGgoAAAANSUhEUgAAAAEAAAABAQMAAAAl21bKAAAAA1BMVEUAAP+KeNJXAAAAAXRSTlMAQObYZgAAAAlwSFlzAAAOxAAADsQBlSsOGwAAAApJREFUCNdjYAAAAAIAAeIhvDMAAAAASUVORK5CYII=\"  alt=\"The Conversation\"  width=\"1\"  height=\"1\"  style=\"border: none !important; box-shadow: none !important; margin: 0 !important; max-height: 1px !important; max-width: 1px !important; min-height: 1px !important; min-width: 1px !important; opacity: 0 !important; outline: none !important; padding: 0 !important\"  referrerpolicy=\"no-referrer-when-downgrade\"  class=\" pk-lazyload\"  data-pk-sizes=\"auto\"  data-pk-src=\"https:\/\/counter.theconversation.com\/content\/236998\/count.gif?distributor=republish-lightbox-basic\" ><!-- End of code. If you don't see any code above, please get new code from the Advanced tab after you click the republish button. The page counter does not collect any personal data. More info: https:\/\/theconversation.com\/republishing-guidelines --><\/p>\n\n  <p><span><a href=\"https:\/\/theconversation.com\/profiles\/nicholas-card-1635329\" target=\"_blank\" rel=\"noopener\">Nicholas Card<\/a>, Postdoctoral Fellow of Neuroscience and Neuroengineering, <em><a href=\"https:\/\/theconversation.com\/institutions\/university-of-california-davis-1312\" target=\"_blank\" rel=\"noopener\">University of California, Davis<\/a><\/em><\/span><\/p>\n\n  <p>This article is republished from <a href=\"https:\/\/theconversation.com\" target=\"_blank\" rel=\"noopener\">The Conversation<\/a> under a Creative Commons license. Read the <a href=\"https:\/\/theconversation.com\/from-thoughts-to-words-how-ai-deciphers-neural-signals-to-help-a-man-with-als-speak-236998\" target=\"_blank\" rel=\"noopener\">original article<\/a>.<\/p>\n<\/div>\n\n","protected":false},"excerpt":{"rendered":"Casey Harrell, who has ALS, works with a brain-computer interface to turn his thoughts into words. Nicholas Card&hellip;\n","protected":false},"author":926,"featured_media":12551,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"nf_dc_page":"","fifu_image_url":"https:\/\/images.pexels.com\/photos\/9599917\/pexels-photo-9599917.jpeg?auto=compress&cs=tinysrgb&w=1260&h=750&dpr=1","fifu_image_alt":"","footnotes":""},"categories":[12,16],"tags":[523,334,474],"class_list":{"0":"post-12549","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-health-and-body","8":"category-tech","9":"tag-als","10":"tag-artificial-intelligence","11":"tag-the-conversation","12":"cs-entry","13":"cs-video-wrap"},"aioseo_notices":[],"_links":{"self":[{"href":"https:\/\/modernsciences.org\/staging\/4414\/wp-json\/wp\/v2\/posts\/12549","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/modernsciences.org\/staging\/4414\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/modernsciences.org\/staging\/4414\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/modernsciences.org\/staging\/4414\/wp-json\/wp\/v2\/users\/926"}],"replies":[{"embeddable":true,"href":"https:\/\/modernsciences.org\/staging\/4414\/wp-json\/wp\/v2\/comments?post=12549"}],"version-history":[{"count":1,"href":"https:\/\/modernsciences.org\/staging\/4414\/wp-json\/wp\/v2\/posts\/12549\/revisions"}],"predecessor-version":[{"id":12550,"href":"https:\/\/modernsciences.org\/staging\/4414\/wp-json\/wp\/v2\/posts\/12549\/revisions\/12550"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/modernsciences.org\/staging\/4414\/wp-json\/wp\/v2\/media\/12551"}],"wp:attachment":[{"href":"https:\/\/modernsciences.org\/staging\/4414\/wp-json\/wp\/v2\/media?parent=12549"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/modernsciences.org\/staging\/4414\/wp-json\/wp\/v2\/categories?post=12549"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/modernsciences.org\/staging\/4414\/wp-json\/wp\/v2\/tags?post=12549"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}