EXPLORING ARTIFICIAL INTELLIGENCE
Editors Note: This article by William M. (’15) is an exploration of neurolinguistics and artificial intelligence, talking about how different types of artificial intelligence work and the effect it can have in the modern world and the issues present in this world. This article also talks about the study of neurolinguistics and how we can use this study to help fabricate artificial intelligence.
Blindness, in addition to the obvious physical limitations it brings, is devastating to the social life and health of the individuals whom it affects. It can be extremely difficult for the visually impaired not only to complete their daily routines but also to interact normally with those around them. One particular challenge for many blind people is the use of technology, since many computer devices require the user to operate a graphical user interface (GUI). However, there is a way to solve this problem. Discovery in the field of neurolinguistics will allow the visually impaired to enjoy the benefits of computer technology more directly because it will aid in the development of innovative systems of artificial intelligence that can replace traditional graphical interfaces.
To understand how to build a computer algorithm that can communicate with human language, the best place to look is the human brain itself. According to classical neurolinguistics, language in the brain is processed in two main regions: Broca’s Area and Wernicke’s Area. Broca’s Area coordinates syntax (the order of words within a sentence) during language production and comprehension, and Wernicke’s area acts as a lexical dictionary which efficiently maps sound to semantics (meaning). Although this distinction may seem straightforward, modern studies in linguistics are finding that language processing is much more complex than previously thought and that there is much overlap between the functions performed by these two regions. For example, scientists are unclear where morphology, a characteristic of language somewhere between syntax and semantics, is handled neurologically according to the current understanding. Another important unknown is how the brain anticipates upcoming words in order to speed up the rate at which it is able to comprehend a full sentence. Processing the subtleties of human language as quickly as it does is clearly one of the most advanced functions of the brain, and before software engineers can efficiently replicate this behavior with computer programs, neuroscientists need to map out exactly how the neurological device for language operates. In order to create advanced systems of artificial intelligence which replicate human language, the traditional model of language in the brain must be expanded upon and the inner workings of neurolinguistic pathways must be understood in greater detail.
A computer which communicates with its user via artificial intelligence as opposed to a traditional graphical user interface is much more accessible to the visually impaired because it can be controlled by voice alone. Although sophisticated artificial intelligence software has yet to be developed, I have personally seen how the relatively primitive algorithm behind Apple’s Siri has allowed my blind grandma to utilize her phone in a way not possible otherwise. Through simple voice commands, my grandma is able to check her email in order to stay in contact with the outside world, play music for her personal enjoyment, and perform countless other tasks which may be difficult otherwise. However, although Siri is a step in the right direction, it does not achieve the true potential of artificial intelligence in such an application. Siri is merely a limited extension of the pre-existing iOS interface, and the voice commands to which it can respond belong to a limited set pre-defined by its developers.
Products similar to Siri with improved natural language algorithms would have a wide range of practical applications. Rather than only parsing simple commands like “play music” or “call (917) 123-4567,” a more sophisticated natural language algorithm could relay information about the surrounding physical world through spoken language to a blind user. For example, a computed response to logically valid questions like “where are my keys?” and “what color is this chair?” could be accomplished by combining a high-quality camera with an advanced language processing algorithm. In addition, a computer able to respond to not only imperative statements but also process indicative ones would revolutionize how we interact with technology. Visually impaired people could use the voice control interface as a way of writing down reminders or notes just as those who have no problems seeing use sticky notes or blackboards. Computers would no longer just “do” stuff, but capture a mental snapshot of any sort of described scenario -- hypothetical or real. The development of statements as opposed to just commands in natural human language revolutionized human communication thousands of years ago, and it is not unreasonable to assume that such a development in computational communication would have far-reaching applications today.
Although sci-fi has given artificial intelligence a bad name, advanced language processing algorithms based upon discoveries in neurolinguistics could revolutionize how the blind interact with technology. Although the classic GUI will by no means disappear, the potential to create computers run by human language will not just be useful to the blind but also the average consumer. Language is one of--if not the most--natural modes of communication for people, and thus not including it in our interactions with computers is not using technology to its greatest potential. In the near future, you will not just type, click, and drag when using a computer, but also talk to it as well.
To understand how to build a computer algorithm that can communicate with human language, the best place to look is the human brain itself. According to classical neurolinguistics, language in the brain is processed in two main regions: Broca’s Area and Wernicke’s Area. Broca’s Area coordinates syntax (the order of words within a sentence) during language production and comprehension, and Wernicke’s area acts as a lexical dictionary which efficiently maps sound to semantics (meaning). Although this distinction may seem straightforward, modern studies in linguistics are finding that language processing is much more complex than previously thought and that there is much overlap between the functions performed by these two regions. For example, scientists are unclear where morphology, a characteristic of language somewhere between syntax and semantics, is handled neurologically according to the current understanding. Another important unknown is how the brain anticipates upcoming words in order to speed up the rate at which it is able to comprehend a full sentence. Processing the subtleties of human language as quickly as it does is clearly one of the most advanced functions of the brain, and before software engineers can efficiently replicate this behavior with computer programs, neuroscientists need to map out exactly how the neurological device for language operates. In order to create advanced systems of artificial intelligence which replicate human language, the traditional model of language in the brain must be expanded upon and the inner workings of neurolinguistic pathways must be understood in greater detail.
A computer which communicates with its user via artificial intelligence as opposed to a traditional graphical user interface is much more accessible to the visually impaired because it can be controlled by voice alone. Although sophisticated artificial intelligence software has yet to be developed, I have personally seen how the relatively primitive algorithm behind Apple’s Siri has allowed my blind grandma to utilize her phone in a way not possible otherwise. Through simple voice commands, my grandma is able to check her email in order to stay in contact with the outside world, play music for her personal enjoyment, and perform countless other tasks which may be difficult otherwise. However, although Siri is a step in the right direction, it does not achieve the true potential of artificial intelligence in such an application. Siri is merely a limited extension of the pre-existing iOS interface, and the voice commands to which it can respond belong to a limited set pre-defined by its developers.
Products similar to Siri with improved natural language algorithms would have a wide range of practical applications. Rather than only parsing simple commands like “play music” or “call (917) 123-4567,” a more sophisticated natural language algorithm could relay information about the surrounding physical world through spoken language to a blind user. For example, a computed response to logically valid questions like “where are my keys?” and “what color is this chair?” could be accomplished by combining a high-quality camera with an advanced language processing algorithm. In addition, a computer able to respond to not only imperative statements but also process indicative ones would revolutionize how we interact with technology. Visually impaired people could use the voice control interface as a way of writing down reminders or notes just as those who have no problems seeing use sticky notes or blackboards. Computers would no longer just “do” stuff, but capture a mental snapshot of any sort of described scenario -- hypothetical or real. The development of statements as opposed to just commands in natural human language revolutionized human communication thousands of years ago, and it is not unreasonable to assume that such a development in computational communication would have far-reaching applications today.
Although sci-fi has given artificial intelligence a bad name, advanced language processing algorithms based upon discoveries in neurolinguistics could revolutionize how the blind interact with technology. Although the classic GUI will by no means disappear, the potential to create computers run by human language will not just be useful to the blind but also the average consumer. Language is one of--if not the most--natural modes of communication for people, and thus not including it in our interactions with computers is not using technology to its greatest potential. In the near future, you will not just type, click, and drag when using a computer, but also talk to it as well.