LIMSwiki

Natural-language user interface (LUI or NLUI) is a type of computer human interface where linguistic phenomena such as verbs, phrases and clauses act as UI controls for creating, selecting and modifying data in software applications.

In interface design, natural-language interfaces are sought after for their speed and ease of use, but most suffer the challenges to understanding wide varieties of ambiguous input.[1] Natural-language interfaces are an active area of study in the field of natural-language processing and computational linguistics. An intuitive general natural-language interface is one of the active goals of the Semantic Web.

Text interfaces are "natural" to varying degrees. Many formal (un-natural) programming languages incorporate idioms of natural human language. Likewise, a traditional keyword search engine could be described as a "shallow" natural-language user interface.

Overview

A natural-language search engine would in theory find targeted answers to user questions (as opposed to keyword search). For example, when confronted with a question of the form 'which U.S. state has the highest income tax?', conventional search engines ignore the question and instead search on the keywords 'state', 'income' and 'tax'. Natural-language search, on the other hand, attempts to use natural-language processing to understand the nature of the question and then to search and return a subset of the web that contains the answer to the question. If it works, results would have a higher relevance than results from a keyword search engine, due to the question being included.[citation needed]

History

Prototype Nl interfaces had already appeared in the late sixties and early seventies.[2]

  • SHRDLU, a natural-language interface that manipulates blocks in a virtual "blocks world"
  • Lunar, a natural-language interface to a database containing chemical analyses of Apollo 11 Moon rocks by William A. Woods.
  • Chat-80 transformed English questions into Prolog expressions, which were evaluated against the Prolog database. The code of Chat-80 was circulated widely, and formed the basis of several other experimental Nl interfaces. An online demo is available on the LPA website.[3]
  • ELIZA, written at MIT by Joseph Weizenbaum between 1964 and 1966, mimicked a psychotherapist and was operated by processing users' responses to scripts. Using almost no information about human thought or emotion, the DOCTOR script sometimes provided a startlingly human-like interaction. An online demo is available on the LPA website.[4]
  • Janus is also one of the few systems to support temporal questions.
  • Intellect from Trinzic (formed by the merger of AICorp and Aion).
  • BBN’s Parlance built on experience from the development of the Rus and Irus systems.
  • IBM Languageaccess
  • Q&A from Symantec.
  • Datatalker from Natural Language Inc.
  • Loqui from BIM Systems.
  • English Wizard from Linguistic Technology Corporation.

Challenges

Natural-language interfaces have in the past led users to anthropomorphize the computer, or at least to attribute more intelligence to machines than is warranted. On the part of the user, this has led to unrealistic expectations of the capabilities of the system. Such expectations will make it difficult to learn the restrictions of the system if users attribute too much capability to it, and will ultimately lead to disappointment when the system fails to perform as expected as was the case in the AI winter of the 1970s and 80s.

A 1995 paper titled 'Natural Language Interfaces to Databases – An Introduction', describes some challenges:[2]

Modifier attachment
The request "List all employees in the company with a driving licence" is ambiguous unless you know that companies can't have driving licences.
Conjunction and disjunction
"List all applicants who live in California and Arizona" is ambiguous unless you know that a person can't live in two places at once.
Anaphora resolution
resolve what a user means by 'he', 'she' or 'it', in a self-referential query.

Other goals to consider more generally are the speed and efficiency of the interface, in all algorithms these two points are the main point that will determine if some methods are better than others and therefore have greater success in the market. In addition, localisation across multiple language sites requires extra consideration - this is based on differing sentence structure and language syntax variations between most languages.

Finally, regarding the methods used, the main problem to be solved is creating a general algorithm that can recognize the entire spectrum of different voices, while disregarding nationality, gender or age. The significant differences between the extracted features - even from speakers who says the same word or phrase - must be successfully overcome.

Uses and applications

The natural-language interface gives rise to technology used for many different applications.

Some of the main uses are:

  • Dictation, is the most common use for automated speech recognition (ASR) systems today. This includes medical transcriptions, legal and business dictation, and general word processing. In some cases special vocabularies are used to increase the accuracy of the system.
  • Command and control, ASR systems that are designed to perform functions and actions on the system are defined as command and control systems. Utterances like "Open Netscape" and "Start a new xterm" will do just that.
  • Telephony, some PBX/Voice Mail systems allow callers to speak commands instead of pressing buttons to send specific tones.
  • Wearables, because inputs are limited for wearable devices, speaking is a natural possibility.
  • Medical, disabilities, many people have difficulty typing due to physical limitations such as repetitive strain injuries (RSI), muscular dystrophy, and many others. For example, people with difficulty hearing could use a system connected to their telephone to convert a caller's speech to text.
  • Embedded applications, some new cellular phones include C&C speech recognition that allow utterances such as "call home". This may be a major factor in the future of automatic speech recognition and Linux.

Below are named and defined some of the applications that use natural-language recognition, and so have integrated utilities listed above.

Ubiquity

Ubiquity, an add-on for Mozilla Firefox, is a collection of quick and easy natural-language-derived commands that act as mashups of web services, thus allowing users to get information and relate it to current and other webpages.

Wolfram Alpha

Wolfram Alpha is an online service that answers factual queries directly by computing the answer from structured data, rather than providing a list of documents or web pages that might contain the answer as a search engine would.[5] It was announced in March 2009 by Stephen Wolfram, and was released to the public on May 15, 2009.[6]

Siri

Siri is an intelligent personal assistant application integrated with operating system iOS. The application uses natural language processing to answer questions and make recommendations.

Siri's marketing claims include that it adapts to a user's individual preferences over time and personalizes results, and performs tasks such as making dinner reservations while trying to catch a cab.[7]

Others

  • Ask.com – The original idea behind Ask Jeeves (Ask.com) was traditional keyword searching with an ability to get answers to questions posed in everyday, natural language. The current Ask.com still supports this, with added support for math, dictionary, and conversion questions.
  • Braina[8] – Braina is a natural language interface for Windows OS that allows to type or speak English language sentences to perform a certain action or find information.
Screenshot of GNOME DO classic interface
  • GNOME Do – Allows for quick finding miscellaneous artifacts of GNOME environment (applications, Evolution and Pidgin contacts, Firefox bookmarks, Rhythmbox artists and albums, and so on) and execute the basic actions on them (launch, open, email, chat, play, etc.).[9]
  • hakia – hakia was an Internet search engine. The company invented an alternative new infrastructure to indexing that used SemanticRank algorithm, a solution mix from the disciplines of ontological semantics, fuzzy logic, computational linguistics, and mathematics. hakia closed in 2014.
  • Lexxe – Lexxe was an Internet search engine that used natural-language processing for queries (semantic search). Searches could be made with keywords, phrases, and questions, such as "How old is Wikipedia?" Lexxe closed its search engine services in 2015.
  • Pikimal – Pikimal used natural-language tied to user preference to make search recommendations by template. Pikimal closed in 2015.
  • Powerset – On May 11, 2008, the company unveiled a tool for searching a fixed subset of Wikipedia using conversational phrases rather than keywords.[10] On July 1, 2008, it was purchased by Microsoft.[11]
  • Q-go – The Q-go technology provides relevant answers to users in response to queries on a company’s internet website or corporate intranet, formulated in natural sentences or keyword input alike. Q-go was acquired by RightNow Technologies in 2011.
  • Yebol – Yebol was a vertical "decision" search engine that had developed a knowledge-based, semantic search platform. Yebol's artificial intelligence human intelligence-infused algorithms automatically clustered and categorized search results, web sites, pages and content that it presented in a visually indexed format that is more aligned with initial human intent. Yebol used association, ranking and clustering algorithms to analyze related keywords or web pages. Yebol integrated natural-language processing, metasynthetic-engineered open complex systems, and machine algorithms with human knowledge for each query to establish a web directory that actually 'learns', using correlation, clustering and classification algorithms to automatically generate the knowledge query, which was retained and regenerated forward.[12]

See also

References

  1. ^ Hill, I. (1983). "Natural language versus computer language." In M. Sime and M. Coombs (Eds.) Designing for Human-Computer Communication. Academic Press.
  2. ^ a b Natural Language Interfaces to Databases – An Introduction, I. Androutsopoulos, G.D. Ritchie, P. Thanisch, Department of Artificial Intelligence, University of Edinburgh
  3. ^ "Chat-80 demo". Archived from the original on 11 November 2016. Retrieved 29 January 2018.
  4. ^ "ELIZA demo". Archived from the original on 26 November 2016. Retrieved 29 January 2018.
  5. ^ Johnson, Bobbie (2009-03-09). "British search engine 'could rival Google'". The Guardian. Retrieved 2009-03-09.
  6. ^ "So Much for A Quiet Launch". Wolfram Alpha Blog. 2009-05-08. Retrieved 2009-10-20.
  7. ^ "iOS - Siri". Apple. Retrieved 29 January 2018.
  8. ^ "Braina - Artificial Intelligence Software for Windows". www.brainasoft.com. Retrieved 29 January 2018.
  9. ^ Ubuntu 10.04 Add/Remove Applications description for GNOME Do
  10. ^ Helft, Miguel (May 12, 2008). "Powerset Debuts With Search of Wikipedia". The New York Times.
  11. ^ Johnson, Mark (July 1, 2008). "Microsoft to Acquire Powerset". Powerset Blog. Archived from the original on February 25, 2009.
  12. ^ Humphries, Matthew. "Yebol.com steps into the search market" Archived 2012-03-15 at the Wayback Machine Geek.com. 31 July 2009.