Friday, September 4, 2015

Turing Tests, Search & Current AI



[H]ow many different automata or moving machines can be made by the industry of man [...] For we can easily understand a machine's being constituted so that it can utter words, and even emit some responses to action on it of a corporeal kind, which brings about a change in its organs; for instance, if touched in a particular part it may ask what we wish to say to it; if in another part it may exclaim that it is being hurt, and so on.

Rene Descartes. 1637. Discourse on Method[1],[2]







I recently saw the film Ex Machina. Leaving aside what I thought about the story, the characters etc. (although there was a lot to like & dislike in both), there were two aspects of the movie that were very interesting (IMHO) & which bear further thought & comment. The first was an updated idea of a Turing Test, & the second was the idea that the infrastructure & function of a general search engine could serve as the basis for a functioning artificial intelligence. An examination of these topics will cover quite a bit of current work on various AI related topics.

First, the Turing Test[3]… In 1950, Alan Turing proposed an “imitation game” for determining a machine’s (computer’s) ability to exhibit intelligent behavior equivalent to, or at least indistinguishable from, a human being. There are many versions of this test, but the essence of it is that natural language is appropriate for this test & that a human interrogator will not be able to distinguish the conversation they are having (by text) as generated by a human or a machine. Turing originally proposed that if 30% of interrogators were not able to determine which other player was a human & which was a machine after five minutes of “conversation”, the machine would be said to have passed the test. There have been many competitions to determine if a machine could pass this test, & in fact the Loebner Prize competition has been held every year since 1991. No machine has yet won the full version of this prize.

In 2014, a competition held at the University of Reading (UK) was said to have been won by the Russian chatbot Evgenyi Goostman. This result was highly controversial (surprise, surprise) as the chatbot was engineered to represent a 13-year old Ukrainian boy who was not a native English speaker. Nevertheless, Evgenyi convinced 33% of the contest’s judges that it was human in a series of five-minute conversations.

Although also controversial, the vast majority of these contests were conducted under the model that the interrogator did not necessarily know that one of the players was a machine. In one of the few competitions to test this principle, the 2008 Loebner Prize (Reading University, UK) examined the interrogator’s ability to distinguish between machine & human even in machine/machine & human/human pairs. No significant difference was found among interrogators who knew that a machine was part of the test & those who did not. Perhaps more interesting was that interrogators used criteria such as spelling errors to distinguish humans (machines did not make such errors) & speed & length of response to distinguish machines.

What about the movie you say… Oh yes, the movie. The primary difference in the movie is that a human is tasked with performing a Turing Test on a humanoid robot that is clearly not human. The interrogator & robot meet face-to-face for the test, & even though the interrogator knows the subject is a robot, he winds up becoming emotionally attached with disastrous consequences. My point is not the moral & real world consequences of emotional attachments to robots, but that this type of “Turing Test” will be upon us sooner rather than later, & that we’ll have to have not only a technical but also a cultural context for dealing with intelligence exhibited by non-humans. Oh, & incidentally, not all of these non-humans will be attractive lifelike robots – most will be entities such as massive networks of devices, or intelligent programmatic agents… How will we deal with these entities? Entities that, in their ability to communicate & analyze information may not be distinguishable from human beings. Perhaps our only clues will be just that these entities will be much better at communicating & analyzing information. What kind of sociocultural adaptations will we have to make in order to function in this “brave new world”? How will people, humans, work & live alongside these entities.  That brings me to search & back to the movie.

As I have already said, one of the most interesting (to me) aspects of the movie was the conceit that a massive, general search engine, that was optimized for certain types of intelligence relevant to natural language, personalization, deep learning & search, could serve as the basis for an artificial intelligence that could pass the modified Turing Test described above. About ten years ago (1/2006), I wrote a paper titled: If Search is the Answer, What’s the Question[4]. In this paper I predicted that in the 2012 timeframe, peoples’ work process would be primarily knowledge & model based, & that “search” would provide the overall structure for this work process that would emphasize information curation & problem solving. I wrote a number of these predictive papers in that time period & I generally got the direction right & the timeframe wrong – I always estimated that change would happen faster than it actually did. That’s the case here too. Now in 2015, we are not yet at the point where models & problem solving are the primary work context for knowledge workers, but there are several trends taking us in this direction. In addition, there is the overhyped, but nevertheless immensely important trend of ultra-large data set analysis (big data) that is also restructuring the concept of work & work process. How has search changed since 2006 & where are we in this evolution? Of course, the question raised by the movie is “does search provide an adequate basis for general artificial intelligence”? Lot’s of stuff to address here…

A lot has happened in search since 2006, or has it. This is true technically, but from an end-user perspective the most visible occurrence has been the emergence of Google as the world’s preeminent search utility. A recent report from AYTM Market research found that 74.3% of worldwide consumers use Google as their primary search engine. SearchStatsBrain[5] reports that Google performed 2.1 trillion (that’s with a T) searches in 2014 – almost 6 billion a day! The number of searches not performed on Google, Bing or Yahoo is trivial. All three engines have deployed almost all new function, at the end-user level, over a relatively short period of time. New capabilities such as local search &/or vertical search are available on the “big three”, but also through specialized apps such as Yelp or De.Li.cious. Perhaps the most interesting, & relevant for this article, new feature is natural language search, as offered by Google (Google Voice), Bing (Smart Search, Microsoft) as well as by such personal assistants as Siri (Apple) & Cortana (Microsoft) & a growing number of others. Natural language provides the underlying interaction context for not only a modified Turing Test, but also for our long-term adoption & adaptation to intelligent systems.

What about the other dimensions I predicted search would have to proceed in to be able to serve as the primary structuring agent for people’s work that was based on model interpretation & problem-solving. In 2006 I said:[6]

“During this transition, search will have to evolve itself from the ubiquitous web & enterprise engines of today that still mainly operate on key word & page rank algorithms to much more deeply focused tools. These tools will be able to refine their operation by using coarse & fine-grained models, not just of business, but also of more general knowledge categories. They will initially be able to structure work process because of their interactions with such models & eventually, in the 2010-2012 timeframe, facilitate work organization & problem solving as well as location of general or specialized knowledge in the context of a person’s work (or personal) process. This will require the integration with search of such areas as classification, ontology-based & advanced metadata (currently RDF & geospatial but also evolving quickly) modeling, rule-based reasoning & non-deductive reasoning of various forms – this is just the beginning, but we are already seeing some of these advances in products such as Mooter, Clusty or Grokker or in the integration of rule-based reasoning with business process management &/or ontology-based modeling (Protégé, Swoop). This work is just at its beginning.”

Most of this has not happened. Search engines, at least the ones used by the vast majority of people, do not currently use models or rule bases or any other type of reasoning, to facilitate work organization & problem solving. Classification is provided by a very few engines that are hardly used (compared to the Big 3). Mooter & Grokker no longer exist except as Source Forge downloads last updated in 2007. Clusty is currently called Yippy (www.yippy.com). It appears to do the same thing that Clusty did in 2007 – provide a sidebar of topics that the search results can be sorted into. There are certainly specialized geospatial search engines (GEOSS, GSE etc.), & geospatial search has been integrated into apps such as Google Maps etc. The larger picture is that search has progressed a lot, but its primary direction has been the facilitation a different types of monetization, thus emphasizing the transactional nature of search engine use.

That’s not to say there hasn’t been some movement in this direction. Two areas do stand out: semantic search & big data. Semantic search has been in development for a long time & can be described as follows:

Semantic search seeks to improve search accuracy by understanding searcher intent and the contextual meaning of terms as they appear in the searchable dataspace, whether on the Web or within a closed system, to generate more relevant results. Semantic search systems consider various points including context of search, location, intent, variation of words, synonyms, generalized and specialized queries, concept matching and natural language queries to provide relevant search results.”[7],[8]

The big 3 search engines all use some aspects of semantic search. Apart from attempting to use language related techniques such as synonym mapping, semantic search engines use additional techniques including, but not limited to: RDF search & RDF path traversal, keyword to concept mapping, analysis of graph patterns to identify relationships, analysis of other (more complex) patterns, use of ontology-based inference (OWL), use of other nonstandard logics[9]. 

The following figure is a relationship matrix generated by the semantic search engine SenseBot for the query “semantic search”. It represents important references by text size & provides hyperlinks for each topic. Other semantic search engines


BARBARA STARR     BING     CONFERENCE     CONTEXT     DISAMBIGUATION     ERIN EVERHART     GOOGLE     INTENT     KEYWORDS     KNOWLEDGE GRAPH     LISTS     MARKETING     MARKUP     MEANING     MICROSOFT     NAVIGATION     ONTOLOGIES     QUERIES     QUERY     SEARCH ENGINE     SEARCH ENGINES     SEMANTIC SEARCH     SEMANTICS     SEO     SOCIAL MEDIA     SOCIALPRO     STANDARDS     STRUCTURED DATA     TECHNOLOGY     UNDERSTANDING    
     

    Relationship Matrix (Sensebot) for Semantic Search, accessed 15 July 2015


represent results differently, some as sidebar topics, some as graphics, but all represent relationships or semantic groupings in some way. The user can choose what semantic dimension to explore. In the above example, the user could investigate companies, data topics or even languages. Search is optimized in the sense that the user gets to emphasize what aspect of the information they were looking for is explored. Each of the mainstream search companies have semantic search projects, & each introduces some aspect of semantic search in updates to their product. I’ll discuss other AI directions next.

Big data – I already hear you rolling your eyes… I’ve been doing a project based on “big data” analysis with healthcare safety-net organizations, & I have a much better appreciation of the strengths & weaknesses of this approach now that I’ve been using various aspects of it for the past year & a half. Why, you are saying, are we talking about it here – this is supposed to be about search. Yes – correct… the underpinnings of current big data analysis was initially all about search. Sometime in 2002, Doug Cutting & Mike Cafarella were working on an Apache search project called Lucerne at the University of Washington. They developed a web indexer called Nutch that eventually was able to run on up to 4-5 nodes & was indexing hundreds of millions of web pages, but still was not operating at “web-scale”, even for 2003-2004 timeframe. Engineers at Google published several seminal papers around this time[10],[11] on the Google File System & MapReduce, a programming model & implementation for processing very large data sets. Cutting & Cafarella decided to use this set of technologies as the basis for an improved indexer & rewrote their systems in Java (Google had implemented them in C++). Cutting then joined Yahoo & over time Hadoop, the system that evolved from the Nutch project was the basis for all search & transactional interaction for Yahoo. By 2011 it was running on 42,000 servers with hundreds of petabytes of storage. Yahoo spun out the distributed file system & MapReduce as open source projects under Apache, & many other companies, research groups & universities started developing tools, apps & applications forming the Hadoop ecosystem. Several companies developing the Hadoop ecosystem were also spun out, either directly or as engineers left Yahoo including Cloudera & Hortonworks.

OK – so back to today when most ultra-large scale projects, whether they are directly search based or analytic, are layered on some flavor of Hadoop (or some flavor of Hadoop-inspired software such as IBM Spark). The point, however, is not that Hadoop is the ultimate answer for search or for analytic processing in general[12] (it’s not…). It is that we have moved from enterprise distributed environments that include relational databases to shared-nothing clusters with massively parallel file & analysis systems. Those systems may be Hadoop based, or Spark[13] based or use Dremel[14] for stream processing or visualization tools for presentation & visual analysis. We are now in an era of massively parallel storage & analysis architectures & these architectures enable a type of processing not previously possible.

So what else is going on that is driving this larger vision of search. Like big data, deep learning has also run up the hype cycle to the extent that mainstream media has already informed the non-technogeek public about the wonders of the technology. The NY Times, the paper of record, ran stories as early as June 2012 & The Economist as early as February 2014[15]. The thing that people know about deep learning is that Watson, a system designed & built by IBM, beat a set of former champions at Jeopardy in February of 2011. While Watson is not purely a deep learning system (it uses a broad variety of techniques), it was the first example of this technology broadly visible outside of the AI community. More recently, companies including Yahoo, Facebook & particularly Google have continued the development of this learning technology. So what is it? & why is it important for search & many other areas going forward?




ABILITY     ALGORITHMS     APPLICATIONS     ARTIFICIAL INTELLIGENCE              BRAIN                             DEEP LEARNING     FACEBOOK              GOOGLE     HINTON     IMAGE RECOGNITION     IMAGES     INDUSTRY                    LANGUAGE                        LEARNING                                          MACHINE LEARNING         MACHINE LEARNING CLOSER               MACHINES                   MICROSOFT     MIT TECHNOLOGY REVIEW     NEURAL NETWORKS          ORIGINAL GOALS                           PATTERNS     RECOGNITION     REPRESENTATION        SOUND     SPEECH                        TECHNIQUES     TECHNOLOGY    
 


      Relationship Matrix (Sensebot) for Deep Learning, accessed 20 July 2015

There is no agreed upon, single definition of deep learning, but most people working on its development would agree that it is a type of machine learning characterized by:[16] 1) the use of multiple layers of nonlinear processing units (usually a neural net), 2) the supervised (through examples) or unsupervised learning of feature sets (patterns) in each layer with the layers organized in a hierarchy from low to high level features where the output of one layer serves as the input for the next higher layer. OK – that didn’t mean a whole lot… Actually what this technology is about is the recognition of patterns or features as information is fed through a neural net, with each layer of the net developing a more & more detailed description of the feature until an interpretation of the information can be made. A good example would be automated facial recognition. The most abstract layers of the net recognize that the overall pattern is a face, then subsequent layers identify & resolve additional “features” or “patterns” such as the mouth, the nose etc. & eventually the entire face is resolved. At this point there is a detailed digital record of the patterns so that this face could potentially be identified from a repository of faces & additional faces could be resolved, as the abstract patterns that make up a face have been defined.

There are specialized architectures for deep learning networks & many different mathematical models & associated algorithms for training & learning. To date, much of the focus in this area has been on image recognition (classification) & speech processing. An example from image recognition is the Google Brain project in which a deep learning network learned to recognize human faces & cats from entirely unlabeled data[17].

What does this mean for search? Imagine a search that instead of keying on specific words or terms instead was able to determine abstract patterns in a request & then respond by specializing that abstraction in a particular context & return information (in whatever form) relevant to the overall pattern or context, instead of just what it was able to match syntactically. Some time ago, I was the lead for a project that built a system[18] that used much more naïve pattern recognition to return information that it determined was “analogous” to a description given by the requester. While this system was not very powerful compared to today’s learning systems, it did often surprise the requestor with an analogy that it had identified. The system provided an explanation of why it had suggested the analogy & often the requestor would be puzzled at first but then would agree with the system once it saw the explanation. Wouldn’t it be nice to be surprised like this by a search engine? Deep learning will provide that capability & much more.

I’ve talked about semantic search, big data & machine learning with reference to where search is going. It’s currently mid-2015 – I’m only three years late with respect to the context for search I predicted back in 2008. We are currently at search on a cusp – a cusp that could easily push it over to the broad-scale service based on patterns, models & analogies that assists in structuring our work (inquiry) & facilitates problem solving in ways we might not have developed or thought of. The ability of search to use deep learning capabilities means that pattern recognition of all sorts will lead to models that allow hypothesis testing (as already done by IBM Watson) & the facilitation of problem solving in context. This latter will require frameworks (ontologies, rule sets & trained networks) that will allow problems to be represented & reasoned about. An additional approach will be the recognition of patterns & models in ultra-large scale data sets & subsequent data characterization or reasoning about the empirical data. Search using these mechanisms will be different than it is today, I’d say it will be better than it is today; better at finding results that match not only the content of our requests, but also match & potentially expand their context. At that point, 3-5 years from now, will it be capable of supporting an independent artificial intelligence. My very strong feeling is that unless such programs are allowed to evolve on their own, this will not happen, but then again, I guess we’ll just have to ask it.

Up next:
·      Intelligent search, big data, deep learning in healthcare information technology (& healthcare in general)
·      Design as a model for the evolution of work… What will future knowledge workers be like?
·      & further in the future… a “meditation” on the evolution of information technology using the “cluster of terms” model[19]






[1] Image originally of the Orion Nebula (NGC 1976), (Hubble) Space Telescope Science Institute, postprocessed by deepdreamr.com (16 July 2015
[2] Rene Descartes. 1637. Discourse on the Method of Rightly Conducting One's Reason and of Seeking Truth in the Sciences. Leiden, Netherlands.
[3] Turing, Alan (October 1950), "Computing Machinery and Intelligence", Mind LIX (236): 433–460, doi:10.1093/mind/LIX.236.433, ISSN 0026-4423, retrieved 2015-07-05
[4]  With respect to Danny Bobrow for his 1985 paper If Prolog is the Answer, What’s the Question?
IEEE Trans. Softw. Eng. 11(11). – perhaps the most insightful paper on the logic of AI languages
ever published (with the possible exception of Doug Lenat’s paper on why AM worked…), presented to MIT ESD Seminar Series, 3/2006.
[5] http://www.statisticbrain.com/Google-searches/
[6] I also started a piece of work, initially in 1994, looking at design practice as a model for knowledge work. Much of my thinking on how work would evolve came from this study.  c.f. Knowledge Work as Design: A Description of a Post-Convergence Work Paradigm. PostTechnical Research Strategist. May 2002.
[7] https://en.wikipedia.org/wiki/Semantic_search
[8] John, Tony (March 15, 2012). "What is Semantic Search?". Techulator. Retrieved 9 July 2015
[9] Mäkelä, Eetu. "Survey of Semantic Search Research" (PDF). Retrieved 9 July 2015.
[10] Chemawat, S., H. Gobioff & S-T Leung. 2003. The Google File System. ACM 1-58113-757-5/03/0010.
[11] J. Dean & S. Chemawat. 2004. MapReduce: Simplified Processing on Large Clusters. 6th Symposium on Operating Systems Design & Implementation. 2004. 137-149. San Francisco, CA.
[13] http://www.computerworld.com/article/2856063/enterprise-software/hadoop-successor-sparks-a-data-analysis-evolution.html
[15] NY Times: 21 May 2015, 12 June 2012; The Economist: 1 February 2014, 13 May 2015
[16] summarized from: https://en.wikipedia.org/wiki/Deep_learning, accessed 15 July 2015
[17] Ng, Andrew, et al. "Building High-level Features Using Large Scale Unsupervised Learning”. 29th Internaltion Conference on Machine Learning. Edinburgh, Scotland. 2012
[18] GNOsys, Digital Equipment Corporation, also see:
Hartzband, D.J. 1987. The provision of inductive problem solving and (some) analogic learning in model-based systems. Group for Artificial Intelligence and Learning (GRAIL), Knowledge Systems Laboratory. Stanford University. Stanford, CA, USA. 6/87.
Hartzband, D.J. 1987. A discussion of inference and problem solving in the GNOsys knowledge model. Problem Solving Systems Group. Artificial Intelligence Technology Group. Digital Equipment Corporation. 5/87.
Hartzband, D.J., L. Holly, and F.J. Maryanski. 1987. The provision of induction in data-model systems: I. Analogy. International Journal of Approximate Reasoning (IJAR) 1(1):1-17.
[19] Foster, Hal. 2015. Bad New Days: Art, Criticism, Emergency. Verso. NYC. 208 pp.

1 comment: