Difference between revisions of "Computing For Human Experience"

From Knoesis wiki
Jump to: navigation, search
(Acknowledgments)
(From Perception to Semantics)
Line 29: Line 29:
 
CHE will bring together many current technological advances in capabilities that are easy and natural for humans but harder for machines, fundamentally combining human sensing with machine sensing and processing. In CHE,
 
CHE will bring together many current technological advances in capabilities that are easy and natural for humans but harder for machines, fundamentally combining human sensing with machine sensing and processing. In CHE,
  
    *  pattern recognition,
+
*  pattern recognition,
    * image analysis,
+
* image analysis,
    * casual text processing,
+
* casual text processing,
    * sentiment and intent detection,
+
* sentiment and intent detection,
    * using domain models to gather factual information, and
+
* using domain models to gather factual information, and
    * polling social media to gather community opinions and build intelligence
+
* polling social media to gather community opinions and build intelligence
  
 
will all come together to enable a system that makes conclusions and decisions with human-like intuition, but much more quickly than humans can do by themselves.
 
will all come together to enable a system that makes conclusions and decisions with human-like intuition, but much more quickly than humans can do by themselves.

Revision as of 23:00, 21 December 2009

Semantics empowered Sensors, Services, and Social Computing on ubiquitous Web

In his influential paper “The Computer for the 21st Century,” Mark Weiser talked about making machines fit the human environment instead of forcing humans to enter the machine’s environment.[MW] He noted, “The most profound technologies are those that disappear. They weave themselves into the fabric of everyday life until they are indistinguishable from it.” Weiser’s vision, outlined two decades ago, led to ubiquitous computing. Now, we must again rethink the relationship and interactions between humans and machines—this time, including a variety of technologies, including computing technologies; communication, social-interaction, and Web technologies; and embedded, fixed, or mobile sensors and devices.

Your browser may not support display of this image. We’re on the verge of an era in which the human experience can be enriched in ways we couldn’t have imagined two decades ago. Rather than depending on a single technology, we’ve progressed with several whose semantics-empowered convergence and integration will enable us to capture, understand, and reapply human knowledge and intellect. Such capabilities will consequently elevate our technological ability to deal with the abstractions, concepts, and actions that characterize human experiences. This will herald computing for human experience (CHE).

The CHE vision is built on a suite of technologies that serves, assists, and cooperates with humans to nondestructively and unobtrusively complement and enrich normal activities, with minimal explicit concern or effort on the humans’ part. CHE will anticipate when to gather and apply relevant knowledge and intelligence. It will enable human experiences that are intertwined with the physical, conceptual, and experiential worlds (emotions, sentiments, and so on [CB]), rather than immerse humans in cyber worlds for a specific task. Instead of focusing on humans interacting with a technology or system, CHE will feature technology-rich human surroundings that often initiate interactions. Interaction will be more sophisticated and seamless compared to today’s precursors such as automotive accident-avoidance systems.

Many components of and ideas associated with the CHE vision have been around for a while. Here, I discuss some of the most important tipping points that I believe will make CHE a reality within a decade.

Bridging the Physical/Digital Divide

We’ve already seen significant progress in technology that enhances human-computer interactions; the iPhone is a good example. Now we’re seeing increasingly intelligent interfaces, as exemplified by Tom Gruber’s “Intelligenc at the Interface” technology , which has demonstrated contextual use of knowledge to develop intelligent human–mobile-device interfaces. We’re also seeing progress in how machines (devices and sensors), surroundings, and humans interact, enabled by advances in sensing the body, the mind, and place. Such research supports the ability to understand human actions, including human gestures and languages in increasingly varied forms. The broadening ability to give any physical object an identity in the cyber world (that is, to associate the object with its representation), as contemplated with the Internet of Things, will let machines leverage extensive knowledge about the object to complement what humans process.

Human-machine interactions are taking place at a new level, as demonstrated by Psyleron’s Mind Lamp, which shows connections between the mind and the physical world, or Neuro Sky’s mind-controlled headset for playing video games. Soon, computers will be able to translate gestures to concrete actionable cues and understand perceptions behind human observations, as shown by MIT’s Sixth Sense project.

In addition, interactions initiated in the cyber world are increasing and becoming richer. Examples range from a location-aware system telling a smart phone user about a sale item’s availability at a nearby store to advanced processing of sensor data and crowd intelligence to recommend a road rerouting or to act on behalf of a human. This bridging of the physical/digital divide is a key part of CHE.

Elevating Abstractions That Machines Understand

Perception is a key aspect of human intelligence and experience. Elevating machine perception to a level closer to that of human perception will be a key enabler of CHE. In 1968, Richard Gregory described perception as a hypothesis over observation.[RG] Such hypothesis building comes naturally to people as an (almost) entirely subconscious activity. Humans often interpret the raw sensory observation before recognizing a conscious thought. On the other hand, hypothesis building is often cumbersome for machines. Nevertheless, to integrate human and machine perception, the convergence must occur at this abstraction level, often termed situation awareness. Therefore, regardless of the source, this integration requires a shared framework for communicating and comparing situation awareness.

Perceptual hypotheses represent the semantics, or meaning, of observation. Beginning with raw observation, we find such meaning by leveraging background knowledge of the interaction between observation and possible causes to determine the most likely hypothesis. Previous experience, schooling, and personality account for much of the background knowledge in a single mind. Machines also must leverage background knowledge for effective perception.[RB] So, effective perception requires a framework for representing background knowledge.

From Perception to Semantics

John Locke, Charles Peirce, Bertrand Russell, and many others have extensively and wonderfully written about semiotics—how we construct and understand meaning through symbols. A key enhancement we’re already seeing is the humanization of data and observation, including social computing extending semantic computing and vice versa. Metadata is no longer confined to structural, syntactic, and semantic metadata but includes units of observations that convey human experience, including perceptions, sentiments, opinions, and intentions.

Soon, we’ll be able to convert massive amounts of raw data and observations into symbolic representations. We’ll make these representations more meaningful through a variety of relationships and associations we can establish with other things we know, via semantics. We’ll then be able to contextually leverage all this to improve human activities and experience.


CHE will bring together many current technological advances in capabilities that are easy and natural for humans but harder for machines, fundamentally combining human sensing with machine sensing and processing. In CHE,

  • pattern recognition,
  • image analysis,
  • casual text processing,
  • sentiment and intent detection,
  • using domain models to gather factual information, and
  • polling social media to gather community opinions and build intelligence

will all come together to enable a system that makes conclusions and decisions with human-like intuition, but much more quickly than humans can do by themselves.

Semantics at an Extraordinary Scale

Semantic computing, aided by Semantic Web technology, is an ideal candidate framework for meaningful representation and sharing of hypotheses and background knowledge. Together with semantic computing, the large-scale adoption of Web 2.0 or social-Web technology has led to the availability of multimodal user-generated content, whether text, audio, video, or simply attention metadata, from a variety of online networks. The most promising aspect of this data is that it truly represents a population and isn’t a biased response or arbitrary sample study. This means that machines now have at their disposal the variety and vastness of data and the local and global contexts that we use in our day-to-day processing of information to gather insights or make decisions.

We also see a move from document- and keyword-centric information processing that relies on search-and-sift to representing information at higher abstraction levels. This involves moving from entity- or object-centric processing to relationship- and event-centric processing. This, in turn, involves improving the ability to extract, represent, and reason about a vast variety of relationships, as well as providing integral support for information’s temporal, spatial, and thematic elements.

With parallel advances in knowledge engineering, large-scale data analytics, and language understanding, we’re able to build systems that can process, represent, and reason over data points much as humans do. In addition, we can provide extremely rich markups of all the observations available to a machine, letting machines connect the dots (contexts) surrounding the observations (data) and draw conclusions that nearly mimic human perception and cognition. All these together are reducing the disparity between humans’ perceptions and the conclusions that machines draw from quantified or qualified observations.

Semantics-empowered social computing, semantics-empowered services computing (currently seen in the context of smart or semantic mashups), and semantics-enhanced sensor computing (exemplified by the semantic sensor Web) are key building blocks of CHE.

CHE can be seen as borrowing from or a synthesis of influential visions and seminal works (see box Influential and Interesting Works That Lead to Computing for Human Experience”).

Semantic Computing as a Starting Point

At the center of the approach to achieving CHE, as we see it, is semantic computing. Importance of semantics in dealing with data heterogeneity has been recognized since 1980s, when we saw emergence of conceptual or semantic models. Starting 1990s through early 2000s we saw the use of conceptual models or ontologies for semantic metadata extraction and annotations, for faceted or semantic search, and subsequently semantic analytics [HH, SR]. Figure 1 shows a contemporary architecture for semantic computing (simplified for brevity). It has four key components: data or resources, models and knowledge, semantic annotation, and semantic analysis or reasoning.

Figure 1. A contemporary architecture for semantic computing. This architecture supports semantic search and browsing, question answering, and situational awareness. To do this, it analyzes any form of Web, social, or sensor data by extracting metadata resulting in comprehensive semantic annotation. This process is aided by conceptual models and knowledge and by a variety of information-retrieval, statistical, and AI (machine learning and natural-language processing) techniques, at the Web scale. Semantic analysis supported by mining, inferencing and reasoning over annotations support applications.

Whereas semantic computing started with primarily enterprise and then Web data, including business and scientific data and literature, it has expanded to include any type of data (structured, semistructured, unstructured, and multimodal) and massive amounts of Web-accessible resources, including services, sensor, and social data. Here are some impressive examples:

   * the capture of comprehensive personal information (for example, MyLifebits),
   * the collection of massive amounts of interlinked curated data (for example, Linked Data [BH] ),
   * data and information contributed by a community of volunteers (for example, Wikipedia) and the record of social discourse of millions of users on a vast number of topics (for example, Facebook and Twitter), and
   * the collection of observations from sensors in, on, or around humans; around the earth (for example, Hewlett-Packard’s Central Nervous System for the Earth [CeNSE] initiative) and in space.


Semantic computing over such a bewildering variety of data is made possible largely by an agreement on what the data means, represented in a manner that’s formal or informal; explicit or implicit; or static (through a deliberate, expert-driven process), periodic, or dynamic (for example, mining Wikipedia to extract a targeted taxonomy). The key forms in which such agreements are modeled include formal ontologies, folksonomies, taxonomies, vocabularies, and dictionaries. Recently we have started to make rapid strides in creating models and background knowledge from human collaborations (exemplified by hundreds of expert created ontologies), by selectively extracting or mining the Web for facts (as demonstrated by Voquette/Semangix [SB]) or scientific literature, and by harvesting community created content (as demonstrated by Yago and Taxonom.com). This has involved large-scale use of statistical, machine-learning and natural language processing techniques. Along with domain-specific or thematic conceptual models, temporal and spatial models (ontologies) have taken their rightful place for capturing meaning, especially as we seek to go from keywords and documents to objects, and then to relationships and events.

Such models and associated background knowledge have provided powerful ways to automatically extract semantic metadata or semantically annotate any type of data to associate meaning with the data. A variety of semantic computations aided by pattern extraction, inferencing, logic- and rule-based reasoning, and so on, then provide a range of applications including semantic search, browsing, integration, and analysis. Such applications can lead to insights, decision support, and situational awareness. Twitris is one such application for extraction of social signals from Twitter.

An Illustrative Example

To get an idea of some of the capabilities I’ve described, consider a scenario in which a farmer observes an unfamiliar disease on his crop and seeks information to manage it (see Figure 2). He clicks a picture, tags it with keywords “crop” and “blight,” and sends a message seeking more information: “Looks bad, but I don’t know what it is. Any help would be great.”

Figure 2. An example embodying some of the computing for human experience (CHE) promises—helping a farmer in his natural context. The farmer sends a message requesting help with an unknown crop disease. The CHE system analyzes the message, contacts appropriate sources, and returns actionable information, while requiring minimal involvement or technology consciousness from the farmer.

A CHE system would analyze the image to help identify the exact crop (for example, sweet corn) and the disease. It would also use location coordinates and related contextual and background domain knowledge, including local weather and soil conditions (for example, to determine whether the disease is northern corn leaf blight). It would analyze the farmer’s message to extract the information-seeking intent and detect an unfavorable sentiment associated with the content.

It would then broadcast this information to online forums and social networks of individuals whose profiles indicate a professional or scientific interest in farming, crop diseases, and plant pathology. It would track responses, prioritizing those from authoritative sources, pulling actionable information based on their suggestions, aggregating duplicate suggestions, filtering spam, and presenting summaries of crowd-contributed intelligence. In this case, the actionable information could be a list of suitable fungicides and pesticides and their prices, buying options, and action plans (for example, if the farmer has a sweet corn variety, he could spray fungicide and then use hybrid seed in the future to provide blight resistance). The CHE system would also have a feedback mechanism, prompting the farmer for progress and informing the community when metrics deviate from known specifications.

The system would do all these things while requiring minimal involvement from the farmer.

This example doesn’t capture all the promise of advances in machines interacting with humans at higher abstraction levels. Perhaps the CHE system, using machine perception, could also detect the disease-ridden crops before the farmer even notices and initiate the search I described, providing actionable intelligence to the farmer. More than just provide automation, CHE would work around the farmer’s natural work pattern and environment, making technology interactions minimal and natural.

Acknowledgments

I developed the ideas proposed here in the context of research in the Semantic Web and semantics-enabled services, sensor, and social computing at Wright State University’s Kno.e.sis Center. I’m grateful for the contributions of Kno.e.sis researchers, especially Karthik Gomadam, Cory Henson, and Meena Nagarajan. I also thank the US National Science Foundation, the US National Institutes of Health, the US Air Force Research Laboratory, IBM, Hewlett-Packard, and Microsoft for their support.