Growth Series BLOG

Its a Premium BLOG template and it contains Instagram Feed, Twitter Feed, Subscription Form, Blog Search, Image CTA, Topic filter and Recent Post.

All Posts

Getting Started with Cognitive Computing

This article originally appeared in KMWorld Magazine.  

Cognitive computing is real. It is here now, and as time goes on, it is only going to become more widely adopted and provide competitive advantages that will allow some enterprises to leapfrog their competition. And getting started may not be as difficult as you imagine.

Most organizations already have applications in which cognitive computing technologies and approaches can be leveraged. In fact, they likely already have some capabilities that can be effectively applied to solving problems with cognitive computing technologies. The elements required to make cognitive computing meaningful and effective are the same ones that are needed to make any information retrieval project meaningful and effective.

New tools and methodologies can provide greater functionality and capabilities, but it is best to think of it as an extension or evolution rather than a revolution. I am sure there are lots of industry pundits and marketers who will breathlessly exclaim how revolutionary this is—and yes, there will be vastly different ways to work in the near future. But organizations don’t deploy revolutions, and even if there were a magic application that would revolutionize work, implementing it and integrating it into business processes, business practices and the organizational culture would take time. Therefore, companies are better off working incrementally to apply the skills they have while moving toward a more sophisticated approach over the long term.

The eternal KM challenge of customer self-service

One application that is getting a great deal of attention is customer self-service. It seems like an obvious choice for cognitive computing and has interesting ramifications. Content can be put into task context when a user is perusing a product list or shopping in a store. Salespeople ask, “How can I help you?” and online chat agents (live or automated) ask the same question. Content needs to be associated with a task or product purchase path, and “help” is what users ask for when they cannot locate what they need or find the answer while on their journey.

 

An intelligent virtual assistant (IVA) is a mechanism for accessing content in a conversational manner where the assistant can provide a specific answer, rather than a long list of responses. Cognitive computing can help optimize performance of virtual assistants, and although some aspects of the technology may seem out of reach, a number of them will be familiar to knowledge practitioners.

Foundational components of cognitive computing

Development of scenarios and use cases: It is important to define the problem to be solved and the boundaries of the problem space—general-purpose artificial intelligence is a challenging problem to solve. The most successful virtual assistant applications target narrow problems where the use cases can be articulated and the process for developing the solution to user problems can be laid out in a predictable fashion. Help requests that are repeatable, or complex but predictable processes that follow a set of procedures, can be structured for access by an intelligent system. The system can then provide the input to the virtual assistant, which is the interface between the customer who is seeking an answer and the underlying knowledgebase.

Knowledgebase, corpus and data sources: Cognitive computing applications operate on a knowledgebase, corpus or content source potentially with other sources of structured and unstructured data. Content can range from highly curated and structured to unstructured information sources such as call transcripts, chat logs and user-generated content in discussion forums and postings. Data can be transactional data, customer attributes and profiles and other sources of structured data, or can be less well formed click streams or streaming sensor data. To process those sources, a degree of curation, while not mandatory with some tools, will improve the effectiveness of the application.

Domain-specific terminology: While many tools contain concept relationships developed over years of development and evolution of the tool, domain-specific terminology is typically layered on top of those vocabularies to help identify the core concepts, issues, products, problems or solutions specific to the company or industry. Those specialized vocabularies are used to improve identification of themes, facts and features.

Feature definition: Features are characteristics of data sets or content that allow for identification of patterns. Patterns are used to either match content (e.g., for a recommendation engine) or make predictions (e.g., about a purchase motivated by a promotion). Features can also be used to reduce the volume of data for processing (imagine putting data in buckets and processing the buckets) and to develop training sets with which to build hypotheses (consider default rates on credit cards, household income and debt load—correlating those variables can predict a borrower’s chance of default).

Features can be defined in advance, and they also emerge from the data. When features are predefined, they can be analogous to facets and attributes. For example, features can be represented by clusters or groupings of content for identification of types of questions or complaints from a customer service or voice of the customer application. The theme can then be used to narrow a selection of potential responses from a broader set of use cases. A customer support analyst can define the categories of questions and problems that can then be used to group both issues and potential solutions.

Dealing with ambiguity

One advantage of using cognitive computing is that it deals more effectively with ambiguity than typical information access approaches. A thesaurus structure used in query expansion (“SOW” being mapped as equivalent to the term “proposal” would allow someone searching on either term to return documents tagged with or containing the other) deals with situations where multiple terms can mean the same thing. When one term can have multiple meanings, we need to disambiguate use of term to understand what the user intended. Ambiguity requires an interpretation based on contextual clues and therefore the answers are not absolute, but instead assessed as more or less likely. When terms have multiple meanings, one mechanism to determine the intended meaning is through natural language processing of the user query. In an insurance processing application, a user could type any of the following phrases and “claim” would have a different meaning in each case:

  • “The adjuster claimed that my house was damaged by rain.”
  • “I want to file a claim for damage to my house.”
  • “My ex-wife has a claim to my house and we need to have it insured. How can we divide responsibility for payment?”

In the first case, the adjuster made a statement that the homeowner disputes; in the second, an insured individual wants to submit information about damage; and in the third, the concept of claim is related to a right in a property. Those differences require processing to interpret the intent of the user.

Confidence based on contextual clues

The system may return different responses for each of the queries with different confidence levels about the appropriate answer. In comparing answers, a cognitive computing system uses a confidence score based on the strength of the input signals. Algorithms approach problems from multiple perspectives and paths in order to converge on the “correct” (actually, the most likely) answer.

One algorithm might process a query by comparing it to crowd-sourced variations in the question to find the closest match. Another might arrive at a normalized query by performing part of speech analysis and word sense disambiguation. In each case, the quality of the match can be ascertained and given a relative score—one algorithm might have different confidence levels for responses, or multiple algorithms might propose candidates with each providing a confidence level. That use of a confidence score is an example of a probabilistic approach.

Learning mechanisms

One of the important features of cognitive computing is a learning mechanism. Cognitive computing systems can learn both from automated process and from human-curated and human-assisted processes. Feedback loops for continual improvement can be based on input from user paths and behaviors: whether they click on results, execute another query or “pogo stick” out from the content once they have clicked, rephrase their question or ask for live help. Those metrics form the performance measures for manual curation and user experience adjustments as well as the inputs for automated machine learning and hybrid approaches.

 

Increasing the size of sampling data sets (providing more occurrences of variations in how a question is phrased, for example) can also provide a mechanism for the system to learn and improve over time. The association of an input (the query) with the output (response) and the user’s subsequent behavior (continuing to the next step or contacting customer service) will provide feedback to the scoring algorithm to adjust confidence the next time a similar query is received.

Extending established approaches

Cognitive computing builds on established approaches such as content curation, metrics-driven governance and sound information architecture. With the addition of natural language processing and learning algorithms and careful use of crowd-sourced elements, intelligent virtual assistants will become pervasive sooner than most people might think.

As additional signals are incorporated as inputs into algorithms, machine learning can begin to determine which signals improve outcomes. Perhaps features extracted from a user’s purchase history can add a signal to improve search result accuracy. A human analyst can add the data set and allow machine learning to detect relevant patterns. The pattern may be too subtle to notice or may emerge from a combination of dozens of elements. That correlation can go on behind the scenes and be used to impact future interactions.

Those aspects of cognitive computing learning mechanisms help intelligent virtual assistants improve through the ability to process information that is large in volume, complex in nature and heterogeneous.

Pattern matching, user profiles and cross-sell recommendations

The same matching algorithm that builds a feature profile of a user for help queries might also apply that feature profile to cross-sell recommendations. That information, when integrated into a comprehensive profile of the customer, might prompt an offer after the help query is resolved. The system begins to understand the customer’s goals and patterns and filters answers by context, considering the customer’s needs and interests. It finds patterns in the data that are both expected and unexpected, across many different kinds of use cases.

Virtual assistants built on cognitive computing technology will be used to make sophisticated recommendations based on higher level needs of users. Perhaps an investor wants to make a stock purchase but changes in the market causes him or her to reconsider. A virtual assistant can understand risk tolerance, investment objectives and market conditions, and then provide a range of alternatives based on human expertise embedded in the system but also correlated with real-time inputs.

Cognitive computing systems build a constantly evolving understanding of users and their goals and behaviors based on continually processing inputs. They learn, build their knowledgebase and develop mechanisms for understanding user queries. They continually fine-tune their outputs. They start with a clearly defined problem and content to solve that problem and then evolve ways of interpreting human interactions and providing responses to help people achieve their goals.

A definition of cognitive computing

Cognitive computing is an overarching concept that describes multiple approaches to making computers more effective at interpreting human interactions and making sense of large amounts of information. It includes various flavors of artificial intelligence and machine learning, including understanding sources of information that have typically been in the realm of human perception—including speech recognition and image identification and interpretation. Cognitive computing deals with ambiguous information and problem solving and can use multiple approaches to dealing with challenges that have defied structured computation up until now.

Earley Information Science Team
Earley Information Science Team
We're passionate about enterprise data and love discussing industry knowledge, best practices, and insights. We look forward to hearing from you! Comment below to join the conversation.

Recent Posts

Designing AI Programs for Success - a 4 Part Series

Recorded - available as on demand webcast AI is plagued by inflated and unrealistic expectations due to a lack of broad understanding of this wide-ranging space by software vendors and customers. Software tools can be extremely powerful, however the services, infrastructure, data quality, architecture, talent and methodologies to fully deploy in the enterprise are frequently lacking. This four-part series by Earley Information Science and Pandata will explore a number of issues that continue to afflict AI projects and reduce the likelihood of success. The sessions will provide actionable steps using proven processes to improve AI program outcomes.

The Missing Ingredient to Digital Transformation: Scaling Knowledge Communities and Processes

The holy grail of digital transformation is the seemingly conflicting goals of high levels of customer service and pressure to reduce costs. “Digital Transformation” has become an all-encompassing term – in a piece in this column about customer data platforms, I asked whether the term has lost its meaning: The phrase “digital transformation” can mean anything and everything — tools, technology, business processes, customer experience, or artificial intelligence, and every buzzword that marketers can come up with. Definitions from analysts and vendors include IT modernization and putting services online; developing new business models; taking a “digital first” approach; and creating new business processes, and customer experiences. The overarching objective of a digital transformation program is to improve end-to-end efficiencies, remove friction from information flows, and create new value streams that differentiate a company’s offerings and strengthen the customer relationship. Having assisted large global enterprises with building the data architecture, supporting processes, and governance for multiple digital transformations, in my experience, there are two broad classes of initiatives that seem to get funding and others that miss the boat in terms of time, attention, and resources.

4 Reasons B2B Manufacturers need Strong Product Data

There are many manufacturers who have started to take the leap forward in the digital space, but there are still a great number who rely solely on their distributors to manage their product data. We are going to look at 4 key reasons why its so important that manufacturers own their product and dedicate the time and resources to build it out.