All Posts

Is IBM Watson Technology Practical for the Enterprise?

This Articles was published online at Baseline Magazine on May 3, 2011.

IBM’s Jeopardy-playing Watson computer has been hailed as a technology triumph – the ability of computers to understand human language and broad knowledge topics – not just facts and trivia but ambiguous language including puns, double entendre’s and idioms.

The technology is impressive and IBM has set its sights on many commercial applications in healthcare, financial services and customer service operations.   Few organizations have the resources it took to build Watson - $3mm worth of hardware (off the shelf servers with almost 3000 processors and 1 terabyte of RAM) – not to mention millions in research.  Nevertheless, the question remains, does Watson embed a solution approach that enterprises can exploit or learn from?  How readily can a “Watson” be applied to the knowledge and content access problems of the typical enterprise? 

A few clues lie in the nature of knowledge access and in some of the challenges that Watson team members discussed in Articless and interviews.   First, here are some principles that Watson exploited:

  • Watson used multiple algorithms to process information.  These included the usual keyword matching algorithms of run-of-the-mill search, “temporal” (time based) reasoning that understand dates and relative time calculations, “statistical paraphrasing” an approach to convey ideas using different words, and “geospatial reasoning” – a way of interpreting locations and geographies, and various approaches to unstructured information processing.    
  •  Watson can be characterized as “semantic search” or natural language search.  That is, questions are asked in plain English as opposed to a structured query and this question is parsed into its semantic and syntactic (meaning and grammatical structure) components.   The parsed question is then processed against the system’s knowledge base derived  from over 200 million pages of information.

Other than keyword matching which parses terms and processes them against a dumb bag of words, more complex and powerful approaches require an underlying structure to the information. These structures are in the form of taxonomies and ontologies which tell the system how concepts relate to one another.  Many organizations are beginning to build these taxonomy frameworks for purposes of e commerce, document management, intranet and knowledge base applications.

The second point is that Watson demonstrates key elements of solutions that do not assume that users know exactly how to frame questions regarding what they want.   As much research on search shows,  more often than not users  ask ambiguous questions and expect precise results.  Therefore we need to build solution that help them with the queries.  These are the same approaches for structuring the information in the first place (the structures that the tools require to make sense of the data are the same ones that help guide users in their choices). Think of the new navigation/search approaches used in ecommerce sites – choosing color, size, brand, price, etc. help users find what they need and precisely navigate to specific information. 

Bottom line is that tools like IBM’s Watson are a great leap forward in capabilities, but there is no free lunch – Watson’s power comes from organizing content.  Tools for gaining insights and finding answers will get better as time goes on, but human judgment needs to be applied to information to develop a foundation of meaning and structure.

Seth Earley
Seth Earley
Seth Earley is the Founder & CEO of Earley Information Science and the author of the award winning book The AI-Powered Enterprise: Harness the Power of Ontologies to Make Your Business Smarter, Faster, and More Profitable. An expert with 20+ years experience in Knowledge Strategy, Data and Information Architecture, Search-based Applications and Information Findability solutions. He has worked with a diverse roster of Fortune 1000 companies helping them to achieve higher levels of operating performance.

Recent Posts

First Party Data: The New Imperative

The need for accurate data to support digital transformation and the emergence of new restrictions on the use of third-party data have prompted many companies to focus their attention on first party data.

Knowledge Graphs, a Tool to Support Successful Digital Transformation Programs

Knowledge graphs are pretty hot these days. While this class of technology is getting a lot of market and vendor attention these days, it is not necessarily a new construct or approach. The core principles have been around for decades. Organizations are becoming more aware of the potential of knowledge graphs, but many digital leaders are puzzled as to how to take the next step and build business capabilities that leverage this technology.

[RECORDED] Powering Personalized Search with Knowledge Graphs

Transforming Legacy Faceted Search into Personalized Product Discovery The latest in e-commerce trends is the transformation of legacy faceted search into a more personalized experience. By applying semantic reasoning over a knowledge graph, contextual information about a customer can be combined with product data, delivering relevant search results tailored to them. The first half of this webinar is designed for the business executive. We’ll focus on why personalized search is an essential e-commerce ingredient. And we’ll demystify the process of implementing a more personalized product discovery experience for your customers. The second half of the webinar is designed for the data strategist. We’ll cover the data modeling required to build knowledge graphs for successful personalized search. We’ll discuss some real-world cases and cover the steps you can take to get started. Who should attend: Executives who care about e-commerce and the data experts who enable them. Speakers: