Interesting

Natural Language Question Answering Systems – Get Quick Answers To Concrete Questions

I’m sure you’ve also spent loads of time searching for an answer to a specific question among the sheer wealth of information available in our digital world. Finding what we need is becoming ever more challenging and time-consuming. Whether in the vastness of the World Wide Web or in the tangled web of corporate data and information systems, a keyword search often only narrows down the search. You then need to filter through a mass of search results to find what you are looking for. However, there’s help at hand. Discover in this article what a Natural Language Question Answering system is, and how it makes it easier to find answers in the data jungle.

As digitalisation progresses, the volume of data worldwide is also rapidly growing. The statistics portal Statista estimates that in 2021, 79 zettabytes of data will have been created, captured and consumed. They expect this number to more than double within 4 years, forecasting 181 zettabytes of data for 2025[1].

Companies using the SEEBURGER Business Integration Suite (BIS) are able to process huge amounts of data every day. This may be from e-invoices, API calls or in other protocols and formats. BIS can communicate between various systems using this structured data.

However, in addition to structured data, there is more and more unstructured data being generated. This includes documents, blog articles, voice messages and videos. Much of this unstructured data contains interesting information on a wide range of topics. However, finding the answer to a specific question you may have on these topics is neither quick nor easy. The number of these documents and files just keeps growing! Classic search engines such as Google, Bing, etc. do help you to find certain documents on the web, but actually sifting through the wealth of documents for specific details is often left to the user.

A new approach is needed to make it quicker and easier to search and find the information end users need. Let us introduce you to natural language question answering systems.

What is a natural language question answering system?

Unlike traditional keyword searches, a Natural Language Question Answering system does not return a complete document to the user. Instead, users ask a question in natural language and receive a specific answer in return.

Figure 1: A natural language question answering system takes questions in natural language and returns a concrete answer

A user who types in “What is the capital of Germany?” not only receives a web page with related information, but also the concrete answer “Berlin”. This saves the user a lot of time in finding the answer, especially if he would otherwise have to trawl through particularly long documents.

How does a natural language question answering system work?

Developing question answering systems has been a hot topic in IT for quite some time. Earlier, there were attempts to set up complex rules to enable a system to understand a user’s naturally worded question and to provide an answer.

The natural language question answering systems of today often take an extractive approach, and consist of a retriever and a reader. The answers to the questions are not stored in large databases, rather the system attempts to find and extract an appropriate answer to a user’s question from a mass of texts. Firstly, documents relevant to the user’s question are loaded from a document store. Then, the reader attempts to extract the answer to the user’s question.

Figure 2: The roles of indexer, document store, retriever and reader in a natural language question answering system

What is a document store?

The document store is responsible for providing relevant documents. There are different ways of doing this. Often, a simple or reverse index is used. One well-known, commonly used such reverse index is Apache Lucene, which is used as an index in Elasticsearch. It’s a fast, efficient way to call up documents.

What is a retriever?

A retriever is responsible for retrieving relevant documents for the user’s question. First of all, it attempts to extract the relevant terms in the question. It then uses these to retrieve relevant documents.

In order to turn a user’s question into the type of query a retriever can process, various Natural Language Processing (NLP) techniques are used. These include:

  1. Removing punctuation
    Full stops, commas and other punctuation are superfluous in retrieving relevant documents. These are therefore removed from the user’s question.
  2. Removing stopwords
    Stopwords are commonly occurring words which don’t have a significant impact on content. Examples include articles such as ‛the’, ‛a’, ‛an’. These words are therefore filtered out.
  3. Tagging entities
    Entities, such as products or names, are usually very relevant to the query. These are therefore incorporated into the query.
  4. Stemming
    Words can appear in different forms or conjugations (walk, walked, walking, etc.). As they may well appear in different forms within in a document, such words are reduced to a base form before being incorporated into the query.

These steps help create a query, which is then made to the document store. The retriever grabs the most relevant documents and passes these on to the reader to extract an answer to the user’s question.

Figure 3: The BERT model in a natural learning question answering system

What is a reader in a question answering system?

A reader is responsible for extracting an answer from the documents it receives. Using a suitable language model, it tries to understand both the question and the documents and to extract the most appropriate answer from the texts.

What is a language model?

A language model is used to extract answers from the texts submitted. One such model is the frequently used BERT model illustrated in Figure 2, however other language models are also available. A language model has been trained to calculate the probability of a word or phrase occurring.

There are several different types of language model. The quality of their results depends on the areas in which they are employed. Many of these models are take a transformer or attention based approach which enables them to comprehend and process language.

How does the BERT model work?

Devlin et al‘s Bidirectional Encoder Representations from Transformers, commonly known as BERT, is a powerful language model. It builds on the encoder from Vaswani et al‘s Transformers and attention mechanism.

It starts by dividing the input into individual words, known as tokens. These are augmented with further BERT-specific tokens and entered into the model. For a Question Answering scenario, the input consists of both the question and the paragraph from which the answer is to be extracted.

There are multiple layers of transformers between the input and output of the model, which are used to compute the answer.

The model outputs start and end tokens to determine the best answer. These are then used to help return an answer to the user.

At this stage, the model is not yet able to compute meaningful answers itself. To make this possible, the model needs to be trained. Initially, the model was trained to have general reading comprehension. This enables it to be used in further areas. To this end, the BERT model was pre-trained using the over 800 million words in BooksCorpus, as well as the English version of Wikipedia. For the latter, the model was only trained on the main body of the individual articles, encompassing over 2.5 billion words. The model was able to learn which words or phrases are dependent on or commonly used with which others, and can use this knowledge for other tasks. The advantage of pre-training a model is that it is later relatively easy to adapt it to a range of specific tasks – such as finding answers in paragraphs without necessarily needing to re-train it.

How the BERT model works for Natural Language Question Answering[2]

The pre-trained model was further prepared for its extractive question answering tasks using the SQuAD data set. The SQuAD data set consists of 100,000 pairs of questions and answers to selected Wikipedia articles. By working through these sample pairs, the model builds on its previously learned general language comprehension and learns to extract and return the right answers to the questions.

A pre-trained BERT model can be directly employed in a Question Answering System. However, by following up by using data sets from specialist interest areas or certain types of document, the model can be trained to deliver even better results in a certain area.

SeeQA –SEEBURGER’s Natural Learning Question Answering System

The SeeQA app is a natural language question answering tool that SEEBURGER uses in-house to retrieve information from various systems. Users enter their questions into an app and the system searches documents and blog articles for relevant words and extracts answers for the user. The most relevant answers are presented in a list, highlighted in yellow, and shown in context. This means that the user not only gets and answer to his question, he is also shown related content.

Figure 5: The SeeQA app – a natural learning question answering system from SEEBURGER

This system has helped our employees find the information they need more quickly. According to an internal survey, it has made it significantly easier to find the information required. Although the system has not always answered every question correctly, most users have been able to find information which has helped them with their issue.

An integrated feedback facility provides the system with information on how relevant answers have been to users. This feedback is collected and stored. It is then used to create new training data so the model can be improved even further. This continuous learning process helps the system better adapt to SEEBURGER-specific texts and questions.

This ongoing adaptation of our chosen model to SEEBURGER-specific texts will help us see to what extent the system can learn and improve on its own, and at what point that particular model cannot be optimized any further.

[1] Statista Inc.: Volume of data/information created, captured, copied, and consumed worldwide from 2010 to 2025 https://www.statista.com/statistics/871513/worldwide-data-created/ (accessed 13th Dec 2021)

[2] Cornell University, BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding, [Submitted on 11 Oct 2018 (v1), last revised 24 May 2019 (this version, v2)], accessed on 13th Dec 2021


Source: https://blog.seeburger.com/natural-language-question-answering-systems-get-quick-answers-to-concrete-questions/

Inline Feedbacks
View all comments
guest

Artificial Intelligence and High Intelligence – SEEBURGER and the Hector Seminar

If you came in to our Bretten offices on a Friday afternoon, you might wonder why one or...

Strengthen Security: Two-Factor Authentication (2FA) Is the Key to Increased Cyber Protection for Your Business

In our digital era, in which security is only becoming more critical to businesses, we are pleased to...

Cyber Security in the Metaverse

Challenges of futuristic business ecosystems Most people are familiar with the term “metaverse” these days, especially since Facebook’s...

NFTs are taking the world by storm. But what is an NFT?

The 20th edition of ArtReview’s Power 100, published in November 2021, had an unusual winner. For the first...

The Digital Product Passport: Driving Sustainability with Digital Transparency

Soon, EU businesses will have to adopt Digital Product Passports (DPPs) for product lifecycle transparency or face penalties....

How Do You Use the Kano Model?

Kano Series Part 2 – Surveying your customers. The Kano model lends itself particularly well to complex products...

The Evolution of the Internet – Web 1.0, Web 2.0, Web 3.0, Web 4.0

From humble beginnings to the groundbreaking innovations of today, the World Wide Web has come a long way....

Hyper-Personalization in Digital Marketing: How Machine Learning is Revolutionizing Customer Experience

The Dawn of Hyper-Personalization Era The digital marketing landscape has undergone a seismic shift from broad demographic targeting...

What Matters Most to Your New Retail Customers?

Improved replenishment strategies The looming question is “How do I address what matters most to today’s New Retail...

Human Intelligence, Artificial Intelligence: Teenager vs. Machine in SEEBURGER Enrichment Project

Machines can learn things. However, just like people, they need to be trained in their specific tasks. SEEBURGER...

Artificial Intelligence and the SEEBURGER BIS

Artificial intelligence is playing an ever greater part in our everyday lives. Intelligent systems are behind ever more...

Industry 4.0 and Lean Management – How to Manage the Chaos

Value chains are becoming increasingly complex and networked. This is not only due to globalisation, but also to...

The Digital Economy – Is Digitalization The Sixth Kondratieff Wave?

Over the last 75 years or so, digitalization has become an indispensable part of our lives, permeating nearly...

What is Digital Twin Energy Management and Why is it Important?

Disruptive technologies have taken the world by storm, and businesses have responded by undertaking digital transformation journeys in...

Extended Reality (XR) in the World of Work: Is A New Dimension of Productivity and Collaboration on the Verge of a Breakthrough?

In recent years, virtual reality (VR) and augmented reality (AR) have developed into fascinating technologies that are not...

What Is the Kano Model?

Kano series: Part 1 – the five categories, from attraction to rejection The Kano model describes the relationship...

IT Carve-Outs: How to Successfully Separate and Integrate Functional IT

Imagine if your business could operate more flexibly, efficiently and sustainably – by adapting to changing market conditions....

Better Than Crypto – What’s the Deal With the Digital Euro?

+++ Update 2025 will be a decisive year for the digital euro. The project has reached its halfway...

Backup, Disaster Recovery or Backup-As-Code — What Really Protects?

Companies face the challenge of protecting their IT systems against data loss, cyber attacks and unexpected outages. But...

What Matters Most to Your New Retail Customers? Real-Time Product Visibility!

In part one of What Matters Most to Your New Retail Customers?, the key message was making certain...