Back to blog
Articles
May 10, 2023
·
3 MIN READ

NLU & NLG Should Go Hand-In-Hand

May 10, 2023
|
3 MIN READ

Latest content

Tutorials
5 min read

Optimizing RAG with Knowledge Base Maintenance

How to find gaps between knowledge base content and real user questions.
April 23, 2024
Tutorials
4 min read

Scaling Quality Assurance with HumanFirst and Google Cloud

How to use HumanFirst with Vertex AI to test, improve, and trust agent performance.
March 14, 2024
Announcements
2 min read

Full Circle: HumanFirst Welcomes Maeghan Smulders as COO

Personal and professional history might not repeat, but it certainly rhymes. I’m thrilled to join the team at HumanFirst, and reconnect with a team of founders I not only trust, but deeply admire.
February 13, 2024
Tutorials
4 min read

Accelerating Data Analysis with HumanFirst and Google Cloud

How to use HumanFirst with CCAI-generated data to accelerate data analysis.
January 24, 2024
Tutorials
4 min read

Exploring Contact Center Data with HumanFirst and Google Cloud

How to use HumanFirst with CCAI-generated data to streamline topic modeling.
January 11, 2024
Articles
5 min

Building In Alignment: The Role of Observability in LLM-Led Conversational Design

Building In Alignment: The Role of Observability in LLM-Led Conversational Design
December 6, 2023
Articles
5 min read

Rivet Is An Open-Source Visual AI Programming Environment

Rivet is suited for building complex agents with LLM Prompts, and it was Open Sourced recently.
September 27, 2023
Articles
6 min read

What Is The Future Of Prompt Engineering?

The skill of Prompt Engineering has been touted as the ultimate skill of the future. But, will prompt engineering be around in the near future? In this article I attempt to decompose how the future LLM interface might look like…considering it will be conversational.
September 26, 2023
Articles
4 min read

LLM Drift

A recent study coined the term LLM Drift. LLM Drift is definite changes in LLM responses and behaviour, over a relatively short period of time.
September 25, 2023
Tutorials
5 min read

Optimizing RAG with Knowledge Base Maintenance

How to find gaps between knowledge base content and real user questions.
April 23, 2024
Tutorials
4 min read

Scaling Quality Assurance with HumanFirst and Google Cloud

How to use HumanFirst with Vertex AI to test, improve, and trust agent performance.
March 14, 2024
Announcements
2 min read

Full Circle: HumanFirst Welcomes Maeghan Smulders as COO

Personal and professional history might not repeat, but it certainly rhymes. I’m thrilled to join the team at HumanFirst, and reconnect with a team of founders I not only trust, but deeply admire.
February 13, 2024

Let your data drive.

NLU & NLG Should Go Hand-In-Hand

COBUS GREYLING
May 10, 2023
.
3 MIN READ

Traditional NLU Can Be Leveraged By Following A Hybrid NLU & NLG Approach

This article considers how Foundation LLMs can be used to leverage existing NLU investments, and improve chatbot experiences.

Considering the Conversational AI landscape and the inroads LLMs are making, there has been a few market phenomenons:

  1. Traditional NLU based chatbot frameworks are adopting Foundation LLM functionality. But mostly in an unimaginative way to generate intent training examples, bot response fallback dialogs or rewriting bot messages. The only exception here is Cognigy, and to some degree Yellow AI.
  2. LLMs are mostly being used in a generative capacity, and not in conjunction with existing predictive capability.
  3. The predictive power of traditional NLU engines with regard to intent classification should not be overlooked.
  4. Hybrid NLU and LLM based NLG implementations are not receiving the consideration in deserves.
Leverage the best of NLU, which is intent detection, with the most accessible LLM feature, which is response generation.

This article considers the following:

  • Increasing the granularity and sophistication of existing inbound user utterance NLU models, by complimenting it with NLG.
  • Using human supervision to monitor generative responses in the same way intents are monitored and maintained.
  • Engineered and curated LLM prompts can be used to create organisation specific fine-tuned LLM models.

Recently Stephen Broadhurst and Gregory Whiteside demonstrated an effective hybrid NLU/NLG approach which combines traditional intent-based logic with dynamic LLM responses.

When it comes to supervision hints, there are three key principles to keep in mind:

  1. Inbound supervision is typically more efficient than outbound supervision.
  2. Light supervision can be a great way to create bot messages that are both responsive and contextual.
  3. Highly directive guidance, while reliable, can lead to more restricted replies.

Consider the following scenario…

When a user utters a statement, it is usually intercepted by a chatbot (1).

After the statement is intercepted, the intent of the statement is retrieved (2).

Then, example prompts or hints associated with the intent are retrieved (3).

These hints help the Large Language Model (LLM) generate accurate responses without any hallucination (4).

In order to keep the hints and prompts up to date, they can be reviewed and maintained, though not in real-time (5).

The hints are sent to the LLM with relevant entities injected via templating, a process known as prompt engineering (6).

The LLM generates a response (7) and sends it back to the user.

The NLU/NLG trainers review the generative results and update the prompt templates daily, in order to give better hints and templating (8).

Below is a detailed diagram of a supervised generative bot process flow:

In Conclusion

It is clear that the successful and established architecture of chatbots and the advanced capabilities of NLU engines should be utilised through a hybrid approach.

This is because companies have consistently demonstrated a tendency to converge around the most effective ideas, as evidenced by the current chatbot landscape.

I’m currently the Chief Evangelist @ HumanFirst. I explore and write about all things at the intersection of AI and language. Including NLU design, evaluation & optimisation. Data-centric prompt tuning and LLM observability, evaluation and fine-tuning.

Subscribe to HumanFirst Blog

Get the latest posts delivered right to your inbox