Back to blog
Articles
May 11, 2023
·
4 MIN READ

Prompt Engineering, OpenAI & Modes

May 11, 2023
|
4 MIN READ

Latest content

Tutorials
5 min read

Optimizing RAG with Knowledge Base Maintenance

How to find gaps between knowledge base content and real user questions.
April 23, 2024
Tutorials
4 min read

Scaling Quality Assurance with HumanFirst and Google Cloud

How to use HumanFirst with Vertex AI to test, improve, and trust agent performance.
March 14, 2024
Announcements
2 min read

Full Circle: HumanFirst Welcomes Maeghan Smulders as COO

Personal and professional history might not repeat, but it certainly rhymes. I’m thrilled to join the team at HumanFirst, and reconnect with a team of founders I not only trust, but deeply admire.
February 13, 2024
Tutorials
4 min read

Accelerating Data Analysis with HumanFirst and Google Cloud

How to use HumanFirst with CCAI-generated data to accelerate data analysis.
January 24, 2024
Tutorials
4 min read

Exploring Contact Center Data with HumanFirst and Google Cloud

How to use HumanFirst with CCAI-generated data to streamline topic modeling.
January 11, 2024
Articles
5 min

Building In Alignment: The Role of Observability in LLM-Led Conversational Design

Building In Alignment: The Role of Observability in LLM-Led Conversational Design
December 6, 2023
Articles
5 min read

Rivet Is An Open-Source Visual AI Programming Environment

Rivet is suited for building complex agents with LLM Prompts, and it was Open Sourced recently.
September 27, 2023
Articles
6 min read

What Is The Future Of Prompt Engineering?

The skill of Prompt Engineering has been touted as the ultimate skill of the future. But, will prompt engineering be around in the near future? In this article I attempt to decompose how the future LLM interface might look like…considering it will be conversational.
September 26, 2023
Articles
4 min read

LLM Drift

A recent study coined the term LLM Drift. LLM Drift is definite changes in LLM responses and behaviour, over a relatively short period of time.
September 25, 2023
Tutorials
5 min read

Optimizing RAG with Knowledge Base Maintenance

How to find gaps between knowledge base content and real user questions.
April 23, 2024
Tutorials
4 min read

Scaling Quality Assurance with HumanFirst and Google Cloud

How to use HumanFirst with Vertex AI to test, improve, and trust agent performance.
March 14, 2024
Announcements
2 min read

Full Circle: HumanFirst Welcomes Maeghan Smulders as COO

Personal and professional history might not repeat, but it certainly rhymes. I’m thrilled to join the team at HumanFirst, and reconnect with a team of founders I not only trust, but deeply admire.
February 13, 2024

Let your data drive.

Prompt Engineering, OpenAI & Modes

COBUS GREYLING
May 11, 2023
.
4 MIN READ

What role can prompt engineering play in preventing LLM hallucination, and what constitutes a good LLM prompt? Furthermore, how are OpenAI's models impacting this?

Intro

Prompt engineering and supervision are essential for successful implementation of Large Language Models (LLMs).

LLMs are known for providing varied responses, giving the feeling of being "alive". However, these variations can become excessive, resulting in false and inaccurate data, which is often referred to as hallucination.

To prevent this, prompt engineering and supervision can be improved through accurate prompt compilation, prompt structure, and observability.

Supervision & Observability

Stephen Broadhurst presented a very good talk recently at the European Chatbot Summit on supervision and observability from a LLM perspective.

You can read more about the basic principles here.

Accurate Prompt Compilation

A real-time conversational interface, such as a chatbot or voicebot, can be improved using a few-shot approach that provides the language model (LLM) with enough context and guidance to give accurate responses within the current dialog context.

The difficulty lies in quickly and accurately compiling the prompt from various data sources and submitting it to the LLM.

You can read more about the architecture for such an implementation here.

Prompt Structure

Prompt structure and compilation is of utmost importance. You can read more about basic prompt engineering principles here.

Text Generation Is A Meta Capability Of Large Language Models & Prompt Engineering Is Key To Unlocking It. You cannot talk directly to a Generative Model, it is not a chatbot. You cannot explicitly request a generative model to do something. But rather you need a vision of what you want to achieve and mimic the initiation of that vision. The process of mimicking is referred to as prompt design, prompt engineering or casting.

Prompts need to hold context and be contextually accurate. A good basic structure is shown below:

Subscribe to HumanFirst Blog

Get the latest posts delivered right to your inbox