Back to blog
Articles
Articles
July 21, 2023
·
6 min read

OpenAI GPT Chat Completions Accounts For 97% Of Usage

July 21, 2023
|
6 min read

Latest content

Tutorials
5 min read

Optimizing RAG with Knowledge Base Maintenance

How to find gaps between knowledge base content and real user questions.
April 23, 2024
Tutorials
4 min read

Scaling Quality Assurance with HumanFirst and Google Cloud

How to use HumanFirst with Vertex AI to test, improve, and trust agent performance.
March 14, 2024
Announcements
2 min read

Full Circle: HumanFirst Welcomes Maeghan Smulders as COO

Personal and professional history might not repeat, but it certainly rhymes. I’m thrilled to join the team at HumanFirst, and reconnect with a team of founders I not only trust, but deeply admire.
February 13, 2024
Tutorials
4 min read

Accelerating Data Analysis with HumanFirst and Google Cloud

How to use HumanFirst with CCAI-generated data to accelerate data analysis.
January 24, 2024
Tutorials
4 min read

Exploring Contact Center Data with HumanFirst and Google Cloud

How to use HumanFirst with CCAI-generated data to streamline topic modeling.
January 11, 2024
Articles
5 min

Building In Alignment: The Role of Observability in LLM-Led Conversational Design

Building In Alignment: The Role of Observability in LLM-Led Conversational Design
December 6, 2023
Articles
5 min read

Rivet Is An Open-Source Visual AI Programming Environment

Rivet is suited for building complex agents with LLM Prompts, and it was Open Sourced recently.
September 27, 2023
Articles
6 min read

What Is The Future Of Prompt Engineering?

The skill of Prompt Engineering has been touted as the ultimate skill of the future. But, will prompt engineering be around in the near future? In this article I attempt to decompose how the future LLM interface might look like…considering it will be conversational.
September 26, 2023
Articles
4 min read

LLM Drift

A recent study coined the term LLM Drift. LLM Drift is definite changes in LLM responses and behaviour, over a relatively short period of time.
September 25, 2023
Tutorials
5 min read

Optimizing RAG with Knowledge Base Maintenance

How to find gaps between knowledge base content and real user questions.
April 23, 2024
Tutorials
4 min read

Scaling Quality Assurance with HumanFirst and Google Cloud

How to use HumanFirst with Vertex AI to test, improve, and trust agent performance.
March 14, 2024
Announcements
2 min read

Full Circle: HumanFirst Welcomes Maeghan Smulders as COO

Personal and professional history might not repeat, but it certainly rhymes. I’m thrilled to join the team at HumanFirst, and reconnect with a team of founders I not only trust, but deeply admire.
February 13, 2024

Let your data drive.

Articles

OpenAI GPT Chat Completions Accounts For 97% Of Usage

COBUS GREYLING
July 21, 2023
.
6 min read

Users are moving from text completion to chat completion.

The initial Completions API, introduced in June 2020, provided a freeform text prompt for interacting with OpenAI’s language models.

However, OpenAI soon realised that a more structured prompt interface could produce far better results.

Chat-based paradigms have proven to be particularly powerful, handling the vast majority of use cases, while offering higher flexibility and specificity.

Considering the image below, there has been a change to the Mode section of the OpenAI playground.

The insert mode has been removed while the Complete and Edit modes are marked as legacy.

The Chat Completions API’s structured interface and multi-turn conversation capabilities (e.g., system messages, function calling) enable developers to build varied conversational experiences and completions tasks.

Furthermore, this structure helps protect against prompt injection attacks, as user-provided content can be kept separated from instructions.

The Chat Completions API’s structured interface and multi-turn conversation capabilities (e.g., system messages, function calling) enable developers to build varied conversational experiences and completions tasks.

Furthermore, this structure helps protect against prompt injection attacks, as user-provided content can be kept separated from instructions.

OpenAI has expressed their intention to focus the majority of platform development efforts to chat completions.

Here are a few complete working code examples for Text Summarisation, Code Completion & Few-Shot Learning.

Text Summarisation Code

Text Summarisation Result:

Code Completion Code

Code Completion Result:

Few-Shot Learning Code

Few-Shot Learning Result:

The last update to Chat Markup Language (ChatML) was four months ago, and I would expect an enhancement soon.

It is clear that OpenAI wants to introduce structure to the format of input data. With the ChatML conversational format extending to generative AI tasks which have been very open in the past, like summarisation, code completion and few-shot contextual prompts.

This structure being imposed from a model level, will have to be absorbed by downstream implementations like autonomous agents, prompt chaining, etc.

As LLM-based generative AI apps can make use of multiple LLMs, the propagation of Chat Completion will surely have benefits, but will surely introduce complexity.

I’m currently the Chief Evangelist @ HumanFirst. I explore & write about all things at the intersection of AI and language; ranging from LLMs, Chatbots, Voicebots, Development Frameworks, Data-Centric latent spaces & more.

Subscribe to HumanFirst Blog

Get the latest posts delivered right to your inbox