Back to blog
Articles
May 4, 2023
·
4 MIN READ

Example Code & Implementation Considerations For GPT 3.5 Turbo, ChatML & Whisper

May 4, 2023
|
4 MIN READ

Latest content

Tutorials
5 min read

Optimizing RAG with Knowledge Base Maintenance

How to find gaps between knowledge base content and real user questions.
April 23, 2024
Tutorials
4 min read

Scaling Quality Assurance with HumanFirst and Google Cloud

How to use HumanFirst with Vertex AI to test, improve, and trust agent performance.
March 14, 2024
Announcements
2 min read

Full Circle: HumanFirst Welcomes Maeghan Smulders as COO

Personal and professional history might not repeat, but it certainly rhymes. I’m thrilled to join the team at HumanFirst, and reconnect with a team of founders I not only trust, but deeply admire.
February 13, 2024
Tutorials
4 min read

Accelerating Data Analysis with HumanFirst and Google Cloud

How to use HumanFirst with CCAI-generated data to accelerate data analysis.
January 24, 2024
Tutorials
4 min read

Exploring Contact Center Data with HumanFirst and Google Cloud

How to use HumanFirst with CCAI-generated data to streamline topic modeling.
January 11, 2024
Articles
5 min

Building In Alignment: The Role of Observability in LLM-Led Conversational Design

Building In Alignment: The Role of Observability in LLM-Led Conversational Design
December 6, 2023
Articles
5 min read

Rivet Is An Open-Source Visual AI Programming Environment

Rivet is suited for building complex agents with LLM Prompts, and it was Open Sourced recently.
September 27, 2023
Articles
6 min read

What Is The Future Of Prompt Engineering?

The skill of Prompt Engineering has been touted as the ultimate skill of the future. But, will prompt engineering be around in the near future? In this article I attempt to decompose how the future LLM interface might look like…considering it will be conversational.
September 26, 2023
Articles
4 min read

LLM Drift

A recent study coined the term LLM Drift. LLM Drift is definite changes in LLM responses and behaviour, over a relatively short period of time.
September 25, 2023
Tutorials
5 min read

Optimizing RAG with Knowledge Base Maintenance

How to find gaps between knowledge base content and real user questions.
April 23, 2024
Tutorials
4 min read

Scaling Quality Assurance with HumanFirst and Google Cloud

How to use HumanFirst with Vertex AI to test, improve, and trust agent performance.
March 14, 2024
Announcements
2 min read

Full Circle: HumanFirst Welcomes Maeghan Smulders as COO

Personal and professional history might not repeat, but it certainly rhymes. I’m thrilled to join the team at HumanFirst, and reconnect with a team of founders I not only trust, but deeply admire.
February 13, 2024

Let your data drive.

Example Code & Implementation Considerations For GPT 3.5 Turbo, ChatML & Whisper

COBUS GREYLING
May 4, 2023
.
4 MIN READ

A while ago OpenAI has released the API for The LLM gpt-3.5-turbo, the same model used in ChatGPT. Additionally, the Whisper speech-to-text large-v2 model is available through an API for transcription.

GPT 3.5 Turbo & Chat Markup Language (ChatML)

The ChatGPT models are accessible via API, with gpt-3.5-turbo being used in the examples below. OpenAI also has a model named gpt-3.5-turbo-0301.

For up-to-date model information, the OpenAI model page is a great resource.

It is important to remember that OpenAI models are non-deterministic, meaning that the same input given at different times or even in a row can lead to different or varying results.

OpenAI states that setting the temperature to 0 will make the output mostly deterministic, but some variability may still exist. The ChatGPT web interface we are used to is very successful in managing conversational context.

ChatGPT web interface by OpenAI

It is important to note that the ChatGPT model (gpt-3.5-turbo) accessed through the API does not keep track of conversational context, as can be seen in the example below:

The ChatGPT model is able to maintain conversational context with a few-shot approach by buffering the prompts, similar to the way OpenAI managed context via the initial web interface.

An example of this few-shot learning prompt in action is shown below, with a very contextual and empathetic response from the ChatGPT model:

Chat Markup Language (ChatML) is an example JSON file that defines the roles of system, user, and assistant.

It is designed to protect against prompt injection attacks, which are the main security vulnerability and avenue of abuse for LLMs.

Below is a working example of a completion request sent to the gpt-3.5-turbo model using the ChatML file. The following Python code snippet can be run in a Colab Notebook:

Notice the role which is defined, the model detail which is gpt-3.5-turbo-0301, and other more in the output from the completion request below.

OpenAI Whisper large-v2 Model

Considering accessing the OpenAI Whisper AI via a Colab Notebook:

The result from uploading the MP3 audio file.

The lines of Python code to transcribe the audio:

And below is the output result…

I find it interesting that Whisper is able to detect the language of the recording before transcribing it.

According to the available Whisper models, languages and Word Error Rates (WER), Spanish has the best rate of 3, followed by Italian with a WER of 4, and English with a WER of 4.2. You can read more about it here.

I’m currently the Chief Evangelist @ HumanFirst. I explore and write about all things at the intersection of AI and language; ranging from LLMs, Chatbots, Voicebots, Development Frameworks, Data-Centric latent spaces and more.

Subscribe to HumanFirst Blog

Get the latest posts delivered right to your inbox