OpenAI has announced that customers can now bring custom data to the lightweight version of GPT-3.5, GPT-3.5 Turbo — making it easier to improve the text-generating AI model’s reliability while building in specific behaviors.
OpenAI also made available two updated GPT-3 base models (babbage-002 and davinci-002), which can be fine-tuned as well, with support for pagination and “more extensibility.” As previously announced, They plan to retire the original GPT-3 base models on January 4, 2024.
Businesses can now fine-tune GPT-3.5 Turbo using their own data to match or exceed the abilities of GPT-4 for certain tasks. Fine-tuning lets businesses essentially hone ChatGPT to a more focused model that’s especially efficient for certain tasks. The supervised training would make a bot that’s unique to the client company so that it offers, say, reliable responses in a specific language or with more concise wording. Until now, business customers were limited to GPT-3 variants for this, like davinci-002 or babbage-002.customers can also shorten their text prompts to speed up API calls and cut costs. “Early testers have reduced prompt size by up to 90% by fine-tuning instructions into the model.
“Since the release of GPT-3.5 Turbo, developers and businesses have asked for the ability to customize the model to create unique and differentiated experiences for their users,” the company wrote in a blog post published this afternoon. “This update gives developers the ability to customize models that perform better for their use cases and run these custom models at scale.”
With fine-tuning enabled, businesses using GPT-3.5 Turbo through OpenAI’s API can make the model better follow instructions, such as having it always respond in a given language. They can also improve the model’s ability to consistently format responses as well as hone the “feel” of the model’s output, like its tone, so that it better fits a brand or voice.
Fine-tuning currently requires prepping data, uploading the necessary files and creating a fine-tuning job through OpenAI’s API. All fine-tuning data must pass through a “moderation” API and a GPT-4-powered moderation system to see if it’s in conflict with OpenAI’s safety standards, says the company. But OpenAI plans to launch a fine-tuning UI in the future with a dashboard for checking the status of ongoing fine-tuning workloads.
Fine-tuning costs are as follows:
- Training: $0.008 / 1K tokens
- Usage input: $0.012 / 1K tokens
- Usage output: $0.016 / 1K tokens
“Tokens” represent raw text — e.g. “fan,” “tas” and “tic” for the word “fantastic.” A GPT-3.5-turbo fine-tuning job with a training file of 100,000 tokens, or about 75,000 words, would cost around $2.40, OpenAI says.
Additionally, OpenAI said that fine-tuning support for GPT-4 — which, unlike GPT-3.5, can understand images in addition to text — will arrive sometime later this fall, but didn’t provide specifics beyond that.