Logo
Back to all models

GPT-3.5 Turbo - In-Depth Overview

OpenAI · GPT-3.5

outdated

Get all the details on GPT-3.5 Turbo, an AI model from OpenAI. This page covers its token limits, pricing structure, key capabilities such as chat, Code Generation, fine_tuning, available API code samples, and performance strengths.

Key Metrics

Input Limit

16.4K tokens

Output Limit

4.1K tokens

Input Cost

$0.50/1M

Output Cost

$1.50/1M

Sample API Code

from openai import OpenAI

client = OpenAI()

completion = client.chat.completions.create(
  model='gpt-3.5-turbo',
  messages=[
    {'role': 'system', 'content': 'You are a helpful assistant.'},
    {'role': 'user', 'content': 'Hello!'}
  ]
)

print(completion.choices[0].message.content)

Required Libraries

openai
openai

Benchmarks

BenchmarkScoreSourceNotes
1106
OpenLLM LeaderboardScore for gpt-3.5-turbo-0125 snapshot, which the gpt-3.5-turbo alias currently points to.

Notes

GPT-3.5 Turbo models can understand and generate natural language or code and have been optimized for chat using the Chat Completions API but work well for non-chat tasks as well. As of July 2024, use gpt-4o-mini in place of GPT-3.5 Turbo, as it is cheaper, more capable, multimodal, and just as fast. GPT-3.5 Turbo is still available for use in the API. Note: Provided documentation snippet states Function Calling and Structured Outputs are not supported for this model.

Capabilities

chat
Code Generation
fine tuning
text generation

Supported Data Types

Input Types

text

Output Types

text

Strengths & Weaknesses

Good at

chat
text generation
code generation

Poor at

advanced reasoning
long context processing
multimodal understanding

Additional Information

Latest Update

Jan 25, 2025

Knowledge Cutoff

Sep 1, 2021