Get all the details on GPT-3.5 Turbo, an AI model from OpenAI. This page covers its token limits, pricing structure, key capabilities such as chat, Code Generation, fine_tuning, available API code samples, and performance strengths.
Key Metrics
Input Limit
16.4K tokens
Output Limit
4.1K tokens
Input Cost
$0.50/1M
Output Cost
$1.50/1M
Sample API Code
from openai import OpenAI
client = OpenAI()
completion = client.chat.completions.create(
model='gpt-3.5-turbo',
messages=[
{'role': 'system', 'content': 'You are a helpful assistant.'},
{'role': 'user', 'content': 'Hello!'}
]
)
print(completion.choices[0].message.content)
Required Libraries
Benchmarks
Benchmark | Score | Source | Notes |
---|---|---|---|
1106 | OpenLLM Leaderboard | Score for gpt-3.5-turbo-0125 snapshot, which the gpt-3.5-turbo alias currently points to. |
Notes
GPT-3.5 Turbo models can understand and generate natural language or code and have been optimized for chat using the Chat Completions API but work well for non-chat tasks as well. As of July 2024, use gpt-4o-mini in place of GPT-3.5 Turbo, as it is cheaper, more capable, multimodal, and just as fast. GPT-3.5 Turbo is still available for use in the API. Note: Provided documentation snippet states Function Calling and Structured Outputs are not supported for this model.
Capabilities
Supported Data Types
Input Types
Output Types
Strengths & Weaknesses
Good at
Poor at
Additional Information
Latest Update
Jan 25, 2025
Knowledge Cutoff
Sep 1, 2021