SkillAgentSearch skills...

Tokencost

Easy token price estimates for 400+ LLMs. TokenOps.

Install / Use

/learn @AgentOps-AI/Tokencost
About this skill

Quality Score

0/100

Supported Platforms

Claude Code
Claude Desktop

README

<p align="center"> <img src="https://raw.githubusercontent.com/AgentOps-AI/tokencost/main/tokencost.png" height="300" alt="Tokencost" /> </p> <p align="center"> <em>Clientside token counting + price estimation for LLM apps and AI agents.</em> </p> <p align="center"> <a href="https://pypi.org/project/tokencost/" target="_blank"> <img alt="Python" src="https://img.shields.io/badge/python-3670A0?style=for-the-badge&logo=python&logoColor=ffdd54" /> <img alt="Version" src="https://img.shields.io/pypi/v/tokencost?style=for-the-badge&color=3670A0"> </a> </p> <p align="center"> <a href="https://twitter.com/agentopsai/">🐦 Twitter</a> <span>&nbsp;&nbsp;•&nbsp;&nbsp;</span> <a href="https://discord.com/invite/FagdcwwXRR">📢 Discord</a> <span>&nbsp;&nbsp;•&nbsp;&nbsp;</span> <a href="https://agentops.ai/?tokencost">🖇️ AgentOps</a> </p>

TokenCost

License: MIT PyPI - Version X (formerly Twitter) Follow

Tokencost helps calculate the USD cost of using major Large Language Model (LLMs) APIs by calculating the estimated cost of prompts and completions.

Building AI agents? Check out AgentOps

Features

  • LLM Price Tracking Major LLM providers frequently add new models and update pricing. This repo helps track the latest price changes
  • Token counting Accurately count prompt tokens before sending OpenAI requests
  • Easy integration Get the cost of a prompt or completion with a single function

Example usage:

from tokencost import calculate_prompt_cost, calculate_completion_cost

model = "gpt-3.5-turbo"
prompt = [{ "role": "user", "content": "Hello world"}]
completion = "How may I assist you today?"

prompt_cost = calculate_prompt_cost(prompt, model)
completion_cost = calculate_completion_cost(completion, model)

print(f"{prompt_cost} + {completion_cost} = {prompt_cost + completion_cost}")
# 0.0000135 + 0.000014 = 0.0000275

Installation

Recommended: PyPI:

pip install tokencost

Usage

Cost estimates

Calculating the cost of prompts and completions from OpenAI requests

from openai import OpenAI

client = OpenAI()
model = "gpt-3.5-turbo"
prompt = [{ "role": "user", "content": "Say this is a test"}]

chat_completion = client.chat.completions.create(
    messages=prompt, model=model
)

completion = chat_completion.choices[0].message.content
# "This is a test."

prompt_cost = calculate_prompt_cost(prompt, model)
completion_cost = calculate_completion_cost(completion, model)
print(f"{prompt_cost} + {completion_cost} = {prompt_cost + completion_cost}")
# 0.0000180 + 0.000010 = 0.0000280

Calculating cost using string prompts instead of messages:

from tokencost import calculate_prompt_cost

prompt_string = "Hello world" 
response = "How may I assist you today?"
model= "gpt-3.5-turbo"

prompt_cost = calculate_prompt_cost(prompt_string, model)
print(f"Cost: ${prompt_cost}")
# Cost: $3e-06

Counting tokens

from tokencost import count_message_tokens, count_string_tokens

message_prompt = [{ "role": "user", "content": "Hello world"}]
# Counting tokens in prompts formatted as message lists
print(count_message_tokens(message_prompt, model="gpt-3.5-turbo"))
# 9

# Alternatively, counting tokens in string prompts
print(count_string_tokens(prompt="Hello world", model="gpt-3.5-turbo"))
# 2

How tokens are counted

Under the hood, strings and ChatML messages are tokenized using Tiktoken, OpenAI's official tokenizer. Tiktoken splits text into tokens (which can be parts of words or individual characters) and handles both raw strings and message formats with additional tokens for message formatting and roles.

For Anthropic models above version 3 (i.e. Sonnet 3.5, Haiku 3.5, and Opus 3), we use the Anthropic beta token counting API to ensure accurate token counts. For older Claude models, we approximate using Tiktoken with the cl100k_base encoding.

Cost table

Units denominated in USD. All prices can be located here.

<!-- PRICING_TABLE_START -->

| Model Name | Prompt Cost (USD) per 1M tokens | Completion Cost (USD) per 1M tokens | Max Prompt Tokens | Max Output Tokens | |:----------------------------------------------------------------------|:----------------------------------|:--------------------------------------|:--------------------|--------------------:| | gpt-4 | $30 | $60 | 8192 | 4096 | | gpt-4o | $2.5 | $10 | 128,000 | 16384 | | gpt-4o-audio-preview | $2.5 | $10 | 128,000 | 16384 | | gpt-4o-audio-preview-2024-10-01 | $2.5 | $10 | 128,000 | 16384 | | gpt-4o-mini | $0.15 | $0.6 | 128,000 | 16384 | | gpt-4o-mini-2024-07-18 | $0.15 | $0.6 | 128,000 | 16384 | | o1-mini | $1.1 | $4.4 | 128,000 | 65536 | | o1-mini-2024-09-12 | $3 | $12 | 128,000 | 65536 | | o1-preview | $15 | $60 | 128,000 | 32768 | | o1-preview-2024-09-12 | $15 | $60 | 128,000 | 32768 | | chatgpt-4o-latest | $5 | $15 | 128,000 | 4096 | | gpt-4o-2024-05-13 | $5 | $15 | 128,000 | 4096 | | gpt-4o-2024-08-06 | $2.5 | $10 | 128,000 | 16384 | | gpt-4-turbo-preview | $10 | $30 | 128,000 | 4096 | | gpt-4-0314 | $30 | $60 | 8,192 | 4096 | | gpt-4-0613 | $30 | $60 | 8,192 | 4096 | | gpt-4-32k | $60 | $120 | 32,768 | 4096 | | gpt-4-32k-0314 | $60 | $120 | 32,768 | 4096 | | gpt-4-32k-0613 | $60 | $120 | 32,768 | 4096 | | gpt-4-turbo | $10 | $30 | 128,000 | 4096 | | gpt-4-turbo-2024-04-09 | $10 | $30 | 128,000 | 4096 | | gpt-4-1106-preview | $10 | $30 | 128,000 | 4096 | | gpt-4-0125-preview | $10 | $30 | 128,000 | 4096 | | gpt-4-vision-preview | $10 | $30 | 128,000 | 4096 | | gpt-4-1106-vision-preview | $10 | $30 | 128,000 | 4096 | | gpt-3.5-turbo | $1.5 | $2 | 16,385 | 4096 | | gpt-3.5-turbo-0301 | $1.5 | $

View on GitHub
GitHub Stars2.0k
CategoryData
Updated7h ago
Forks103

Languages

Python

Security Score

100/100

Audited on Mar 30, 2026

No findings