PyPI Stats
  • Insights
  • PyPI
  • GitHub
  • Search
  • Compare
  • Advisories
  • Ecosystem
  • About
Home

Search Packages

Find Python packages by name, description, GitHub topic, or filter by metrics
izikeros
count-tokens

Count tokens in a text file.

46K 13 0
gweidart
rs-bpe

A ridiculously fast Python BPE (Byte Pair Encoder) implementation written in Rust

24K 38 5
cahya-wirawan
pyrwkv-tokenizer

A fast RWKV Tokenizer written in Rust

13K 54 5
cahya-wirawan
rwkv-tokenizer

A fast RWKV Tokenizer written in Rust

3K 54 5
kgruiz
pytokencounter

A simple Python library for tokenizing text and counting tokens. While currently only supporting OpenAI LLMs, it helps with text processing and managing token limits in AI applications.

2K 2 0
stef41
toksight

Tokenizer analysis toolkit. Compare vocabulary coverage, compression ratios, and token boundaries across GPT-4o, Llama 3, Mistral, and any HuggingFace tokenizer.

695 1 0
DelvyG
promptminify

Minify LLM prompts to save tokens — domain-aware, tiktoken-validated, zero regressions

689 1 0
Thibault00
runtoken

A blazing-fast BPE tokenizer for LLMs. Drop-in tiktoken replacement, 20-80x faster.

502 1 0
Reversehobo
openai-function-tokens

Predict the exact openai token usage of functions

414 19 1
Wolfe-Jam
slash-tokens

Token Optimization for Context Engineers. 4.8 KB WASM. Sub-millisecond. Zero dependencies.

209 4 0
unitythemaker
tokdu

tokdu (Token Disk Usage) is a terminal-based utility that helps you analyze and visualize token usage in your codebase. Similar to the classic du (disk usage) command, tokdu shows you how many tokens your files and directories consume, which is essential when working with Large Language Models (LLMs) that have token limits.

175 5 0
ElmiraGhorbani
chatgpt-long-term-memory

The ChatGPT Long Term Memory package is a powerful tool designed to empower your projects with the ability to handle a large number of simultaneous users and external sources.

114 62 3
DevelopedBy-Siva
llm-tokenscope

Profile your LLM payloads. Find the waste. Cut the cost. Field-level token attribution, cost leak detection, and payload optimization for any LLM API.

109 0 0
    • Data from PyPI, GitHub, ClickHouse, and BigQuery