PyPI Stats
  • Insights
  • PyPI
  • GitHub
  • Search
  • Compare
  • Advisories
  • Ecosystem
  • About
Home

Search Packages

Find Python packages by name, description, GitHub topic, or filter by metrics
vllm-project
vllm

A high-throughput and memory-efficient inference and serving engine for LLMs

9.4M 79K 16K
vllm-project
vllm-tpu

A high-throughput and memory-efficient inference and serving engine for LLMs

143K 79K 16K
ThreeFish-AI
coding-proxy

A High-Availability, Transparent, and Smart Multi-Vendor Proxy for Claude Code. Support Claude Plans, GitHub Copilot, Google Antigravity, ZAI/GLM, MiniMax, Qwen, Xiaomi, Kimi, Doubao...

15K 15 1
Shelpuk-AI-Technology-Consulting
kitty-bridge

Universal LLM bridge for AI agents. Use Claude Code with MiniMax, Codex with GLM, or Gemini CLI with OpenRouter — one command, any provider. Works with coding agents, OpenClaw, Hermes, and others.

7K 8 2
Amanbig
devorch

A terminal-native, multi-provider intelligent assistant that plans, executes, and tracks developer tasks, not just answers prompts, similar to Claude Code and Gemini CLI.

559 4 0
vllm-project
vllm-hust

A high-throughput and memory-efficient inference and serving engine for LLMs

437 79K 16K
vllm-project
wxy-test

A high-throughput and memory-efficient inference and serving engine for LLMs

375 2K 1K
LLMPages
llm-onesdk

OneSDK is a Python library that provides a unified interface for interacting with various Large Language Model (LLM) providers.

363 2 0
vllm-project
vllm-xft

A high-throughput and memory-efficient inference and serving engine for LLMs

345 79K 16K
vllm-project
ai-dynamo-vllm

A high-throughput and memory-efficient inference and serving engine for LLMs

344 79K 16K
vllm-project
vllm-acc

A high-throughput and memory-efficient inference and serving engine for LLMs

342 79K 16K
vllm-project
nextai-vllm

A high-throughput and memory-efficient inference and serving engine for LLMs

273 79K 16K
vllm-project
vllm-consul

A high-throughput and memory-efficient inference and serving engine for LLMs

219 79K 16K
vllm-project
vllm-npu

A high-throughput and memory-efficient inference and serving engine for LLMs

209 79K 16K
SertraFurr
kimi4free

Simple API Wrapper for Kimi

207 4 1
vllm-project
vllm-musa

A high-throughput and memory-efficient inference and serving engine for LLMs

194 79K 16K
vllm-project
vllm-rocm

A high-throughput and memory-efficient inference and serving engine for LLMs

176 79K 16K
shibing624
chatpilot

ChatPilot: Chat Agent Web UI,实现Chat对话前端,支持Google搜索、文件网址对话(RAG)、代码解释器功能,复现了Kimi Chat(文件,拖进来;网址,发出来)。

134 599 59
vllm-project
vllm-emissary

A high-throughput and memory-efficient inference and serving engine for LLMs

132 79K 16K
vllm-project
vllm-usf

A high-throughput and memory-efficient inference and serving engine for LLMs

115 79K 16K
AirTouch666
aether-cli

A command-line interface for interacting with various AI models.

113 0 0
vllm-project
tilearn-infer

A high-throughput and memory-efficient inference and serving engine for LLMs

107 79K 16K
vllm-project
vllm-online

A high-throughput and memory-efficient inference and serving engine for LLMs

82 79K 16K
vllm-project
vllm-test-tpu

A high-throughput and memory-efficient inference and serving engine for LLMs

80 79K 16K
    • Data from PyPI, GitHub, ClickHouse, and BigQuery