91 dependents
| Package | Description | Downloads/month |
|---|---|---|
| A high-throughput and memory-efficient inference and serving engine for LLMs | 9.4M | |
| Turn any computer or edge device into a command center for your computer vision ... | 1.1M | |
| Lepton AI Platform | 447K | |
| A high-throughput and memory-efficient inference and serving engine for LLMs | 143K | |
| Turn any computer or edge device into a command center for your computer vision ... | 119K | |
| Wheels & Docker images for running vLLM on CPU-only systems, optimized for diffe... | 31K | |
| Open source alternative to LangGraph Platform (now LangSmith Deployments) - Self... | 29K | |
| TensorRT LLM provides users with an easy-to-use Python API to define Large Langu... | 16K | |
| Agentic Repo Maintenance | 13K | |
| A package that provides a set of tools to build a FastAPI application with a Cla... | 13K | |
| Rammearkitektur integrations framework | 12K | |
| OpenRAG is a comprehensive Retrieval-Augmented Generation platform that enables ... | 11K | |
| Application server components for LlamaDeploy | 11K | |
| A lightweight plugin framework for building extensible AI systems | 9K | |
| Turn any computer or edge device into a command center for your computer vision ... | 9K | |
| Large-scale LLM inference engine | 7K | |
| Turn any computer or edge device into a command center for your computer vision ... | 7K | |
| Package for fastAPI models | 6K | |
| An AI Gateway, registry, and proxy that sits in front of any MCP, A2A, or REST/g... | 5K | |
| Python SDK for RockAI.online | 5K | |
| Công cụ tải truyện CỰC ĐỘC ĐÁO từ Valvrareteam.net | 4K | |
| Wheels & Docker images for running vLLM on CPU-only systems, optimized for diffe... | 3K | |
| Application server components for LlamaDeploy | 3K | |
| Adaptive Agentic AI Reasoning using Microsoft Agent Framework -- Join the Discor... | 3K | |
| vLLM CPU inference engine (AVX512 + VNNI optimized) | 3K | |
| Wheels & Docker images for running vLLM on CPU-only systems, optimized for diffe... | 3K | |
| Matter's Observability Library - Includes all observability functions, including... | 3K | |
| vLLM CPU inference engine (AVX512 optimized) | 2K | |
| General Package for Microservices based on FastAPI like Profiler, Scheduler, Sys... | 2K | |
| A shared tools for other services | 2K | |
| Common utilities for Camptocamp ASGI applications | 2K | |
| gULP - (generic) Unified Log Processor. | 2K | |
| Estimate Energy Consumption | 1K | |
| Feature flags server | 931 | |
| API看板服务 | 873 | |
| Citation-verified RAG service with deterministic + semantic claim verification. | 769 | |
| Open-source High-performance authorization engine for RBAC, ReBAC, and ACL. Mult... | 761 | |
| Generative AI components | 722 | |
| General Information, model certifications, and benchmarks for nm-vllm enterprise... | 666 | |
| Expose Great Expectations data-quality checks via MCP | 646 | |
| Package for IoC knowledge management | 645 | |
| GRID - Geometric Resonance Intelligence Driver: A comprehensive framework for ex... | 635 | |
| FastAPI server for Z3rno: REST API, authentication, rate limiting, and Celery wo... | 625 | |
| Defense-in-depth input safety for LLMs — perplexity gate + FAISS + ModernBERT +... | 515 | |
| vLLM Kunlun3 backend plugin | 464 | |
| TextEmbed is a REST API crafted for high-throughput and low-latency embedding in... | 440 | |
| A high-throughput and memory-efficient inference and serving engine for LLMs | 437 | |
| A high-throughput and memory-efficient inference and serving engine for LLMs | 375 | |
| Python package with core python to use in microservices | 357 | |
| Add your description here | 355 |