53 dependents
| Package | Description | Downloads/month |
|---|---|---|
| SageMaker Serve package for model serving and deployment | 1.6M | |
| MLServer | 334K | |
| Visualize Deep Learning | 207K | |
| NeMo Retriever Library is a scalable, performance-oriented document content and ... | 76K | |
| NeMo Retriever Library is a scalable, performance-oriented document content and ... | 31K | |
| PyTriton is a Flask/FastAPI-like interface that simplifies Triton's deployment i... | 23K | |
| 15K | ||
| Triton Model Analyzer is a CLI tool to help with better understanding of the com... | 5K | |
| FEDML - The unified and scalable ML library for large-scale distributed training... | 3K | |
| Real-time and batch prediction service for Open Food Facts. | 2K | |
| Democratizing ML in proteomics | 2K | |
| Inference utilities for keynet - Triton Inference Server integration | 1K | |
| MONAI Deploy App SDK offers a framework and associated tools to design, develop ... | 1K | |
| Triton Model Navigator: An inference toolkit for optimizing and deploying machin... | 1K | |
| NVIDIA's package for core modules common across TAO Toolkit DNNs. | 739 | |
| llama-index llms nvidia triton integration | 714 | |
| 713 | ||
| End of utterance detection for LiveKit Agents | 686 | |
| Inference-as-a-Service deployment made simple | 605 | |
| easy to use bert with nvidia triton inference server | 601 | |
| Tiny configuration for Triton Inference Server | 564 | |
| vnc_tool | 543 | |
| Provides a buildkit for constructing Triton Inference Server model repositories | 534 | |
| MLflow plugin for Triton Inference Server with secure Python function execution | 493 | |
| 409 | ||
| A framework for deploying serializable and optimizable neural net models at scal... | 352 | |
| A powerful terminal user interface (TUI) for real-time GPU monitoring and proces... | 330 | |
| Adding Grammarly (and other) open source ML models to LanguageTool | 314 | |
| a python package for computer vision | 305 | |
| 290 | ||
| A lightweight and modular Python package for handling computer vision inference ... | 287 | |
| A utterly useless package that imports everything for you. Now with top 1000 PyP... | 247 | |
| Unified SDK for Face detection, Faiss vector match, and Triton gRPC inference | 244 | |
| A re-implementation of ViT containing utilities to convert to TensorRT engines a... | 236 | |
| 雪浪模型推理服务的客户端 | 193 | |
| A Python package with a built-in web application | 177 | |
| 171 | ||
| A high level package for Nvidia Triton requests | 168 | |
| Client utilities for the triton inference server | 145 | |
| 137 | ||
| Triton Inference Server Client | 116 | |
| Guided molecule generation via the BioNemo cloud service | 84 | |
| Agent compiler with orchestration-level memoization for LLM inference | 77 | |
| An integration package connecting TritonTensorRT and LangChain | 69 | |
| Inference client for msir inference service | 64 | |
| A unified Triton client for speech recognition and object detection. | 55 | |
| Easier Shared Memory management for Triton Inference Client | 49 | |
| project descriptions here | 46 | |
| AIO Triton Utilities | 39 | |
| 基于python标准库concurrent & requests封装tritonclient | 37 |