53 dependents
Package Description Downloads/month
SageMaker Serve package for model serving and deployment 1.6M
MLServer 334K
Visualize Deep Learning 207K
NeMo Retriever Library is a scalable, performance-oriented document content and ... 76K
NeMo Retriever Library is a scalable, performance-oriented document content and ... 31K
PyTriton is a Flask/FastAPI-like interface that simplifies Triton's deployment i... 23K
15K
Triton Model Analyzer is a CLI tool to help with better understanding of the com... 5K
FEDML - The unified and scalable ML library for large-scale distributed training... 3K
Real-time and batch prediction service for Open Food Facts. 2K
Democratizing ML in proteomics 2K
Inference utilities for keynet - Triton Inference Server integration 1K
MONAI Deploy App SDK offers a framework and associated tools to design, develop ... 1K
Triton Model Navigator: An inference toolkit for optimizing and deploying machin... 1K
NVIDIA's package for core modules common across TAO Toolkit DNNs. 739
llama-index llms nvidia triton integration 714
713
End of utterance detection for LiveKit Agents 686
Inference-as-a-Service deployment made simple 605
easy to use bert with nvidia triton inference server 601
Tiny configuration for Triton Inference Server 564
vnc_tool 543
Provides a buildkit for constructing Triton Inference Server model repositories 534
MLflow plugin for Triton Inference Server with secure Python function execution 493
409
A framework for deploying serializable and optimizable neural net models at scal... 352
A powerful terminal user interface (TUI) for real-time GPU monitoring and proces... 330
Adding Grammarly (and other) open source ML models to LanguageTool 314
a python package for computer vision 305
290
A lightweight and modular Python package for handling computer vision inference ... 287
A utterly useless package that imports everything for you. Now with top 1000 PyP... 247
Unified SDK for Face detection, Faiss vector match, and Triton gRPC inference 244
A re-implementation of ViT containing utilities to convert to TensorRT engines a... 236
雪浪模型推理服务的客户端 193
A Python package with a built-in web application 177
171
A high level package for Nvidia Triton requests 168
Client utilities for the triton inference server 145
137
Triton Inference Server Client 116
Guided molecule generation via the BioNemo cloud service 84
Agent compiler with orchestration-level memoization for LLM inference 77
An integration package connecting TritonTensorRT and LangChain 69
Inference client for msir inference service 64
A unified Triton client for speech recognition and object detection. 55
Easier Shared Memory management for Triton Inference Client 49
project descriptions here 46
AIO Triton Utilities 39
基于python标准库concurrent & requests封装tritonclient 37