PyPI Stats
  • Insights
  • PyPI
  • GitHub
  • Search
  • Compare
  • Advisories
  • Ecosystem
  • About
Home

Search Packages

Find Python packages by name, description, GitHub topic, or filter by metrics
pytorch
torchao

PyTorch native quantization and sparsity for training and inference

3.4M 3K 502
openvinotoolkit
nncf

Neural Network Compression Framework for enhanced OpenVINO™ inference

456K 1K 293
tensorflow
tensorflow-model-optimization

A toolkit to optimize ML models for deployment for Keras and TensorFlow, including quantization and pruning.

105K 2K 347
intel
neural-compressor

SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, and ONNX Runtime

22K 3K 304
tensorflow
tf-model-optimization-nightly

A suite of tools that users, both novice and advanced can use to optimize machine learning models for deployment and execution.

10K 2K 347
neuralmagic
sparseml

Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models

3K 2K 156
satabios
sconce

E2E AutoML Model Compression Package

2K 45 4
bwohlberg
sporco

Sparse Optimisation Research Code

2K 273 40
intel
neural-compressor-pt

SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, and ONNX Runtime

1K 3K 304
FasterAI-Labs
fasterai

FasterAI: Prune and Distill your models with FastAI and PyTorch

1K 261 19
adrhill
asdex

Automatic Sparse Differentiation in JAX.

1K 25 2
intel
neural-speed

An innovative library for efficient LLM inference via low-bit quantization

1K 352 38
intel
neural-compressor-tf

Repository of Intel® Neural Compressor

739 3K 304
intel
neural-compressor-full

SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, and ONNX Runtime

476 3K 304
intel
neural-solution

Repository of Intel® Neural Compressor

384 3K 304
intel
lpot

Repository of Intel® Low Precision Optimization Tool

384 3K 302
intel
neural-insights

SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, and ONNX Runtime

340 3K 304
SIP-AAU
magni

A package for AFM image reconstruction and compressed sensing in general

191 43 10
RabadanLab
randomly

A Library for Denoising Single-Cell Data with Random Matrix Theory

142 38 10
intel
ilit

SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, and ONNX Runtime

94 3K 304
ryantd
veloce

Veloce: An instant distributed computing library based on Ray stack

88 17 0
danielathome19
tensorflow-engram

A Python package for Engram Neural Networks, adding biologically-inspired Hebbian memory and engram layers to TensorFlow/Keras models, supporting memory traces, plasticity, attention, and sparsity for neural sequence learning.

76 4 0
fraenkel-lab
gslr

An algorithm for classification from a graph-sparse support

73 15 4
intel
neural-compressor-3x-tf

SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, and ONNX Runtime

14 3K 304
    • Data from PyPI, GitHub, ClickHouse, and BigQuery