37 dependents
| Package | Description | Downloads/month |
|---|---|---|
| MLX-Embeddings is the best package for running Vision and Language Embedding mod... | 51K | |
| vMLX - Home of JANG_Q - Cont Batch, Prefix, Paged, KV Cache Quant, VL - Powers M... | 37K | |
| A high-performance API server that provides OpenAI-compatible endpoints for MLX ... | 22K | |
| The fastest local AI engine for Apple Silicon. 4.2x faster than Ollama, 0.08s ca... | 21K | |
| Fine-tune LLMs on your Mac with Apple Silicon. SFT, DPO, GRPO, Vision, TTS, STT,... | 14K | |
| vLLM-like inference for Apple Silicon - GPU-accelerated Text, Image, Video & Aud... | 11K | |
| port of mlx-video but attempting to get audio ported as well finally | 6K | |
| This is LLM interface library. | 2K | |
| Unified MLX server & CLI (language and vision) with OpenAI-compatible endpoints | 2K | |
| HuggingFace model management for MLX on Apple Silicon | 2K | |
| RETIX - The Optic Nerve for Autonomous Agents | 2K | |
| Standalone MLX-based LLM inference service with OpenAI compatible API | 2K | |
| Photonamer: Autonomous photo file renaming tool using local Visual-Language Mode... | 1K | |
| An optimized MLX (Apple Silicon Metal) Server for running local LLMs with higher... | 997 | |
| Offline meeting recorder & summarizer for macOS | 977 | |
| Set-of-Mark detection pipeline for macOS — Apple Vision, YOLO11, and VLM on MLX.... | 773 | |
| Automatically caption images using various LLaVA multimodal models. This tool pr... | 749 | |
| vLLM hardware plugin for Apple Silicon - unifies MLX and PyTorch under a single ... | 669 | |
| ('A library for augmenting large language models using MLX',) | 552 | |
| Ollama-style daemon and CLI over vllm-mlx on Apple Silicon | 446 | |
| Intelligent mobile simulator control — AI-driven device testing for vibe coding | 443 | |
| PDF processing pipeline: remove headers/footers, convert to markdown, and genera... | 414 | |
| Anthropic-compatible local LLM server for Claude Code, backed by mlx-lm / mlx-vl... | 386 | |
| A python package for serving LLM on OpenAI-compatible API endpoints with prompt ... | 386 | |
| Experimental: MLX model provider for Strands Agents - Build, train, and deploy A... | 334 | |
| A local/offline-capable voice assistant with speech recognition, LLM processing,... | 329 | |
| FastMLX is a high performance production ready API to host MLX models. | 290 | |
| Medical AI on Apple Silicon – MedGemma 1.5 4B via MLX | 277 | |
| A Python package with a built-in web application | 276 | |
| Travel Vlog Automation System - Automate vlog ingestion, junk detection, and DaV... | 254 | |
| CLI for running LLMs on Apple Silicon via MLX | 221 | |
| Local OpenAI API runtime and tool-search foundation for multimodal Cappuccino-cl... | 203 | |
| MLX-GUI MLX Inference Server for Apple Silicone | 190 | |
| An OpenAI-compatible inference server optimized for Apple Silicon. Hot-swap betw... | 174 | |
| Local-first alt-text generator built on top of MLX-VLM | 122 | |
| FastMLX is a high performance production ready API to host MLX models. | 85 | |
| A Python package with a built-in web application | 84 |