PyPI Stats
  • Insights
  • PyPI
  • GitHub
  • Search
  • Compare
  • Advisories
  • Ecosystem
  • About
Home

Search Packages

Find Python packages by name, description, GitHub topic, or filter by metrics
lucidrains
ring-attention-pytorch

Implementation of 💍 Ring Attention, from Liu et al. at Berkeley AI, in Pytorch

5K 548 35
lucidrains
megabyte-pytorch

Implementation of MEGABYTE, Predicting Million-byte Sequences with Multiscale Transformers, in Pytorch

4K 655 55
lucidrains
recurrent-memory-transformer-pytorch

Implementation of Recurrent Memory Transformer, Neurips 2022 paper, in Pytorch

2K 423 19
lucidrains
infini-transformer-pytorch

Implementation of Infini-Transformer in Pytorch

2K 112 4
Neuranox
titans-memory

Titans in PyTorch: Infinite Context models that learn to remember. Faster than Transformers, smarter than RNNs, and fully HuggingFace compatible.

1K 1 2
compactbench
compactbench

Open benchmark for LLM context compaction methods — measures what survives when you replace conversation history with a compacted artifact. Multi-cycle drift, hidden ranked set.

997 1 0
lucidrains
perceiver-ar-pytorch

Implementation of Perceiver AR, Deepmind's new long-context attention network based on Perceiver architecture, in Pytorch

585 95 4
forhaoliu
ringattention

RingAttention for Transformers with Arbitrarily Large Context.

491 770 52
jagmarques
nexusquant-kv

Training-free KV cache compression for LLMs. 10-33x compression via E8 lattice quantization + attention-aware token eviction. One line of code.

476 13 0
denial-web
hard-needle

Semantically hard multi-needle long-context data generator. Stop testing LLMs with random-password needles.

398 0 0
FluffyAIcode
kakeyalattice

Discrete Kakeya cover for LLM KV cache: D4/E8 nested-lattice quantisation realising a Kakeya-style tube-cover over the direction sphere. 2.4x-2.8x compression at <1% perplexity loss on Qwen3, Llama-3, DeepSeek, GLM-4, Gemma. Drop-in transformers.DynamicCache. pip install kakeyalattice.

365 7 2
melvinebenezer
liah

needle in a haystack for LLMs

79 2 0
yuplin2333
mcp-long-context-reader

A tool to help agents read and query long documents.

75 5 1
dschulmeist
replm

Recursive Language Models — process arbitrarily long prompts by offloading context into a REPL with symbolic recursion via sub-LLM calls. Based on Zhang, Kraska & Khattab (2025).

73 1 0
bytedance
shadowkv

shadow kv cache

67 294 21
    • Data from PyPI, GitHub, ClickHouse, and BigQuery