PyPI Stats
  • Insights
  • PyPI
  • GitHub
  • Search
  • Compare
  • Advisories
  • Ecosystem
  • About
Home

Search Packages

Find Python packages by name, description, GitHub topic, or filter by metrics
Trusted-AI
aif360

A comprehensive set of fairness metrics for datasets and machine learning models, explanations for these metrics, and algorithms to mitigate bias in datasets and models.

40K 3K 914
aria-ml
dataeval

Python library for analyzing data quality and its impact on model performance across classification and object-detection tasks.

4K 17 6
aria-ml
dataeval-plots

Python library for analyzing data quality and its impact on model performance across classification and object-detection tasks.

2K 17 6
cvs-health
langfair

LangFair is a Python library for conducting use-case level LLM bias and fairness assessments

2K 257 43
lorentzenchr
model-diagnostics

Tools for diagnostics and assessment of (machine learning) models

2K 45 5
peremartra
optipfair

Structured pruning and bias visualization for Large Language Models. Tools for LLM optimization and fairness analysis.

1K 38 9
Khanz9664
trustlens

Open-source Python library for evaluating ML model reliability beyond accuracy — with calibration, failure, and fairness diagnostics for informed deployment decisions.

995 10 12
ankurpand3y
judicator

Who evaluates the evaluator? Judicator audits LLM-as-a-Judge systems for 7 documented bias types. Zero config. Works with any LLM.

973 5 1
NahuelGiudizi
ai-safety-tester

LLM security testing framework with CVE-style severity scoring and multi-model benchmarking

847 0 0
aria-ml
daml

Python library for analyzing data quality and its impact on model performance across classification and object-detection tasks.

827 17 6
dccuchile
wefe

The Word Embedding Fairness Evaluation Framework

662 182 14
mishi93999
seatbelt

Responsible AI auditing for LLMs and SLMs — deception, fairness, sociotechnical risk, regulatory compliance

536 0 0
SolomonB14D3
rho-eval

Behavioral auditing toolkit for LLMs — audit any model across 8 dimensions (factual, toxicity, bias, sycophancy, reasoning, refusal, deception, over-refusal) using teacher-forced confidence probes.

498 4 0
SolomonB14D3
knowledge-fidelity

Compress LLMs while auditing whether they still know truth vs myths. SVD compression + false-belief detection in one toolkit.

313 4 0
antrixsh
trusteval-ai

Enterprise LLM Evaluation & Responsible AI Framework — Benchmark bias, hallucination, PII leakage, and toxicity across Healthcare, BFSI, Retail & Legal industries. Supports OpenAI, Anthropic, Gemini & HuggingFace. Python SDK + CLI + Web Dashboard. 191 tests. Compliance-ready reports.

252 7 5
VectorInstitute
unbias-plus

Python package that finds biased language in text, explains why, suggests neutral wording, and rewrites the whole text. Use from CLI, API, or Python.

243 2 1
ethical-spectacle
the-fairly-project

Bias detection Toolkit: Chrome Extension, Python Package, SOTA research paper docs.

207 4 0
TaimoorKhan10
ai-fairness-toolkit

AI Fairness and Explainability Toolkit (AFET) is an open-source project aimed at providing tools and frameworks to assess, visualize, and mitigate bias in machine learning models. It supports multiple ML frameworks and offers a comprehensive suite of metrics and visualization components to enhance model transparency and fairness.

206 0 2
IQTLabs
aiscan

Scan your AI/ML models for problems before you put them into production.

184 11 7
jorgeMFS
phenoqc

PhenoQC is a lightweight, efficient, and user-friendly toolkit designed to perform comprehensive quality control (QC) on phenotypic datasets. It ensures that data adheres to standardized formats, maintains consistency, and is harmonized with recognized ontologies.

174 0 0
whis-19
whis-ethical-ai

Ethical AI Validator detects bias and assesses fairness in AI models with statistical parity analysis, real-time monitoring, and automated GDPR/AI Act compliance reporting. Python 2.7+ compatible.

156 0 0
VectorInstitute
fairsense-agentix

An agentic fairness and AI-risk analysis platform for detecting bias in text and images

154 2 1
Trusted-AI
aif360-fork2

A comprehensive set of fairness metrics for datasets and machine learning models, explanations for these metrics, and algorithms to mitigate bias in datasets and models.

130 3K 914
bws82
biasclear

Structural bias detection engine built on Persistent Influence Theory (PIT). Detect framing, anchoring, false consensus, and 30+ rhetorical distortion patterns.

98 1 0
    • Data from PyPI, GitHub, ClickHouse, and BigQuery