A curated list of awesome open source libraries to deploy, monitor, version and scale your machine learning
-
Updated
Jul 1, 2024
A curated list of awesome open source libraries to deploy, monitor, version and scale your machine learning
🐢 Open-Source Evaluation & Testing for LLMs and ML models
A Python package to assess and improve fairness of machine learning models.
The Python Risk Identification Tool for generative AI (PyRIT) is an open access automation framework to empower security professionals and machine learning engineers to proactively find risks in their generative AI systems.
moDel Agnostic Language for Exploration and eXplanation
Responsible AI Toolbox is a suite of tools providing model and data exploration and assessment user interfaces and libraries that enable a better understanding of AI systems. These interfaces and libraries empower developers and stakeholders of AI systems to develop and monitor AI more responsibly, and take better data-driven actions.
Deliver safe & effective language models
A toolkit that streamlines and automates the generation of model cards
💡 Adversarial attacks on explanations and how to defend them
Official code repo for the O'Reilly Book - Machine Learning for High-Risk Applications
[NeurIPS 2023] Sentry-Image: Detect Any AI-generated Images
Reading list for adversarial perspective and robustness in deep reinforcement learning.
Carefully curated list of awesome data science resources.
Référentiel d'évaluation data science responsable et de confiance
A detailed summary of "Designing Machine Learning Systems" by Chip Huyen. This book gives you and end-to-end view of all the steps required to build AND OPERATE ML products in production. It is a must-read for ML practitioners and Software Engineers Transitioning into ML.
PyTorch package to train and audit ML models for Individual Fairness
A collection of news articles, books, and papers on Responsible AI cases. The purpose is to study these cases and learn from them to avoid repeating the failures of the past.
[ICCV 2023 Oral, Best Paper Finalist] ITI-GEN: Inclusive Text-to-Image Generation
Python library for implementing Responsible AI mitigations.
Credo AI Lens is a comprehensive assessment framework for AI systems. Lens standardizes model and data assessment, and acts as a central gateway to assessments created in the open source community.
Add a description, image, and links to the responsible-ai topic page so that developers can more easily learn about it.
To associate your repository with the responsible-ai topic, visit your repo's landing page and select "manage topics."