Search
⌃K

Getting Started

Welcome to the official documentation of AI Vulnerability Database (AVID)!
As the first open-source, extensible knowledge base of failures across the AI Ecosystem (e.g. data sets, models, systems), AVID aims to
  • encompass coordinates of responsible ML such as security, ethics, and performance
  • build out a taxonomy of potential harms across these coordinates
  • house full-fidelity information (e.g. metadata, measurements, benchmarks) on evaluation use cases of a harm (sub)category
  • evaluate models and datasets that are either open-source or accessible through APIs
This site contains information to get you started with different components of AVID.
  • ​Taxonomy: a landing place of instances of AI system/model/dataset failures.
  • ​Database: stores information on such instances in a structured manner.
  • ​Developer SDK: the official Python toolkit for working with AVID.
​
​
Last modified 7mo ago