Analysis
Card

cover a broad picture of imbalances
Generates a fairness and bias report using the FairBench library. This explores many kinds of bias to paint a broad picture and help you decide on what is problematic and what is acceptable behavior. Imbalanced distributions of benefits are uncovered to serve as points of discussion of real-world impact.
Details for experts.
The generated report can be viewed in three different formats, where the model card contains a subset of results but attaches to these socio-technical concerns to be taken into account:
- A summary table of results.
- A simplified model card that includes concerns.
- The full report, including details.
The module's report summarizes how a model behaves on a provided dataset across different population groups. These groups are based on sensitive attributes like gender, age, and race. Each attribute can have multiple values, such as several genders or races. Numeric attributes, like age, are normalized to the range [0,1] and treated as fuzzy values, where 0 indicates membership to a fuzzy group of "small" values, and 1 indicates membership to a fuzzy group of "large" values. A separate set of fairness metrics is calculated for each prediction label.
If intersectional subgroup analysis is enabled, separate subgroups are created for each combination of sensitive attribute values. However, if there are too many attributes, some groups will be small or empty. Empty groups are ignored in the analysis.
ParametersSpecific concerns

focus on a specific definition of fairness
Computes a fairness or bias measure that matches a specific type of numerical evaluation using the FairBench library. The measure is built by combining simpler options to form more than 300 valid alternatives.
This computes a specific fairness concerns and does not paint a broad enough picture. Make sure that you explore prospective biases with other modules first, like model card.
Technical details.
The assessment is conducted over sensitive attributes like gender, age, and race. Each attribute can have multiple values, such as several genders or races. Numeric attributes, like age, are normalized to the range [0,1] and treated as fuzzy values, where 0 indicates membership to a fuzzy group of "small" values, and 1 indicates membership to a fuzzy group of "large" values. A separate set of fairness metrics is calculated for each prediction label.
If intersectional subgroup analysis is enabled, separate subgroups are created for each combination of sensitive attribute values. However, if there are too many attributes, some groups will be small or empty. Empty groups are ignored in the analysis.
ParametersInteractive report

for data scientists: explore several biases and their intermediate quantities
Creates an interactive report using the FairBench library. The report creates traceable evaluations that you can shift through to find actual sources of unfairness.
ParametersSklearn audit

report on the biases of a simple predictor
One way to evaluate the fairness of a dataset is by testing for biases using simple models with limited degrees of freedom. This module audits datasets by training such models provided by the scikit-learn library on half of the dataset. The second half is then used as test data to assess predictive performance and detect classification or scoring biases. Test data are used to generate a fairness and bias report with the FairBench library.
Details for experts.
If strong biases appear in the simple models that are explored, they may also persist in more complex models trained on the same data. To focus on the most significant biases, adjust the minimum shown deviation parameter. The report provides multiple types of fairness and bias assessments and can be viewed in three different formats, where the model card contains a subset of results but attaches to these socio-technical concerns to be taken into account:
- A summary table of results.
- A simplified model card with key fairness concerns.
- A full detailed report.
The module's report summarizes how a model behaves on a provided dataset across different population groups. These groups are based on sensitive attributes like gender, age, and race. Each attribute can have multiple values, such as several genders or races. Numeric attributes, like age, are normalized to the range [0,1] and treated as fuzzy values, where 0 indicates membership to a fuzzy group of "small" values, and 1 indicates membership to a fuzzy group of "large" values. A separate set of fairness metrics is calculated for each prediction label.
If intersectional subgroup analysis is enabled, separate subgroups are created for each combination of sensitive attribute values. However, if there are too many attributes, some groups will be small or empty. Empty groups are ignored in the analysis. The report may also include information about built-in datasets.
ParametersSklearn visual analysis

for data scientists: barplots with the biases of a simple predictor
One way to evaluate the fairness of a dataset is by testing for biases using simple models with limited degrees of freedom. This module audits datasets by training such models provided by the scikit-learn library on half of the dataset. The second half is then used as test data to assess predictive performance and detect classification or scoring biases. Test data are used to generate a fairness and bias report with the FairBench library.
Details for experts.
If strong biases appear in the simple models that are explored, they may also persist in more complex models trained on the same data. To focus on the most significant biases, adjust the minimum shown deviation parameter. The report provides multiple types of fairness and bias assessments and can be viewed in three different formats, where the model card contains a subset of results but attaches to these socio-technical concerns to be taken into account:
- A summary table of results.
- A simplified model card with key fairness concerns.
- A full detailed report.
The module's report summarizes how a model behaves on a provided dataset across different population groups. These groups are based on sensitive attributes like gender, age, and race. Each attribute can have multiple values, such as several genders or races. Numeric attributes, like age, are normalized to the range [0,1] and treated as fuzzy values, where 0 indicates membership to a fuzzy group of "small" values, and 1 indicates membership to a fuzzy group of "large" values. A separate set of fairness metrics is calculated for each prediction label.
If intersectional subgroup analysis is enabled, separate subgroups are created for each combination of sensitive attribute values. However, if there are too many attributes, some groups will be small or empty. Empty groups are ignored in the analysis. The report may also include information about built-in datasets.
ParametersAif360 metrics
popular types of bias
Use IBM's AIF360 to compute common group fairness metrics for each sensitive attribute provided. If attributes are non-binary, they are binarized into one-hot encoded columns. Only categorical attributes are allowed.
This module is based on AIF360, which does not support generalized intersectional analysis. When needed, generate sensitive intersectional group labels in your dataset. However, these will always be computed against the rest of the population.
ParametersOptimal transport
representational disparities in predictions
Expert details.
Creates an optimal transport evaluation based on the implementation provided by the AIF360 library. The evaluation computes the Wasserstein distance that reflects the cost of transforming the predictive distributions between sensitive attribute groups.
Optimal Transport (OT) is a field of mathematics which studies the geometry of probability spaces. Among its many contributions, OT provides a principled way to compare and align probability distributions by taking into account the underlying geometry of the considered metric space. As a mathematical problem, it was first introduced by Gaspard Monge in 1781. It addresses the task of determining the most efficient method for transporting mass from one distribution to another. In this problem, the cost associated with moving a unit of mass from one position to another is referred to as the ground cost. The primary objective of OT is to minimize the total cost incurred when moving one mass distribution onto another.
OT can be used to detect model-induced bias by calculating the a cost known as Earth Mover's distance or Wasserstein distance between the distribution of ground truth labels and model predictions for each of the protected groups. If its value is close to 1, the model is biased towards this group.
License
Parts of the above description are adapted from AIF360 (https://github.com/Trusted-AI/AIF360), which is licensed under Apache License 2.0.
ParametersBias scan
scan for biased attributes or attribute intersections
Use AIF360 to scans your dataset to estimate the most biased attributes or combinations of attributes. For example, gender may only show bias when combined with socioeconomic status, despite the latter not bein inherently sensitive. If you have already marked some attributes as sensitive (such as race or gender), the module will exclude them from the scan. This allows searching for additional patterns that contribute to unfair outcomes.
To get started, run the module without setting any sensitive attributes. After the first scan, advanced users can review the results and mark any problematic attributes it identifies as sensitive. Then, run the scan again to uncover additional potential issues—these may be less prominent but still worth investigating.
Technical details
A paper describing how this approach estimates biased intersection candidates in linear rather than exponential time is available here. Instead of checking every possible combination (which can be very time-consuming), it uses a more efficient method.
For convenience, there is a discovery mode available in the parameters. This automatically adds attributes suspected of contributing to bias to the list of ignored (already known sensitive) ones, then reruns the scan. While this automation helps streamline the process, it removes all attributes contributing to biased intersections. A domain expert may prefer to manually remove one attribute at a time by adding it to known sensitive attributes and rerun the module to investigate more granular effects on the results.
ParametersImage bias analysis

for data scientists: solutions for imbalanced image models
This module provides a comprehensive solution for analyzing image bias and recommending effective mitigation strategies. It can be used for both classification tasks (e.g., facial attribute extraction) and face verification. The core functionality revolves around evaluating how well different population groups, defined by a given protected attribute (such as gender, age, or ethnicity), are represented in the dataset. Representation bias occurs when some groups are overrepresented or underrepresented, leading to models that may perform poorly or unfairly on certain groups.
Additionally, the module detects spurious correlations between the target attribute (e.g., the label a model is trying to predict) and other annotated attributes (such as image features like color or shape). Spurious correlations are misleading patterns that do not reflect meaningful relationships and can cause a model to make biased or inaccurate predictions. By identifying and addressing these hidden biases, the module helps improve the fairness and accuracy of your model.
When you run the analysis, the module identifies specific biases within the dataset and suggests tailored mitigation approaches. Specifically, the suitable mitigation methodologies are determined based on the task and the types of the detected biases in the data. The analysis is conducted based on the CV Bias Mitigation Library.
ParametersFacex regions
![]()
facial features contributing to image categorization
FaceX is used to analyze 19 key face regions, such as the eyes, nose, mouth, hair, and skin. Then it identifies which parts of the face the model focuses on when making predictions about attributes like age, gender, or race.
GPU access is recommended for large datasets.
Technical approach at a glance.
Why is this useful?
Rather than explaining each individual image separately, FaceX aggregates information across the entire dataset, offering a broader view of the model's behavior. It looks at how the model activates different regions of the face for each decision. This aggregated information helps you see which facial features are most influential in the model's predictions, and whether certain features are being emphasized more than others.
In addition to providing an overall picture of which regions are important, FaceX also zooms in on specific areas of the face - such as a section of the skin or a part of the hair - showing which patches of the image have the highest impact on the model's decision. This makes it easier to identify potential biases or problems in how the model is interpreting the face.
Overall, with FaceX, you can quickly and easily get a better understanding of your model's decision-making process. This is especially useful for ensuring that your model is fair and transparent, and for spotting any potential biases that may affect its performance.
ParametersFacex embeddings
![]()
facial features contributing to image comparisons
FaceX is used to analyze
19 key face regions, such as the eyes, nose, mouth,
hair, and skin. Then it identifies where face verification models would focus
on matching pairs of images.
Technical approach at a glance.
Why is this useful?
Image embeddings are analyzed. In tasks like face recognition, models
generate a feature vector (embedding) for each image. These embeddings capture the unique
characteristics of the image and can be compared to determine how similar or different two images
are. FaceX helps explain which parts of an image contribute to its similarity or difference to a
reference embedding, allowing you to understand the model's focus on specific facial
features during the comparison.
The key idea is that FaceX analyzes the facial regions in the image that most influence how the model compares the reference embedding with the new image's embedding. Rather than providing an explanation for individual images in isolation, FaceX aggregates information across the dataset, offering insights into how different parts of the face, such as the eyes, mouth, or hair, contribute to the similarity or difference between the reference and the image being compared.
FaceX works by using a "reference" image (e.g., identity image) and evaluating how other images (e.g., selfies) compare to it. It looks at how various facial regions of the new image align with the reference image in the feature space. The tool then highlights which facial regions are most influential in the comparison, showing what aspects of the image are similar to or different from the concept embedding.
Overall, FaceX gives you a better understanding of how the model processes and compares facial features by highlighting the specific regions that influence the similarity or difference in feature embeddings. This is especially useful for improving transparency and identifying potential biases in how face verification models represent and compare faces.
GPU access is recommended for large datasets.
ParametersMulti objective report

for data scientists: interactive trade-off exploration
Presents an interactive Plotly 3D plot visualizing multiple objectives to evaluate model fairness and performance trade-offs. The report highlights three primary objectives: accuracy loss, balanced accuracy loss, and discrimination (MMM-fairness) loss. Each point plotted within the 3D space represents a Pareto-optimal solution, which achieves an optimal balance between these objectives where no single objective can improve without worsening another.
Users can hover over any solution point to display the corresponding loss values for each objective. Additionally, each point includes a theta value, indicating up to which sequence in the ONNX ensemble the particular solution is achieved. This allows users to observe performance changes throughout different stages of the ensemble, helping them better understand the trade-offs involved in each model configuration.
The multi-objective report generates predictions at each step of the partial ensemble. This may result in slower processing times when the number of Pareto solutions is high.
Viz fairness plots
structured visualization of common types of bias
This module visualizes fairness metrics using the Fairlearn library and interactive Plotly charts. It provides visual insights into how a model performs across different groups defined by sensitive features such as gender, race, or age.
What to expect?
Two sets of visual outputs are processed:
- Group-wise metrics: Shown as grouped bar charts, these display performance metrics (e.g., false positive rate) across subgroups.
- Scalar metrics: Displayed as horizontal bar charts, these summarize disparities (e.g., equal opportunity difference) in a compact, interpretable format.
Interactive charts allow users to hover for precise values, compare metrics between groups, and quickly identify fairness gaps. An explanation panel is included to define each metric and guide interpretation. This module is well suited for exploratory analysis, presentations, and fairness monitoring. It makes group disparities visible and intuitive, helping identify where further scrutiny or mitigation may be needed.
Viz fairness report
structured report on common types of bias
This module generates a structured fairness report using the Fairlearn library. It assesses whether a machine learning model behaves similarly across different population groups, as defined by sensitive attributes such as gender, race, or age.
What to expect?
Two types of fairness metrics are considered:
- Group-wise metrics: These show how the model performs for each group separately (e.g., true positive rates for Group A vs. Group B).
- Scalar metrics: These summarize disparities across groups into single numeric values. Small differences and ratios close to 1 indicate balanced treatment.
Results are presented in aligned tables with clear formatting, allowing users to compare outcomes across groups at a glance. Each metric is briefly explained to help interpret whether the model exhibits performance or outcome disparities for different groups. This module is particularly useful in evaluation pipelines, audit reports, and model reviews where transparency and fairness are essential. It helps teams assess group-level equity in model behavior using interpretable, tabular summaries.
Croissant
![]()
for data scientists: croissant specification
Generate some json dataset metadata that bootstraps conversion of your datasets into the Croissant format. That format is used to standardized how datasets may be indexed and loaded. If your dataset is stored locally, such as in minio instances, you can consider either sharing the metadata to explain to others what you are working with, or using publicly hosted data by providing https links for files. Metadata are displayed as HTML to help you get an overview and are presented as a copy-able block of json.
ParametersLanguage License Name Description Citation Qualitative creators Distribution
Exposure distance comparison
for network analysts: exposure distance
Compute the exposure distance between the protected and non-protected groups in the dataset and ranking. Sensitive attributes is a comma-separated list of the attributes relevant for fairness analysis. Currently, only Gender is supported.
ParametersAugmentation report

intersectional representation imbalances in data
This module uses the MMM-fair library to generate an interactive sunburst pie chart of dataset imbalances under class and sensitive attribute intersections. This helps to quickly identify representational imbalances in the dataset, allowing users to assess potential biases and identify areas that may require intervention, such augmentation.
Recommends data augmentation strategies.
Results for experts include bar charts for comparing different augmentation strategies designed to
mitigate data imbalances per sensitive attribute. These strategies adjust the distribution of attributes
in the dataset by oversampling specific subgroups with synthetic samples,
ensuring more equitable representation of sensitive attributes and target classes. Investigation includes
the following options:
- Class: Balances the class distribution within each subgroup by sampling the minority class.
- Class & Protected: Ensures equal sample distribution across all subgroups by sampling both majority
and minority classes.
- Protected: Balances the number of instances across different groups without considering class labels.
- Class (Ratio): Maintains the same class ratio across all groups as found in the largest group.
These visualizations allow users to observe the effects of each strategy on the data distribution, helping to understand how the dataset is being augmented.
For more information, refer to our full paper: "Synthetic Tabular Data Generation for Class Imbalance and Fairness: A Comparative Study" Link to paper.
Why is this needed?
This report provides is generated by MAI-BIAS to generate an overview of dataset imbalances across
the intersection of sensitive attributes and prediction targets using the MMM-Fair library.
Intersectionality emphasizes that people experience overlapping systems of discrimination
based on multiple identity characteristics (race, gender, class, sexual orientation,
disability, etc.). This is reflected also in how AI systems reproduce forms of discrimination.
As an example of intersectional bias [1] race and
gender together affect algorithmic performance of commercial facial-analysis systems;
worst performance for darker-skinned women demonstrates a compounded disparity
that would be missed if the analysis looked only at race or only at gender.
[1] Buolamwini, J., & Gebru, T. (2018, January). Gender shades: Intersectional accuracy disparities
in commercial gender classification. In Conference on fairness, accountability and transparency
(pp. 77-91). PMLR.
Imbalanced datasets can lead to biased model behavior, as underrepresented subgroups are more difficult to predict correctly. Intersections of many sensitive attributes (e.g., low-income hispanic woman) may create tiny or empty groups. Augmentation strategies increase the representation of such subgroups, which can improve fairness and robustness of downstream models.
ParametersText debias
This module uses DBias library to perform unsupervised auditing of text biases in text resembling article titles. If library identifies biases, it is used to mitigate them with a more neutral phrasing. Results show a judgement and prediction confidence for the original and adjusted text, as well as potentially offending keywords.
The DBias library employs three transformer models, one pretrained for the English language and two trained on the MBIC dataset.
Llm audit
![]()
use an LLM as text auditor
This assessment methodology sets an LLM at the role of fairness auditor and asks it to provide a sequence of votes, obtaining an assessment of whether given text is biased or neutral. Then, it follows a chain-of-thought approach for summarizing the reasoning associated with all valid votes (some votes may be invalid due to erroneous formatting) and eventually identifying actionable insights or explanations.
Parameters