Analysis

Card

Generates a fairness and bias report using the FairBench library. This explores many kinds of bias to paint a broad picture and help you decide on what is problematic and what is acceptable behavior. The generated report can be viewed in three different formats, where the model card contains a subset of results but attaches to these socio-technical concerns to be taken into account:

  1. A summary table of results.
  2. A simplified model card that includes concerns.
  3. The full report, including details.

The module's report summarizes how a model behaves on a provided dataset across different population groups. These groups are based on sensitive attributes like gender, age, and race. Each attribute can have multiple values, such as several genders or races. Numeric attributes, like age, are normalized to the range [0,1] and treated as fuzzy values, where 0 indicates membership to a fuzzy group of "small" values, and 1 indicates membership to a fuzzy group of "large" values. A separate set of fairness metrics is calculated for each prediction label.

If intersectional subgroup analysis is enabled, separate subgroups are created for each combination of sensitive attribute values. However, if there are too many attributes, some groups will be small or empty. Empty groups are ignored in the analysis. The report may also include information about built-in datasets.

Parameters

Specific concerns

Computes a fairness or bias measure that matches a specific type of numerical evaluation using the FairBench library. The measure is built by combining simpler options to form more than 300 valid alternatives.

This computes a specific fairness concerns and does not paint a broad enough picture. Make sure that you explore prospective biases with other modules first, like model card.

The assessment is conducted over sensitive attributes like gender, age, and race. Each attribute can have multiple values, such as several genders or races. Numeric attributes, like age, are normalized to the range [0,1] and treated as fuzzy values, where 0 indicates membership to a fuzzy group of "small" values, and 1 indicates membership to a fuzzy group of "large" values. A separate set of fairness metrics is calculated for each prediction label.

If intersectional subgroup analysis is enabled, separate subgroups are created for each combination of sensitive attribute values. However, if there are too many attributes, some groups will be small or empty. Empty groups are ignored in the analysis. The report may also include information about built-in datasets.

Parameters

Interactive report

Creates an interactive report using the FairBench library. The report creates traceable evaluations that you can shift through to find actual sources of unfairness.

Parameters

Sklearn audit

One way to evaluate the fairness of a dataset is by testing for biases using simple models with limited degrees of freedom. This module audits datasets by training such models provided by the scikit-learn library on half of the dataset. The second half is then used as test data to assess predictive performance and detect classification or scoring biases.

Test data are used to generate a fairness and bias report with the FairBench library. If strong biases appear in the simple models that are explored, they may also persist in more complex models trained on the same data. To focus on the most significant biases, adjust the minimum shown deviation parameter. The report provides multiple types of fairness and bias assessments and can be viewed in three different formats, where the model card contains a subset of results but attaches to these socio-technical concerns to be taken into account:

  1. A summary table of results.
  2. A simplified model card with key fairness concerns.
  3. A full detailed report.

The module's report summarizes how a model behaves on a provided dataset across different population groups. These groups are based on sensitive attributes like gender, age, and race. Each attribute can have multiple values, such as several genders or races. Numeric attributes, like age, are normalized to the range [0,1] and treated as fuzzy values, where 0 indicates membership to a fuzzy group of "small" values, and 1 indicates membership to a fuzzy group of "large" values. A separate set of fairness metrics is calculated for each prediction label.

If intersectional subgroup analysis is enabled, separate subgroups are created for each combination of sensitive attribute values. However, if there are too many attributes, some groups will be small or empty. Empty groups are ignored in the analysis. The report may also include information about built-in datasets.

Parameters

Optimal transport

Creates an optimal transport evaluation based on the implementation provided by the AIF360 library. The evaluation computes the Wasserstein distance that reflects the cost of transforming the predictive distributions between sensitive attribute groups.

Optimal Transport (OT) is a field of mathematics which studies the geometry of probability spaces. Among its many contributions, OT provides a principled way to compare and align probability distributions by taking into account the underlying geometry of the considered metric space. As a mathematical problem, it was first introduced by Gaspard Monge in 1781. It addresses the task of determining the most efficient method for transporting mass from one distribution to another. In this problem, the cost associated with moving a unit of mass from one position to another is referred to as the ground cost. The primary objective of OT is to minimize the total cost incurred when moving one mass distribution onto another.

OT can be used to detect model-induced bias by calculating the a cost known as Earth Mover's distance or Wasserstein distance between the distribution of ground truth labels and model predictions for each of the protected groups. If its value is close to 1, the model is biased towards this group.

License

Parts of the above description are adapted from AIF360 (https://github.com/Trusted-AI/AIF360), which is licensed under Apache License 2.0.

Parameters

Bias scan

This module scans your dataset to estimate the most biased attributes or combinations of attributes. For example, gender may only show bias when combined with socioeconomic status, despite the latter not bein inherently sensitive. If you have already marked some attributes as sensitive (such as race or gender), the module will **exclude** them from the scan. This allows searching for additional patterns that contribute to unfair outcomes.

A paper describing how this approach estimates biased intersection candidates in **linear** rather than **exponential** time is available here. Instead of checking every possible combination (which can be very time-consuming), it uses a more efficient method.

To get started, run the module without setting any sensitive attributes. After the first scan, advanced users can review the results and mark any problematic attributes it identifies as sensitive. Then, run the scan again to uncover additional potential issues—these may be less prominent but still worth investigating.

For convenience, there is a discovery mode available in the parameters. This automatically adds attributes suspected of contributing to bias to the list of ignored (already known sensitive) ones, then reruns the scan. While this automation helps streamline the process, it removes all attributes contributing to biased intersections. A domain expert may prefer to manually remove one attribute at a time by adding it to known sensitive attributes and rerun the module to investigate more granular effects on the results.

Parameters

Image bias analysis

This module provides a comprehensive solution for analyzing image bias and recommending effective mitigation strategies. It can be used for both classification tasks (e.g., facial attribute extraction) and face verification. The core functionality revolves around evaluating how well different population groups, defined by a given protected attribute (such as gender, age, or ethnicity), are represented in the dataset. Representation bias occurs when some groups are overrepresented or underrepresented, leading to models that may perform poorly or unfairly on certain groups.

Additionally, the module detects spurious correlations between the target attribute (e.g., the label a model is trying to predict) and other annotated attributes (such as image features like color or shape). Spurious correlations are misleading patterns that do not reflect meaningful relationships and can cause a model to make biased or inaccurate predictions. By identifying and addressing these hidden biases, the module helps improve the fairness and accuracy of your model.

When you run the analysis, the module identifies specific biases within the dataset and suggests tailored mitigation approaches. Specifically, the suitable mitigation methodologies are determined based on the task and the types of the detected biases in the data. The analysis is conducted based on the CV Bias Mitigation Library.

Parameters

Facex regions

FaceX is a tool designed to help you understand how face attribute classifiers make decisions. It provides clear explanations by analyzing 19 key regions of the face, such as the eyes, nose, mouth, hair, and skin. This method helps reveal which parts of the face the model focuses on when making predictions about attributes like age, gender, or race.

Rather than explaining each individual image separately, FaceX aggregates information across the entire dataset, offering a broader view of the model's behavior. It looks at how the model activates different regions of the face for each decision. This aggregated information helps you see which facial features are most influential in the model's predictions, and whether certain features are being emphasized more than others.

In addition to providing an overall picture of which regions are important, FaceX also zooms in on specific areas of the face - such as a section of the skin or a part of the hair - showing which patches of the image have the highest impact on the model's decision. This makes it easier to identify potential biases or problems in how the model is interpreting the face.

Overall, with FaceX, you can quickly and easily get a better understanding of your model's decision-making process. This is especially useful for ensuring that your model is fair and transparent, and for spotting any potential biases that may affect its performance.

GPU access is recommended as this analysis can be computationally intensive, especially with large datasets.

Parameters

Facex embeddings

FaceX for feature extractors is designed to help you understand how face verification models process images by comparing feature embeddings. In tasks like face recognition, models generate a feature vector (embedding) for each image. These embeddings capture the unique characteristics of the image and can be compared to determine how similar or different two images are. FaceX helps explain which parts of an image contribute to its similarity or difference to a reference embedding, allowing you to understand the model's focus on specific facial features during the comparison.

In this context, the key idea is that FaceX analyzes the facial regions in the image that most influence how the model compares the reference embedding with the new image's embedding. Rather than providing an explanation for individual images in isolation, FaceX aggregates information across the dataset, offering insights into how different parts of the face, such as the eyes, mouth, or hair, contribute to the similarity or difference between the reference and the image being compared.

FaceX works by using a "reference" image (e.g., identity image) and evaluating how other images (e.g., selfies) compare to it. It looks at how various facial regions of the new image align with the reference image in the feature space. The tool then highlights which facial regions are most influential in the comparison, showing what aspects of the image are similar to or different from the concept embedding.

Overall, FaceX gives you a better understanding of how the model processes and compares facial features by highlighting the specific regions that influence the similarity or difference in feature embeddings. This is especially useful for improving transparency and identifying potential biases in how face verification models represent and compare faces.

GPU access is recommended as this analysis can be computationally intensive, especially with large datasets.

Parameters

Multi objective report

This module presents an interactive Plotly 3D plot visualizing multiple objectives to evaluate model fairness and performance trade-offs. The report highlights three primary objectives: accuracy loss, balanced accuracy loss, and discrimination (MMM-fairness) loss. Each point plotted within the 3D space represents a Pareto-optimal solution, which achieves an optimal balance between these objectives where no single objective can improve without worsening another.

Users can hover over any solution point to display the corresponding loss values for each objective. Additionally, each point includes a theta value, indicating up to which sequence in the ONNX ensemble the particular solution is achieved. This allows users to observe performance changes throughout different stages of the ensemble, helping them better understand the trade-offs involved in each model configuration.

The multi-objective report generates predictions at each step of the partial ensemble. This may result in slower processing times when the number of Pareto solutions is high.

Viz fairness plots

This module visualizes fairness metrics using the Fairlearn library and interactive Plotly charts. It provides visual insights into how a model performs across different groups defined by sensitive features such as gender, race, or age.

The module produces two sets of visual outputs:

  • Group-wise metrics: Shown as grouped bar charts, these display performance metrics (e.g., false positive rate) across subgroups.
  • Scalar metrics: Displayed as horizontal bar charts, these summarize disparities (e.g., equal opportunity difference) in a compact, interpretable format.

Interactive charts allow users to hover for precise values, compare metrics between groups, and quickly identify fairness gaps. An explanation panel is included to define each metric and guide interpretation.

This module is well suited for exploratory analysis, presentations, and fairness monitoring. It makes group disparities visible and intuitive, helping identify where further scrutiny or mitigation may be needed.

Viz fairness report

This module generates a structured fairness report using the Fairlearn library. It assesses whether a machine learning model behaves similarly across different population groups, as defined by sensitive attributes such as gender, race, or age.

The report includes two types of fairness metrics:

  • Group-wise metrics: These show how the model performs for each group separately (e.g., true positive rates for Group A vs. Group B).
  • Scalar metrics: These summarize disparities across groups into single numeric values. Small differences and ratios close to 1 indicate balanced treatment.

Results are presented in aligned tables with clear formatting, allowing users to compare outcomes across groups at a glance. Each metric is briefly explained to help interpret whether the model exhibits performance or outcome disparities for different groups.

This module is particularly useful in evaluation pipelines, audit reports, and model reviews where transparency and fairness are essential. It helps teams assess group-level equity in model behavior using interpretable, tabular summaries.

Exposure distance comparison

Compute the exposure distance between the protected and non-protected groups in the dataset and ranking. Sensitive attributes is a comma-separated list of the attributes relevant for fairness analysis. Currently, only Gender is supported.

Parameters

Augmentation report

This module generates an interactive HTML report featuring a Plotly sunburst pie chart visualization to explore imbalances in the dataset based on subgroups defined by sensitive attributes and the target variable. This helps to quickly identify any potential imbalances in the dataset, allowing users to assess fairness and identify areas that may require intervention, such augmentation.

Furthermore, the report visually presents different augmentation strategies designed to mitigate data imbalances, per sensitive attribute, as an interactive Plotly bar chart. These strategies adjust the distribution of the dataset by oversampling specific subgroups with synthetic samples, ensuring more equitable representation of sensitive attributes and target classes. The following augmentation strategies are visualized:

  • Class: Balances the class distribution within each subgroup by sampling the minority class.
  • Class & Protected: Ensures equal sample distribution across all subgroups by sampling both majority and minority classes.
  • Protected: Balances the number of instances across different groups without considering class labels.
  • Class (Ratio): Maintains the same class ratio across all groups as found in the largest group.

These visualizations allow users to observe the effects of each strategy on the data distribution, helping to understand how the dataset is being augmented.

For more information, refer to our full paper: "Synthetic Tabular Data Generation for Class Imbalance and Fairness: A Comparative Study" Link to paper.

Text debias

This module uses DBias library to perform unsupervised auditing of text biases in text resembling article titles. If library identifies biases, it is used to mitigate them with a more neutral phrasing. Results show a judgement and prediction confidence for the original and adjusted text, as well as potentially offending keywords.

The DBias library employs three transformer models, one pretrained for the English language and two trained on the MBIC dataset.