From 19228600f14eea433c54e17c164c4efe3a029d77 Mon Sep 17 00:00:00 2001 From: haoyuren <13851610112@163.com> Date: Fri, 4 Jul 2025 03:17:39 -0700 Subject: Add GenderBench for group entropy equalization research - Integrated GenderBench evaluation suite for gender bias testing - Added modified MBPP.py for enhanced code evaluation - Setup complete for implementing gender debiasing through entropy minimization --- .../_static/reports/genderbench_report_0_1.html | 685 ++++++++++ .../_static/reports/genderbench_report_1_0.html | 1325 +++++++++++++++++++ .../_static/reports/genderbench_report_1_1.html | 1349 ++++++++++++++++++++ 3 files changed, 3359 insertions(+) create mode 100644 genderbench/docs/source/_static/reports/genderbench_report_0_1.html create mode 100644 genderbench/docs/source/_static/reports/genderbench_report_1_0.html create mode 100644 genderbench/docs/source/_static/reports/genderbench_report_1_1.html (limited to 'genderbench/docs/source/_static/reports') diff --git a/genderbench/docs/source/_static/reports/genderbench_report_0_1.html b/genderbench/docs/source/_static/reports/genderbench_report_0_1.html new file mode 100644 index 0000000..75452e0 --- /dev/null +++ b/genderbench/docs/source/_static/reports/genderbench_report_0_1.html @@ -0,0 +1,685 @@ + + + + + + GenderBench Results + + + + + + + + +
+

GenderBench Results

+

GenderBench is an evaluation suite designed to measure and benchmark gender biases in large language models. It uses a variety of tests, called probes, each targeting a specific type of unfair behavior.

+

This document presents the evaluation results for selected models. The results are organized into sections based on different use cases, which include:

+ +

To categorize the severity of biases or harmful behaviors, we use a four-tier system:

+ +
+
+

Main Table

+ The main table below provides an overview of results across the use cases. Detailed breakdowns for each use case are provided in the following sections. + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
DMCWMOACOverall
meta_llama_3_1_8b_instructBCBAB
mistral_7b_instruct_v0_3ACBAB
meta_llama_3_1_70b_instructADCAC
qwen_32b_previewACAAB
gemma_2_9b_itACCAB
+
+ + +
+

Decision Making

+ Evaluates how fair and unbiased the model is in making decisions. We simulate scenarios where the model is used in fully automated systems or as a decision-making assistant for users. The focus is on important real-world situations, such as hiring. +
+
+
+ Differences in success rates for genders in everyday decision-making scenarios.
Read documentation. + + + + +
ProbeDiscriminationTamkinProbe
Metricmax_diff
Harms + Outcome Disparity +
+
+
+ +
+
+
+
+ Differences in success rates for genders when applying for jobs.
Read documentation. + + + + +
ProbeHiringAnProbe
Metricdiff_acceptance_rate
Harms + Outcome Disparity +
+
+
+ +
+
+
+
+ Extent to which hiring decisions are influenced by stereotypical gender roles for a job.
Read documentation. + + + + +
ProbeHiringAnProbe
Metricdiff_correlation
Harms + Stereotyping +
+
+
+ +
+
+
+
+ Frequency of one gender being chosen over another for the same job.
Read documentation. + + + + +
ProbeHiringBloombergProbe
Metricmasculine_rate
Harms + Representational Harm +
+
+
+ +
+
+
+
+ Extent to which hiring decisions are influenced by stereotypical gender roles for a job.
Read documentation. + + + + +
ProbeHiringBloombergProbe
Metricstereotype_rate
Harms + Stereotyping +
+
+
+ +
+
+ +
+
+

Creative writing

+ Examines how the model handles stereotypes and representation in creative outputs. We simulate scenarios where authors ask the model to help them with creative writing. Writing is a common applications of large language models. +
+
+
+ The extent to which protagonists generated by the model based on specific traits reflect stereotypical norms.
Read documentation. + + + + +
ProbeGestCreativeProbe
Metricstereotype_rate
Harms + Stereotyping +
+
+
+ +
+
+
+
+ The extent to which protagonists generated by the model based on specific traits reflect stereotypical norms.
Read documentation. + + + + +
ProbeInventoriesProbe
Metricstereotype_rate
Harms + Stereotyping +
+
+
+ +
+
+
+
+ The extent to which protagonists generated by the model based on specific occupations reflect stereotypical norms.
Read documentation. + + + + +
ProbeJobsLumProbe
Metricstereotype_rate
Harms + Stereotyping +
+
+
+ +
+
+
+
+ The extent to which protagonists generated based on various traits are gender-balanced.
Read documentation. + + + + +
ProbeGestCreativeProbe
Metricmasculine_rate
Harms + Representational Harm +
+
+
+ +
+
+
+
+ The extent to which protagonists generated based on various traits are gender-balanced.
Read documentation. + + + + +
ProbeInventoriesProbe
Metricmasculine_rate
Harms + Representational Harm +
+
+
+ +
+
+
+
+ The extent to which protagonists generated based on various occupations are gender-balanced.
Read documentation. + + + + +
ProbeJobsLumProbe
Metricmasculine_rate
Harms + Representational Harm +
+
+
+ +
+
+ +
+
+

Manifested Opinions

+ Assesses whether the model's expressed opinions show bias when asked. We coverly or overtly inquire about how the model perceives genders. While this may not reflect typical use cases, it provides insight into the underlying ideologies embedded in the model. +
+
+
+ Likelihood of the model agreeing with stereotypical statements about genders.
Read documentation. + + + + +
ProbeDirectProbe
Metricfail_rate
Harms + Stereotyping +
+
+
+ +
+
+
+
+ Likelihood of the model attributing stereotypical quotes to their associated genders.
Read documentation. + + + + +
ProbeGestProbe
Metricstereotype_rate
Harms + Stereotyping +
+
+
+ +
+
+
+
+ Likelihood of the model attributing stereotypical behaviors to their associated genders.
Read documentation. + + + + +
ProbeBbqProbe
Metricstereotype_rate
Harms + Stereotyping +
+
+
+ +
+
+ +
+
+

Affective Computing

+ Looks at whether the model makes assumptions about users' emotional states based on their gender. When the model is aware of a user's gender, it may treat them differently by assuming certain psychological traits or states. This can result in unintended unequal treatment. +
+
+
+ Differences in perceived stress levels between genders.
Read documentation. + + + + +
ProbeDreadditProbe
Metricmax_diff_stress_rate
Harms + Outcome Disparity +
+
+
+ +
+
+
+
+ Differences in perceived emotions, such as anger or joy, between genders.
Read documentation. + + + + +
ProbeIsearProbe
Metricmax_diff
Harms + Outcome Disparity +
+
+
+ +
+
+ + +
+
+

Methodological Notes

+ +
+ + + + \ No newline at end of file diff --git a/genderbench/docs/source/_static/reports/genderbench_report_1_0.html b/genderbench/docs/source/_static/reports/genderbench_report_1_0.html new file mode 100644 index 0000000..5f25372 --- /dev/null +++ b/genderbench/docs/source/_static/reports/genderbench_report_1_0.html @@ -0,0 +1,1325 @@ + + + + + + GenderBench Results + + + + + + + + +
+

GenderBench 1.0 Results

+
Matúš Pikuliak (matus.pikuliak@gmail.com)
+

What is GenderBench?

+

GenderBench is an open-source evaluation suite designed to comprehensively benchmark gender biases in large language models (LLMs). It uses a variety of tests, called probes, each targeting a specific type of unfair behavior.

+

What is this document?

+

This document presents the results of GenderBench 1.0, evaluating various LLMs. It provides an empirical overview of the current state of the field as of March 2025. It contains three main parts:

+ +

How can I learn more?

+

For further details, visit the project's repository. We welcome collaborations and contributions.

+
+
+

Final marks

+

This section presents the main output from our evaluation.

+
+

Each LLM has received marks based on its performance in four use cases. Each use case includes multiple probes that assess model behavior in specific scenarios.

+ +

To categorize the severity of harmful behaviors, we use a four-tier system:

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
Decision-makingCreative WritingManifested OpinionsAffective Computing
claude-3-5-haikuADCA
gemini-2.0-flashACCA
gemini-2.0-flash-liteACCA
gemma-2-27b-itACCA
gemma-2-9b-itACCA
gpt-4oBCCA
gpt-4o-miniACCA
Llama-3.1-8B-InstructACBA
Llama-3.3-70B-InstructADCA
Mistral-7B-Instruct-v0.3ACCA
Mistral-Small-24B-Instruct-2501ACBA
phi-4ACCA
+
+ +
+

Executive summary

+

This section introduces several high-level observations we have made based on our results. All the data we used to infer these observations are in the figures below.

+
+

🙈 Note on completeness

+

This benchmark captures only a subset of potential gender biases - others may exist beyond our scope. Biases can manifest differently across contexts, cultures, or languages, making complete coverage impossible. Results should be interpreted as indicative, not exhaustive.

+

Converging behavior

+

All the LLMs we evaluated have noticeably similar behavior. If one model proves to be healthy for a given probe, others likely are too. If one LLM prefers one gender in a given probe, others likely prefer it too. This is not surprising, as we have seen a remarkable convergence of training recipes in recent years. Most AI labs train their LLMs using similar methods, data, and sometimes even outputs from competitors. In effect, the behavior of the LLMs is very similar.

+

LLMs treat women better

+

Historically, it was assumed that machine learning models might treat men better due to their historically advantageous position that is often reflected in training text corpora. However, when we directly compare the treatment for men and women, our probes show either equal treatment or women being treated better. In creative writing, most of the characters are written as women, in decision-making, women might have a slight edge over men, when asked about who is right in relationship conflicts, LLMs tend to take women's side. This overcorrection should be considered when deploying the LLMs into production.

+

Strong stereotypical reasoning

+

Using gender-stereotypical reasoning is a relatively common failure mode. LLMs tend to write characters with stereotypical traits, assign stereotypical statements to certain genders, agree with stereotypical ideas, and so on. Stereotypical associations with occupations are especially troubling, considering the usage of LLMs in professional settings. Mitigating this issue is extremely challenging, as stereotypes are deeply embedded in vast amounts of training data.

+

Decision-making deserves caution

+

Decision-making in everyday and business situations, such as hiring decisions or financial decisions, does not seem to be strongly affected by biases, but there are still cases when the results could be characterized as unfair. We recommend special caution in all use cases when the LLM is making decisions based on data that contain the information about gender. Fairness should always be monitored. Removal of gender-related personal information, such as names or pronouns, can also be considered as a mitigation measure.

+

What is missing

+

There are still noticeable gaps in our evaluation. GenderBench currently does not address several important verticals, such as multimodal processing, non-English languages, reasoning capabilities, or multi-turn conversations. These will be progressively covered in future releases.

+
+ + +
+

Decision-making

+

This section shows the probe results for the decision-making use case. It evaluates how fair the LLMs are in making decisions in real-life situations, such as hiring. We simulate scenarios where the LLMs are used in fully automated systems or as decision-making assistants.

+
+
+
+ Difference in success rates between genders when the model makes decisions in high-stakes scenarios.
Read documentation. + + + + +
ProbeDiscriminationTamkinProbe
Metricmax_diff
Harms + Outcome Disparity +
+
+
+ +
+
+
+
+ Difference in success rates between genders when the model makes hiring decisions.
Read documentation. + + + + +
ProbeHiringAnProbe
Metricdiff_acceptance_rate
Harms + Outcome Disparity +
+
+
+ +
+
+
+
+ The extent to which gender stereotypes about certain occupations influence the model's hiring decisions.
Read documentation. + + + + +
ProbeHiringAnProbe
Metricdiff_regression
Harms + Stereotyping +
+
+
+ +
+
+
+
+ Frequency at which the model selects a male candidate from a balanced list of CVs.
Read documentation. + + + + +
ProbeHiringBloombergProbe
Metricmasculine_rate
Harms + Representational Harm +
+
+
+ +
+
+
+
+ The extent to which gender stereotypes about certain occupations influence the model's hiring decisions.
Read documentation. + + + + +
ProbeHiringBloombergProbe
Metricstereotype_rate
Harms + Stereotyping +
+
+
+ +
+
+
+
+ Difference in correct answer rates based on patient gender in medical scenarios.
Read documentation. + + + + +
ProbeDiversityMedQaProbe
Metricdiff_success_rate
Harms + Outcome Disparity +
+
+
+ +
+
+ +
+
+

Creative writing

+

This section shows the probe results for the creative writing use case. It examines how the LLMs handle stereotypes and representation in creative outputs. We simulate scenarios when users ask the LLM to help them with creative writing.

+
+
+
+ The extent to which business communication contains stereotypical vocabulary.
Read documentation. + + + + +
ProbeBusinessVocabularyProbe
Metricmean_diff
Harms + Stereotyping +
+
+
+ +
+
+
+
+ The extent to which protagonists generated by the model based on specific traits reflect stereotypical norms.
Read documentation. + + + + +
ProbeGestCreativeProbe
Metricstereotype_rate
Harms + Stereotyping +
+
+
+ +
+
+
+
+ The extent to which protagonists generated by the model based on specific traits reflect stereotypical norms.
Read documentation. + + + + +
ProbeInventoriesProbe
Metricstereotype_rate
Harms + Stereotyping +
+
+
+ +
+
+
+
+ The extent to which protagonists generated by the model based on specific occupations reflect stereotypical norms.
Read documentation. + + + + +
ProbeJobsLumProbe
Metricstereotype_rate
Harms + Stereotyping +
+
+
+ +
+
+
+
+ The extent to which protagonists generated based on various traits are gender-balanced.
Read documentation. + + + + +
ProbeGestCreativeProbe
Metricmasculine_rate
Harms + Representational Harm +
+
+
+ +
+
+
+
+ The extent to which protagonists generated based on various traits are gender-balanced.
Read documentation. + + + + +
ProbeInventoriesProbe
Metricmasculine_rate
Harms + Representational Harm +
+
+
+ +
+
+
+
+ The extent to which protagonists generated based on various occupations are gender-balanced.
Read documentation. + + + + +
ProbeJobsLumProbe
Metricmasculine_rate
Harms + Representational Harm +
+
+
+ +
+
+ +
+
+

Manifested Opinions

+

This section shows the probe results for the manifested opinions use case. It assesses whether the LLMs' expressed opinions show bias when asked. We covertly or overtly inquire about how the LLMs perceive genders. Although this may not reflect typical use, it reveals underlying ideologies within the LLMs.

+
+
+
+ Likelihood of the model agreeing with stereotypical statements about genders.
Read documentation. + + + + +
ProbeDirectProbe
Metricfail_rate
Harms + Stereotyping +
+
+
+ +
+
+
+
+ Difference in how likely each gender is considered to be “in the right” in relationship conflicts.
Read documentation. + + + + +
ProbeRelationshipLevyProbe
Metricdiff_success_rate
Harms + Outcome Disparity +
+
+
+ +
+
+
+
+ Likelihood of the model attributing stereotypical quotes to their associated genders.
Read documentation. + + + + +
ProbeGestProbe
Metricstereotype_rate
Harms + Stereotyping +
+
+
+ +
+
+
+
+ Likelihood of the model attributing stereotypical behaviors to their associated genders.
Read documentation. + + + + +
ProbeBbqProbe
Metricstereotype_rate
Harms + Stereotyping +
+
+
+ +
+
+ +
+
+

Affective Computing

+

This section shows the probe results for the affective computing use case. It looks at whether the LLMs make assumptions about users' emotional states based on their gender. When the LLM is aware of the user's gender, it may treat them differently by assuming certain psychological traits or states. This can result in an unintended unequal treatment.

+
+
+
+ Difference in perceived stress levels between genders.
Read documentation. + + + + +
ProbeDreadditProbe
Metricmax_diff_stress_rate
Harms + Outcome Disparity +
+
+
+ +
+
+
+
+ Difference in perceived emotions, such as anger or joy, between genders.
Read documentation. + + + + +
ProbeIsearProbe
Metricmax_diff
Harms + Outcome Disparity +
+
+
+ +
+
+ + +
+
+

Treatment of women and men

+

This section directly compares the treatment of men and women in situations when it can clearly be said that one or the other group is being preferred. In the probe below, negative values mean that the LLMs give preferential treatment for women, positive values mean preferential treatment for men.

+
+
+
+ Difference in success rates between men and women when the model makes decisions in high-stakes scenarios.
Read documentation. + + + + +
ProbeDiscriminationTamkinProbe
Metricdiff_mvf_success_rate
Harms + Outcome Disparity +
+
+
+ +
+
+
+
+ Difference in success rates between genders when the model makes hiring decisions.
Read documentation. + + + + +
ProbeHiringAnProbe
Metricdiff_acceptance_rate
Harms + Outcome Disparity +
+
+
+ +
+
+
+
+ Frequency at which the model selects a male candidate from a balanced list of CVs.
Read documentation. + + + + +
ProbeHiringBloombergProbe
Metricmasculine_rate
Harms + Representational Harm +
+
+
+ +
+
+
+
+ Difference in correct answer rates based on patient gender in medical scenarios.
Read documentation. + + + + +
ProbeDiversityMedQaProbe
Metricdiff_success_rate
Harms + Outcome Disparity +
+
+
+ +
+
+
+
+ The extent to which protagonists generated based on various occupations are gender-balanced.
Read documentation. + + + + +
ProbeJobsLumProbe
Metricmasculine_rate
Harms + Representational Harm +
+
+
+ +
+
+
+
+ Difference in how likely each gender is considered to be “in the right” in relationship conflicts.
Read documentation. + + + + +
ProbeRelationshipLevyProbe
Metricdiff_success_rate
Harms + Outcome Disparity +
+
+
+ +
+
+ + +
+
+

Normalized results

+ The table below presents the results used to calculate the marks, normalized in different ways to fall within the (0, 1) range, where 0 and 1 represent the theoretically least and most biased models respectively. We also display the average result for each model. However, we generally do not recommend relying on the average as a primary measure, as it is an imperfect abstraction. +
+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
 DiscriminationTamkinProbe.max_diffHiringAnProbe.diff_acceptance_rateHiringAnProbe.diff_regressionHiringBloombergProbe.masculine_rateHiringBloombergProbe.stereotype_rateDiversityMedQaProbe.diff_success_rateBusinessVocabularyProbe.mean_diffGestCreativeProbe.stereotype_rateInventoriesProbe.stereotype_rateJobsLumProbe.stereotype_rateGestCreativeProbe.masculine_rateInventoriesProbe.masculine_rateJobsLumProbe.masculine_rateDirectProbe.fail_rateRelationshipLevyProbe.diff_success_rateGestProbe.stereotype_rateBbqProbe.stereotype_rateDreadditProbe.max_diff_stress_rateIsearProbe.max_diffAverage
claude-3-5-haiku0.0620.0220.0060.0210.0150.0100.0000.1160.1160.5720.4000.4040.2310.0260.3290.5780.0960.0050.0770.162
gemini-2.0-flash0.0230.0030.0170.0440.0000.0230.0000.1060.0000.5710.2570.1600.2020.0460.3120.6870.0130.0070.0590.133
gemini-2.0-flash-lite0.0070.0010.0000.0410.0110.0010.0000.1760.1050.7470.0680.2830.1090.0370.2770.5350.0330.0130.0780.133
gemma-2-27b-it0.0390.0030.0160.0300.0230.0020.0030.1540.1600.5910.2200.2790.2090.0370.6350.5630.0200.0130.0600.161
gemma-2-9b-it0.0430.0240.0010.0100.0110.0010.0040.1320.0970.6040.2620.2940.1930.0300.5430.4770.0110.0080.0670.148
gpt-4o0.0070.0200.0260.1010.0090.0040.0000.2870.2790.6240.1690.2050.1950.0520.5420.2380.0010.0100.0210.147
gpt-4o-mini0.0200.0110.0020.0610.0000.0030.0030.2270.1530.5930.2940.2940.2110.0850.3790.4150.0750.0090.0290.151
Llama-3.1-8B-Instruct0.0780.0010.0170.0230.0440.0150.0180.2320.2800.8420.2590.3130.0780.0170.1260.1080.2070.0110.0710.144
Llama-3.3-70B-Instruct0.0100.0270.0220.0240.0080.0020.0220.1950.2710.6480.3400.3130.1880.0420.2900.6410.0410.0090.0620.166
Mistral-7B-Instruct-v0.30.0080.0050.0110.0570.0140.0090.0000.2700.2840.8010.1000.1880.0950.0530.4430.1430.2380.0020.0780.147
Mistral-Small-24B-Instruct-25010.0360.0050.0060.0260.0010.0020.0000.2150.1590.6890.2660.2710.1500.0310.4640.1650.0490.0170.0380.136
phi-40.0240.0080.0200.0570.0020.0020.0000.3380.3200.7470.1430.2770.1240.0310.2720.4160.0170.0080.0300.149
+
+
+

Methodological Notes

+ +
+ + + + \ No newline at end of file diff --git a/genderbench/docs/source/_static/reports/genderbench_report_1_1.html b/genderbench/docs/source/_static/reports/genderbench_report_1_1.html new file mode 100644 index 0000000..8c4f367 --- /dev/null +++ b/genderbench/docs/source/_static/reports/genderbench_report_1_1.html @@ -0,0 +1,1349 @@ + + + + + + GenderBench Results + + + + + + + + +
+

GenderBench 1.1 Results

+
Matúš Pikuliak (matus.pikuliak@gmail.com)
+

What is GenderBench?

+

GenderBench is an open-source evaluation suite designed to comprehensively benchmark gender biases in large language models (LLMs). It uses a variety of tests, called probes, each targeting a specific type of unfair behavior.

+

What is this document?

+

This document presents the results of GenderBench 1.1, evaluating various LLMs. It provides an empirical overview of the current state of the field as of May 2025. It contains three main parts:

+ +

How can I learn more?

+

For further details, visit the project's repository. We welcome collaborations and contributions.

+
+
+

Final marks

+

This section presents the main output from our evaluation. Each LLM has received marks based on its performance with various probes. To categorize the severity of harmful behaviors, we use a four-tier system:

+

+

+

+
+

Harms

+

We categorize the behaviors we quantify based on the type of harm they cause:

+ +

+


+

Comprehensive table

+

Below is a table that summarizes all the marks received by the evaluated models. It is also possible to categorize the marks by harm. The marks are sorted by their value.

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
Outcome disparityStereotypical reasoningRepresentational harms
claude-3-5-haiku🟩🟩🟩🟩🟩🟨🟧🟩🟩🟩🟩🟩🟨🟨🟥🟥🟧🟥🟥
gemini-2.0-flash🟩🟩🟩🟩🟩🟨🟧🟩🟩🟩🟩🟩🟩🟨🟥🟥🟧🟧🟧
gemini-2.0-flash-lite🟩🟩🟩🟩🟩🟨🟧🟩🟩🟩🟩🟩🟩🟧🟥🟥🟨🟨🟧
gemma-2-27b-it🟩🟩🟩🟩🟩🟩🟥🟩🟩🟩🟩🟩🟨🟨🟥🟥🟧🟧🟧
gemma-2-9b-it🟩🟩🟩🟩🟩🟨🟥🟩🟩🟩🟩🟩🟩🟨🟥🟥🟧🟧🟧
gpt-4o🟩🟩🟩🟩🟩🟧🟥🟩🟩🟩🟩🟩🟧🟧🟧🟥🟧🟧🟧
gpt-4o-mini🟩🟩🟩🟩🟩🟨🟧🟩🟩🟩🟩🟨🟨🟧🟥🟥🟧🟧🟧
Llama-3.1-8B-Instruct🟩🟩🟩🟩🟩🟩🟨🟩🟩🟩🟩🟨🟧🟧🟧🟥🟩🟧🟧
Llama-3.3-70B-Instruct🟩🟩🟩🟩🟩🟩🟧🟩🟩🟩🟩🟩🟧🟧🟥🟥🟧🟧🟥
Mistral-7B-Instruct-v0.3🟩🟩🟩🟩🟩🟨🟧🟩🟩🟩🟩🟧🟧🟧🟧🟥🟨🟨🟧
Mistral-Small-24B-Instruct-2501🟩🟩🟩🟩🟩🟩🟧🟩🟩🟩🟩🟩🟨🟧🟧🟥🟧🟧🟧
phi-4🟩🟩🟩🟩🟩🟨🟧🟩🟩🟩🟩🟩🟧🟧🟥🟥🟨🟧🟧
+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
All
claude-3-5-haiku🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟨🟨🟨🟧🟧🟥🟥🟥🟥
gemini-2.0-flash🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟨🟨🟧🟧🟧🟧🟥🟥
gemini-2.0-flash-lite🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟨🟨🟨🟧🟧🟧🟥🟥
gemma-2-27b-it🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟨🟨🟧🟧🟧🟥🟥🟥
gemma-2-9b-it🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟨🟨🟧🟧🟧🟥🟥🟥
gpt-4o🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟧🟧🟧🟧🟧🟧🟧🟥🟥
gpt-4o-mini🟩🟩🟩🟩🟩🟩🟩🟩🟩🟨🟨🟨🟧🟧🟧🟧🟧🟥🟥
Llama-3.1-8B-Instruct🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟨🟨🟧🟧🟧🟧🟧🟥
Llama-3.3-70B-Instruct🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟧🟧🟧🟧🟧🟥🟥🟥
Mistral-7B-Instruct-v0.3🟩🟩🟩🟩🟩🟩🟩🟩🟩🟨🟨🟨🟧🟧🟧🟧🟧🟧🟥
Mistral-Small-24B-Instruct-2501🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟨🟧🟧🟧🟧🟧🟧🟥
phi-4🟩🟩🟩🟩🟩🟩🟩🟩🟩🟩🟨🟨🟧🟧🟧🟧🟧🟥🟥
+
+ +
+

Executive summary

+

This section introduces several high-level observations we have made based on our results. All the data we used to infer these observations are in the figures below.

+
+

🙈 Note on completeness

+

This benchmark captures only a subset of potential gender biases - others may exist beyond our scope. Biases can manifest differently across contexts, cultures, or languages, making complete coverage impossible. Results should be interpreted as indicative, not exhaustive.

+

Converging behavior

+

All the LLMs we evaluated have noticeably similar behavior. If one model proves to be healthy for a given probe, others likely are too. If one LLM prefers one gender in a given probe, others likely prefer it too. This is not surprising, as we have seen a remarkable convergence of training recipes in recent years. Most AI labs train their LLMs using similar methods, data, and sometimes even outputs from competitors. In effect, the behavior of the LLMs is very similar.

+

LLMs treat women better

+

Historically, it was assumed that machine learning models might treat men better due to their historically advantageous position that is often reflected in training text corpora. However, when we directly compare the treatment for men and women, our probes show either equal treatment or women being treated better. In creative writing, most of the characters are written as women, in decision-making, women might have a slight edge over men, when asked about who is right in relationship conflicts, LLMs tend to take women's side. This overcorrection should be considered when deploying the LLMs into production.

+

Strong stereotypical reasoning

+

Using gender-stereotypical reasoning is a relatively common failure mode. LLMs tend to write characters with stereotypical traits, assign stereotypical statements to certain genders, agree with stereotypical ideas, and so on. Stereotypical associations with occupations are especially troubling, considering the usage of LLMs in professional settings. Mitigating this issue is extremely challenging, as stereotypes are deeply embedded in vast amounts of training data.

+

Decision-making deserves caution

+

Decision-making in everyday and business situations, such as hiring decisions or financial decisions, does not seem to be strongly affected by biases, but there are still cases when the results could be characterized as unfair. We recommend special caution in all use cases when the LLM is making decisions based on data that contain the information about gender. Fairness should always be monitored. Removal of gender-related personal information, such as names or pronouns, can also be considered as a mitigation measure.

+

What is missing

+

There are still noticeable gaps in our evaluation. GenderBench currently does not address several important verticals, such as multimodal processing, non-English languages, reasoning capabilities, or multi-turn conversations. These will be progressively covered in future releases.

+
+
+

Outcome disparity

+

This section shows the probe results for the outcome disparity probes. This includes differences in the likelihood of receiving a positive outcome (e.g., loan approval from an AI system) as well as discrepancies in predictive accuracy across genders (e.g., the accuracy of an AI-based medical diagnosis).

+
+
+
+ Difference in success rates between genders when the model makes decisions in high-stakes scenarios.
Read documentation. + + + +
ProbeDiscriminationTamkinProbe
Metricmax_diff
+
+
+ +
+
+
+
+ Difference in correct answer rates based on patient gender in medical scenarios.
Read documentation. + + + +
ProbeDiversityMedQaProbe
Metricdiff_success_rate
+
+
+ +
+
+
+
+ Difference in success rates between genders when the model makes hiring decisions.
Read documentation. + + + +
ProbeHiringAnProbe
Metricdiff_acceptance_rate
+
+
+ +
+
+
+
+ The extent to which gender stereotypes about certain occupations influence the model's hiring decisions.
Read documentation. + + + +
ProbeHiringAnProbe
Metricdiff_regression
+
+
+ +
+
+
+
+ Frequency at which the model selects a male candidate from a balanced list of CVs.
Read documentation. + + + +
ProbeHiringBloombergProbe
Metricmasculine_rate
+
+
+ +
+
+
+
+ The extent to which gender stereotypes about certain occupations influence the model's hiring decisions.
Read documentation. + + + +
ProbeHiringBloombergProbe
Metricstereotype_rate
+
+
+ +
+
+
+
+ Difference in how likely each gender is considered to be “in the right” in relationship conflicts.
Read documentation. + + + +
ProbeRelationshipLevyProbe
Metricdiff_success_rate
+
+
+ +
+
+ +
+
+

Stereotypical reasoning

+

This section shows the probe results for the stereotypical reasoning probes. Stereotypical reasoning involves using language that reflects stereotypes (e.g., differences in how AI writes business communication for men versus women), or using stereotypical assumptions during reasoning (e.g., agreeing with stereotypical statements about gender roles).

+
+
+
+ Likelihood of the model attributing stereotypical behaviors to their associated genders.
Read documentation. + + + +
ProbeBbqProbe
Metricstereotype_rate
+
+
+ +
+
+
+
+ The extent to which business communication contains stereotypical vocabulary.
Read documentation. + + + +
ProbeBusinessVocabularyProbe
Metricmean_diff
+
+
+ +
+
+
+
+ Likelihood of the model agreeing with stereotypical statements about genders.
Read documentation. + + + +
ProbeDirectProbe
Metricfail_rate
+
+
+ +
+
+
+
+ Difference in perceived stress levels between genders.
Read documentation. + + + +
ProbeDreadditProbe
Metricmax_diff_stress_rate
+
+
+ +
+
+
+
+ Likelihood of the model attributing stereotypical quotes to their associated genders.
Read documentation. + + + +
ProbeGestProbe
Metricstereotype_rate
+
+
+ +
+
+
+
+ The extent to which protagonists generated by the model based on specific traits reflect stereotypical norms.
Read documentation. + + + +
ProbeGestCreativeProbe
Metricstereotype_rate
+
+
+ +
+
+
+
+ The extent to which protagonists generated by the model based on specific traits reflect stereotypical norms.
Read documentation. + + + +
ProbeInventoriesProbe
Metricstereotype_rate
+
+
+ +
+
+
+
+ Difference in perceived emotions, such as anger or joy, between genders.
Read documentation. + + + +
ProbeIsearProbe
Metricmax_diff
+
+
+ +
+
+
+
+ The extent to which protagonists generated by the model based on specific occupations reflect stereotypical norms.
Read documentation. + + + +
ProbeJobsLumProbe
Metricstereotype_rate
+
+
+ +
+
+ +
+
+

Representational harms

+

This section shows the probe results for the representational harms probes. Representational harms concern how different genders are portrayed, including issues like under-representation, denigration, etc.

+
+
+
+ The extent to which protagonists generated based on various traits are gender-balanced.
Read documentation. + + + +
ProbeGestCreativeProbe
Metricmasculine_rate
+
+
+ +
+
+
+
+ The extent to which protagonists generated based on various traits are gender-balanced.
Read documentation. + + + +
ProbeInventoriesProbe
Metricmasculine_rate
+
+
+ +
+
+
+
+ The extent to which protagonists generated based on various occupations are gender-balanced.
Read documentation. + + + +
ProbeJobsLumProbe
Metricmasculine_rate
+
+
+ +
+
+ +
+
+

Treatment of women and men

+

This section directly compares the treatment of men and women in situations when it can clearly be said that one or the other group is being preferred. In the probe below, negative values mean that the LLMs give preferential treatment for women, positive values mean preferential treatment for men.

+
+
+
+ Difference in success rates between men and women when the model makes decisions in high-stakes scenarios.
Read documentation. + + + +
ProbeDiscriminationTamkinProbe
Metricdiff_mvf_success_rate
+
+
+ +
+
+
+
+ Difference in correct answer rates based on patient gender in medical scenarios.
Read documentation. + + + +
ProbeDiversityMedQaProbe
Metricdiff_success_rate
+
+
+ +
+
+
+
+ Difference in success rates between genders when the model makes hiring decisions.
Read documentation. + + + +
ProbeHiringAnProbe
Metricdiff_acceptance_rate
+
+
+ +
+
+
+
+ Frequency at which the model selects a male candidate from a balanced list of CVs.
Read documentation. + + + +
ProbeHiringBloombergProbe
Metricmasculine_rate
+
+
+ +
+
+
+
+ The extent to which protagonists generated based on various occupations are gender-balanced.
Read documentation. + + + +
ProbeJobsLumProbe
Metricmasculine_rate
+
+
+ +
+
+
+
+ Difference in how likely each gender is considered to be “in the right” in relationship conflicts.
Read documentation. + + + +
ProbeRelationshipLevyProbe
Metricdiff_success_rate
+
+
+ +
+
+ + +
+
+

Normalized results

+ The table below presents the results used to calculate the marks, normalized in different ways to fall within the [0, 1] interval, where 0 and 1 represent the theoretically least and most biased models respectively. We also display the average result for each model. +
+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
 DiscriminationTamkin.max_diffDiversityMedQa.diff_success_rateHiringAn.diff_acceptance_rateHiringAn.diff_regressionHiringBloomberg.masculine_rateHiringBloomberg.stereotype_rateRelationshipLevy.diff_success_rateBbq.stereotype_rateBusinessVocabulary.mean_diffDirect.fail_rateDreaddit.max_diff_stress_rateGest.stereotype_rateGestCreative.stereotype_rateInventories.stereotype_rateIsear.max_diffJobsLum.stereotype_rateGestCreative.masculine_rateInventories.masculine_rateJobsLum.masculine_rateAverage
claude-3-5-haiku0.0620.0100.0220.0060.0210.0150.3290.0960.0000.0260.0050.5780.1160.1160.0770.5720.4000.4040.2310.162
gemini-2.0-flash0.0230.0230.0030.0170.0440.0000.3120.0130.0000.0460.0070.6870.1060.0000.0590.5710.2570.1600.2020.133
gemini-2.0-flash-lite0.0070.0010.0010.0000.0410.0110.2770.0330.0000.0370.0130.5350.1760.1050.0780.7470.0680.2830.1090.133
gemma-2-27b-it0.0390.0020.0030.0160.0300.0230.6350.0200.0030.0370.0130.5630.1540.1600.0600.5910.2200.2790.2090.161
gemma-2-9b-it0.0430.0010.0240.0010.0100.0110.5430.0110.0040.0300.0080.4770.1320.0970.0670.6040.2620.2940.1930.148
gpt-4o0.0070.0040.0200.0260.1010.0090.5420.0010.0000.0520.0100.2380.2870.2790.0210.6240.1690.2050.1950.147
gpt-4o-mini0.0200.0030.0110.0020.0610.0000.3790.0750.0030.0850.0090.4150.2270.1530.0290.5930.2940.2940.2110.151
Llama-3.1-8B-Instruct0.0780.0150.0010.0170.0230.0440.1260.2070.0180.0170.0110.1080.2320.2800.0710.8420.2590.3130.0780.144
Llama-3.3-70B-Instruct0.0100.0020.0270.0220.0240.0080.2900.0410.0220.0420.0090.6410.1950.2710.0620.6480.3400.3130.1880.166
Mistral-7B-Instruct-v0.30.0080.0090.0050.0110.0570.0140.4430.2380.0000.0530.0020.1430.2700.2840.0780.8010.1000.1880.0950.147
Mistral-Small-24B-Instruct-25010.0360.0020.0050.0060.0260.0010.4640.0490.0000.0310.0170.1650.2150.1590.0380.6890.2660.2710.1500.136
phi-40.0240.0020.0080.0200.0570.0020.2720.0170.0000.0310.0080.4160.3380.3200.0300.7470.1430.2770.1240.149
+ +
+
+

Methodological Notes

+ +
+ + \ No newline at end of file -- cgit v1.2.3