Towards Region-aware Bias Evaluation Metrics (2024)

Angana Borah1 Aparna Garimella2 Rada Mihalcea1
1University of Michigan, Ann Arbor, USA,2Adobe Research, India

anganab@umich.edu garimell@adobe.com mihalcea@umich.edu

Abstract

When exposed to human-generated data, language models are known to learn and amplify societal biases. While previous works introduced benchmarks that can be used to assess the bias in these models, they rely on assumptions that may not be universally true. For instance, a gender bias dimension commonly used by these metrics is that of family–career, but this may not be the only common bias in certain regions of the world. In this paper, we identify topical differences in gender bias across different regions and propose a region-aware bottom-up approach for bias assessment. Our proposed approach uses gender-aligned topics for a given region and identifies gender bias dimensions in the form of topic pairs that are likely to capture gender societal biases. Several of our proposed bias topic pairs are on par with human perception of gender biases in these regions in comparison to the existing ones, and we also identify new pairs that are more aligned than the existing ones. In addition, we use our region-aware bias topic pairs in a Word Embedding Association Test (WEAT)-based evaluation metric to test for gender biases across different regions in different data domains. We also find that LLMs have a higher alignment to bias pairs for highly-represented regions showing the importance of region-aware bias evaluation metric.

Towards Region-aware Bias Evaluation Metrics


Angana Borah1 Aparna Garimella2 Rada Mihalcea11University of Michigan, Ann Arbor, USA,2Adobe Research, Indiaanganab@umich.edu garimell@adobe.com mihalcea@umich.edu


1 Introduction

Human bias refers to the tendency of prejudice or preference towards a certain group or an individual and can reflect social stereotypes with respect to gender, age, race, religion, and so on.

Bias in machine learning refers to prior information which is a necessary prerequisite for intelligence Bishop (2006). However, biases can be problematic when prior information is derived from harmful precedents like prejudices and social stereotypes. Early work in detecting biases includes the Word Embedding Association Test (WEAT) Caliskan etal. (2017) and the Sentence Encoder Association Test (SEAT) May etal. (2019). WEAT is inspired by the Implicit Association Test (IAT) Greenwald etal. (1998) in psychology, which gauges people’s propensity to unconsciously link particular characteristics—like family versus career—with specific target groups—like female (F) versus male (M).WEAT measures the distances between target and attribute word sets in word embeddings using dimensions111We use ‘topic pairs’ and ‘topic dimensions’ interchangeably. similar to those used in IAT.

Biases toward or against a group can vary across different regions due to the influence of an individual’s culture and demographics Grimm and Church (1999); Kiritchenko and Mohammad (2018a); Garimella etal. (2022). Psychological studies and experiments that demonstrate human stereotypes vary by continental regions Damann etal. (2023); Blog (2017) and even larger concepts like western and eastern worlds Markus and Kitayama (2003); Jiang etal. (2019) serve as an inspiration for the use of regions to determine differences across cultures. However, existing bias evaluation metrics like WEAT and SEAT follow a “one-size-fits-all” approach to detect biases across different regions.As biases can be very diverse depending on the demographic lens, a fixed or a small set of dimensions (such as family–career, math–arts) may not be able to cover all the possible biases in the society.In this paper, we address two main research questions about gender bias: (1) Is it possible to use current NLP techniques to automatically identify gender bias characteristics (such as family, career) specific to various regions? (2) How do these gender dimensions compare to the current generic dimensions included in WEAT/SEAT?

Our paper makes four main contributions:

  1. 1.

    An automatic method to uncover gender bias dimensions in various regions that uses (a) topic modeling to identify dominant topics aligning with the F/M groups for different regions, and (b) an embedding-based approach to identify F-M topic pairs for different regions that can be viewed as gender bias dimensions in those regions.

  2. 2.

    An IAT-style test to assess our predicted gender bias dimensions with human subjects. To the best of our knowledge, this is the first study to use a data-driven, bottom-up method to evaluate bias dimensions across regional boundaries.

  3. 3.

    A WEAT-based evaluation setup using our region-aware topic pairs to evaluate gender biases in different data domains (Reddit and UN General Debates) across regions.

  4. 4.

    An analysis of how well our predicted bias dimensions align with those of custom LLMs. We consider several LLMs that include open-source models like Llama-3-8b and Mistral-7b-Instruct; as well as closed-source models such as GPT-4, Gemini-Pro and Claude-3-Sonnet.

2 Data

We use GeoWAC Dunn and Adams (2020a), a geographically balanced corpus that consists of web pages from Common Crawl. Language samples are geo-located using country-specific domains, such as an .in.in. italic_i italic_n domain suggesting Indian origin Dunn and Adams (2020b). The GeoWAC’s English corpus spans 150 countries. We select the top three countries with the most examples per region: Asia, Africa, Europe, North America, and Oceania as in Garimella etal. (2022). We randomly choose 282,000 examples (after pre-processing) for each region, with 94,000 examples belonging to each country within the regions. Dataset details are included in AppendixA.

3 Variations in Gender Bias Tests Across Regions

We start by investigating the differences in existing gender bias tests across different regions using WEAT. WEAT takes in target words such as male names and female names, to indicate a specific group, and attribute words that can be associated with the target words, such as math and art. It computes bias by finding the cosine distance between the embeddings of the target and attribute words.We compute WEAT scores using word2vec embeddings Mikolov etal. (2013) trained on the five regions separately. Table1 shows the region-wise scores for the three gender tests in WEAT.

Target words - Attribute wordsRegionWEAT
Africa1.798
Asia1.508
career vs family - Male names vs Female namesNorth America1.885
Europe1.610
Oceania1.727
Africa1.429
Asia1.187
Math vs Arts - Male terms vs Female termsNorth America0.703
Europe0.334
Oceania1.158
Africa1.247
Asia0.330
Science vs Arts - Male terms vs Female termsNorth America0.036
Europe-0.655
Oceania0.725

Although we see a positive bias for most gender bias dimensions, the scores vary across regions.For example, the highest scoring regions vary for the target words-attribute words groups. For family–career dimension, North America shows the highest bias, however for math–arts and math–science dimensions, Africa shows the highest bias.Europe has a negative bias on science–arts (indicating a stronger F-science and M-arts association).

These results provide preliminary support to our hypothesis that gender bias dimensions vary across regions, thus propelling a need to come up with further bias measurement dimensions to better capture gender biases in these regions in addition to the existing generic ones in WEAT.

4 A Method to Automatically Detect Bias Dimensions Across Regions

Building upon our WEAT findings, we propose a two-stage approach to automatically detect region-aware bias dimensions that likely capture the biases in specific regions in a bottom-up manner.In the first stage, we utilize topic modeling to identify prominent topics in each region. In the second stage, we use an embedding-based approach to find pairs of topics among those identified in the first stage that are likely to represent prominent gender bias dimensions in each region.Fig1 shows the pipeline of our methodology.

Towards Region-aware Bias Evaluation Metrics (1)

4.1 Identifying Region-wise Bias Topics

We use topic modeling to identify dominant topics in the male and female examples in each region.

We first build F(emale)- and M(ale)-aligned datasets using the examples from GeoWAC for each region.We use the 52 pairs of gender-defined words that are non-stereotypically F/M (e.g., wife, brother, see AppendixG) from Bolukbasi etal. (2016), and find examples that contain these words.These datasets are used to find gender-aligned topics from GeoWAC.The dataset statisticsare specified in Table6 in AppendixB.

We then use topic modeling to identify dominant topics in the male and female examples in each region.We use Bertopic Grootendorst (2022), which identifies an optimal number of topics n𝑛nitalic_n for a given dataset (see AppendixL.1 for implementation details). We further refine the resulting topics using Llama2 Touvron etal. (2023) to label and better understand the topic clusters identified by Bertopic. The prompting mechanism for Llama2 is provided in AppendixH.

We next compute the alignment of the topics to either of the F/M groups. We first compute the topic distribution of a data point, which gives the probability pitsubscript𝑝𝑖𝑡p_{it}italic_p start_POSTSUBSCRIPT italic_i italic_t end_POSTSUBSCRIPT of an example i𝑖iitalic_i belonging to each topic t𝑡titalic_t. For a topic t𝑡titalic_t, we take n examples that dominantly belong to topic t𝑡titalic_t: i1,i2,.,ini_{1},i_{2},....,i_{n}italic_i start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , italic_i start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT , … . , italic_i start_POSTSUBSCRIPT italic_n end_POSTSUBSCRIPT. If m out of n data points belong to the F group in the F-M dataset, and the other (n - m) belongs to the M group, we compute the average of topic probabilities for both groups separately: pFt=(pi1t+pi2t++pimt)msubscript𝑝𝐹𝑡subscript𝑝subscript𝑖1𝑡subscript𝑝subscript𝑖2𝑡subscript𝑝subscript𝑖𝑚𝑡𝑚p_{Ft}=\frac{(p_{i_{1}t}+p_{i_{2}t}+......+p_{i_{m}t})}{m}italic_p start_POSTSUBSCRIPT italic_F italic_t end_POSTSUBSCRIPT = divide start_ARG ( italic_p start_POSTSUBSCRIPT italic_i start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT + italic_p start_POSTSUBSCRIPT italic_i start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT + … … + italic_p start_POSTSUBSCRIPT italic_i start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ) end_ARG start_ARG italic_m end_ARG and pMt=(pim+1t+pim+2t++pint)(nm)subscript𝑝𝑀𝑡subscript𝑝subscript𝑖𝑚1𝑡subscript𝑝subscript𝑖𝑚2𝑡subscript𝑝subscript𝑖𝑛𝑡𝑛𝑚p_{Mt}=\frac{(p_{i_{m+1}t}+p_{i_{m+2}t}+......+p_{i_{n}t})}{(n-m)}italic_p start_POSTSUBSCRIPT italic_M italic_t end_POSTSUBSCRIPT = divide start_ARG ( italic_p start_POSTSUBSCRIPT italic_i start_POSTSUBSCRIPT italic_m + 1 end_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT + italic_p start_POSTSUBSCRIPT italic_i start_POSTSUBSCRIPT italic_m + 2 end_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT + … … + italic_p start_POSTSUBSCRIPT italic_i start_POSTSUBSCRIPT italic_n end_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ) end_ARG start_ARG ( italic_n - italic_m ) end_ARG, where pFtsubscript𝑝𝐹𝑡p_{Ft}italic_p start_POSTSUBSCRIPT italic_F italic_t end_POSTSUBSCRIPT and pMtsubscript𝑝𝑀𝑡p_{Mt}italic_p start_POSTSUBSCRIPT italic_M italic_t end_POSTSUBSCRIPT refer to the average probability by which a topic dominantly belongs to the F and M groups respectively. If pFt>pMtsubscript𝑝𝐹𝑡subscript𝑝𝑀𝑡p_{Ft}>p_{Mt}italic_p start_POSTSUBSCRIPT italic_F italic_t end_POSTSUBSCRIPT > italic_p start_POSTSUBSCRIPT italic_M italic_t end_POSTSUBSCRIPT, we say the topic is a bias topic that aligns with the F group and vice-versa.

4.2 Finding Topic Pairs as Region-wise Bias Dimension Indicators

We use an embedding-based approach to identify F-M topic pairs from the pool of topics identified in the previous stage, to generate topic pairs (bias dimensions) that are comparable to IAT/WEAT pairs.

We use BERT-large (stsb-bert-large) from SpaCy’s Honnibal and Montani (2017) sentencebert library to extract contextual embeddings for topic words extracted from the GeoWAC dataset for each region.For a topic t𝑡titalic_t consisting of topic words w1,..wnw_{1},..w_{n}italic_w start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , . . italic_w start_POSTSUBSCRIPT italic_n end_POSTSUBSCRIPT, the topic embedding is given by the average of embeddings of the top ten topic words in that topic.

We identify topic pairs from the embeddings taking inspiration from Bolukbasi etal. (2016): let the embeddings of the words she𝑠𝑒sheitalic_s italic_h italic_e and he𝑒heitalic_h italic_e be Eshesubscript𝐸𝑠𝑒E_{she}italic_E start_POSTSUBSCRIPT italic_s italic_h italic_e end_POSTSUBSCRIPT and Ehesubscript𝐸𝑒E_{he}italic_E start_POSTSUBSCRIPT italic_h italic_e end_POSTSUBSCRIPT respectively. The embedding of a topic tisubscript𝑡𝑖t_{i}italic_t start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT be Etisubscript𝐸subscript𝑡𝑖E_{t_{i}}italic_E start_POSTSUBSCRIPT italic_t start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT. A female topic Ftisubscript𝐹subscript𝑡𝑖F_{t_{i}}italic_F start_POSTSUBSCRIPT italic_t start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT and a male topic Mtjsubscript𝑀subscript𝑡𝑗M_{t_{j}}italic_M start_POSTSUBSCRIPT italic_t start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT end_POSTSUBSCRIPT are a topic pair if: cos(EFti,Eshe)cos(EMtj,Ehe)similar-to𝑐𝑜𝑠subscript𝐸subscript𝐹subscript𝑡𝑖subscript𝐸𝑠𝑒𝑐𝑜𝑠subscript𝐸subscript𝑀subscript𝑡𝑗subscript𝐸𝑒cos(E_{F_{t_{i}}},E_{she})\sim\ cos(E_{M_{t_{j}}},E_{he})italic_c italic_o italic_s ( italic_E start_POSTSUBSCRIPT italic_F start_POSTSUBSCRIPT italic_t start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT end_POSTSUBSCRIPT , italic_E start_POSTSUBSCRIPT italic_s italic_h italic_e end_POSTSUBSCRIPT ) ∼ italic_c italic_o italic_s ( italic_E start_POSTSUBSCRIPT italic_M start_POSTSUBSCRIPT italic_t start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT end_POSTSUBSCRIPT end_POSTSUBSCRIPT , italic_E start_POSTSUBSCRIPT italic_h italic_e end_POSTSUBSCRIPT ) and/or cos(EFti,Ehe)cos(EMtj,Eshe)similar-to𝑐𝑜𝑠subscript𝐸subscript𝐹subscript𝑡𝑖subscript𝐸𝑒𝑐𝑜𝑠subscript𝐸subscript𝑀subscript𝑡𝑗subscript𝐸𝑠𝑒cos(E_{F_{t_{i}}},E_{he})\sim\ cos(E_{M_{t_{j}}},E_{she})italic_c italic_o italic_s ( italic_E start_POSTSUBSCRIPT italic_F start_POSTSUBSCRIPT italic_t start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT end_POSTSUBSCRIPT , italic_E start_POSTSUBSCRIPT italic_h italic_e end_POSTSUBSCRIPT ) ∼ italic_c italic_o italic_s ( italic_E start_POSTSUBSCRIPT italic_M start_POSTSUBSCRIPT italic_t start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT end_POSTSUBSCRIPT end_POSTSUBSCRIPT , italic_E start_POSTSUBSCRIPT italic_s italic_h italic_e end_POSTSUBSCRIPT ), where cos(i,j)𝑐𝑜𝑠𝑖𝑗cos(i,j)italic_c italic_o italic_s ( italic_i , italic_j ) refers to the cosine similarity between embeddings i𝑖iitalic_i and j𝑗jitalic_j, given by cos(i,j)=i,jij𝑐𝑜𝑠𝑖𝑗𝑖𝑗norm𝑖norm𝑗cos(i,j)=\frac{i,j}{||i||||j||}italic_c italic_o italic_s ( italic_i , italic_j ) = divide start_ARG italic_i , italic_j end_ARG start_ARG | | italic_i | | | | italic_j | | end_ARG. The threshold for the difference between the cosine similarities we consider for two topics to be a pair is 0.01, i.e., two topics (t1,t2)𝑡1𝑡2(t1,t2)( italic_t 1 , italic_t 2 ) are considered a pair if the difference of cosine similarities cos(t1𝑡1t1italic_t 1, she𝑠𝑒sheitalic_s italic_h italic_e)/cos(t1𝑡1t1italic_t 1, he𝑒heitalic_h italic_e) and cos(t2𝑡2t2italic_t 2, he𝑒heitalic_h italic_e)/cos(t2𝑡2t2italic_t 2, she𝑠𝑒sheitalic_s italic_h italic_e) respectively is <0.01absent0.01<0.01< 0.01. We manually choose 0.01 since differences close to 0.01 are almost =0absent0=0= 0.

Towards Region-aware Bias Evaluation Metrics (2)

4.3 Human Validation Setup

We design an IAT-style test to validate our topic pairs with annotators from different regions. We recruit six annotators from each region controlled by gender (three female and three male). In addition to our topics, we also test for existing WEAT dimensions relating to gender, namely family–career, math–arts, and science–arts. For each region, we validate all the region-aware topic pairs using the assistance of our annotators.

As done in IAT, to verify a topic pair, we show the topic names and male/female faces to our annotators along with a set of guidelines. As shown in Fig2, each topic pair test form contains two tasks.First, the annotators have to press one key for a female face f𝑓fitalic_f and a female topic Tfsubscript𝑇𝑓T_{f}italic_T start_POSTSUBSCRIPT italic_f end_POSTSUBSCRIPT and another key for a male face m𝑚mitalic_m with a male topic Tmsubscript𝑇𝑚T_{m}italic_T start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT, timing responses as r1subscript𝑟1r_{1}italic_r start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT and r2subscript𝑟2r_{2}italic_r start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT.In the reverse task, they pair Tmsubscript𝑇𝑚T_{m}italic_T start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT with f𝑓fitalic_f and Tfsubscript𝑇𝑓T_{f}italic_T start_POSTSUBSCRIPT italic_f end_POSTSUBSCRIPT with m𝑚mitalic_m, timing these as r3subscript𝑟3r_{3}italic_r start_POSTSUBSCRIPT 3 end_POSTSUBSCRIPT and r4subscript𝑟4r_{4}italic_r start_POSTSUBSCRIPT 4 end_POSTSUBSCRIPT. We average r1subscript𝑟1r_{1}italic_r start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT and r2subscript𝑟2r_{2}italic_r start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT for the ‘un-reversed’ case and r3subscript𝑟3r_{3}italic_r start_POSTSUBSCRIPT 3 end_POSTSUBSCRIPT and r4subscript𝑟4r_{4}italic_r start_POSTSUBSCRIPT 4 end_POSTSUBSCRIPT for the ‘reversed’ case. The annotators’ implicit association of a gender to a topic may influence their response time. A lower response time suggests easier recollection of the guidelines and potential implicit gender-topic associations, and thus lower bias with respect to these topics. We also varied the test order for different annotators to avoid initial pairing bias.We conduct the survey with six annotators each from Africa, Asia, Europe, and North America, also including a family-career topic pair, a standard WEAT bias dimension. We provide screenshots of our annotation framework in AppendixM.

4.4 Results: Bias Dimensions across Regions

4.4.1 Region-wise Bias Topics

Table2 displays the top topics based on umasssubscript𝑢𝑚𝑎𝑠𝑠u_{mass}italic_u start_POSTSUBSCRIPT italic_m italic_a italic_s italic_s end_POSTSUBSCRIPT Mimno etal. (2011) coherence for each region.

RegionFemaleMale
AfricaCredit cards and finances, Royalty and Media, Trading strategies and market analysis, Dating and relationships guides, Parenting and family relationshipsFashion and Lifestyle, Male enhancement and sexual health, Nollywood actresses and movies, Nigerian politics and government, Essay writing and research
AsiaHobbies and Interests, Healthy eating habits for children, Social media platforms, Royal wedding plans, Online Dating and ChattingDC comic characters, Mobile Application, Phillippine Politics and Government, Sports and Soccer, Career
EuropePets and animal care, Fashion and Style, Education, Obituaries and Genealogy, Luxury sailingPolitical developments in Northern Ireland, Christian Theology and Practice, Crime and murder investigation, EU Referendum and Ministerial Positions, Criminal Justice System
North AmericaPets, Cooking: culinary delights and chef recipes, Fashion and style, Family dynamics and relationships, Reading and fictionCivil War and history, Middle East conflict and political tensions, Movies and filmmaking, Political leadership and party dynamics in Bermuda, Rock Music and songwriting
OceaniaCooking and culinary delights, Romance, Weight loss and nutrition for women, Water travel experience, Woodworking plans and projectsHarry Potter adventures, Art and Photography, Superheroes and their Universes, Music recording and Artists, Football in Vanuatu

Several topics are exclusive to certain regions. Some topics like family and parenting; cooking; pets and animal care are common across some regions for F. Similarly we have movies; politics and government; sports; and music for M. Finally, there are differences between regions in terms of education, reading, and research (F-Europe, NA, and M-Africa), and fashion and lifestyle (F-Europe, NA, and M-Africa). Some other popular topics across regions are religion and spirituality, Christian theology in M; obituaries and genealogy, online dating, travel, and sailing in F (see AppendixD for a comprehensive list of topics). We provide an example of topic clusters in AppendixJ.

4.4.2 Region-wise Bias Dimensions

Table3 shows the top five topic pairs per region, chosen based on the umasssubscript𝑢𝑚𝑎𝑠𝑠u_{mass}italic_u start_POSTSUBSCRIPT italic_m italic_a italic_s italic_s end_POSTSUBSCRIPT score from the top 10 topics each for F and M from the topic modeling scheme. As expected, topic pairs differ by region, and we also note new topic pairs that do not appear in the WEAT tests. Among the top ones, there are recurring topics in F such as dating and marriage, family and relationships, luxury sailing, and education, whereas in M, we have politics, religion, sports, and movies.These region-specific pairs may supplement generic tests to detect regional biases.

RegionF-M topic pair
AfricaParenting and family relationships-Nollywood Actress and Movies (P1)
Marriage and relationships - Sports and Football (P2)
Womens’ lives and successes - Fashion and Lifestyle (P3)
Music - Social Media (P4)
Dating and relationships advice - Religious and Spiritual growth (P5)
AsiaHotel royalty - Political leadership in India (P1)
Healthy eating habits for children - Sports and Soccer (P2)
Royal wedding plans - Social Media platforms for video sharing (P3)
Royal wedding plans - Religious devotion and spirituality (P4)
Marriage - Bollywood actors and films (P5)
EuropeEducation - Music (P1)
Comfortable hotels - Political decision and impact on society (P2)
Luxury sailing - UK Government Taxation policies (P3)
Obituaries and Genealogy - Christian Theology and Practice (P4)
Fashion and style - Christian theology and practice (P5)
Online Dating for Singles - Religion and Spirituality (P1)
NorthFashion and Style - Reproductive Health (P2)
AmericaEducation and achievements - Reinsurance and capital markets (P3)
Family dynamics and relationships - Nike shoes and fashion (P4)
Reading and fiction - Cape Cod news (P5)
OceaniaFamily relationships - Religious beliefs and figures (P1)
Woodworking plans and projects - Music record and Artists (P2)
Weight loss and nutrition for women - Building and designing boats (P3)
Exercises for hormone development - Superheroes and their Universes (P4)
Kids’ furniture and decor - Building and designing boats (P5)
Towards Region-aware Bias Evaluation Metrics (3)
Towards Region-aware Bias Evaluation Metrics (4)
Towards Region-aware Bias Evaluation Metrics (5)
Towards Region-aware Bias Evaluation Metrics (6)

4.4.3 Unigram/Bigram Analysis

We find several topics that are common across regions. However, they may differ across cultures and may reveal varied perceptions of biases. Several topics also change associations to genders based on regions. For example, ‘fashion and lifestyle’ in Africa is associated with males, however, it is associated with females in Europe and North America. Several topics like ‘family and parenting’ are commonly associated with females across different regions while ‘politics’ is associated with males. To this end, we compute the top uni-grams and bi-grams for topic pairs that are common across regions in AppendixE.

4.4.4 Human Validation Results

Fig3 shows response times for top five topic pairs in each region for un-reversed and reversed scenarios. Larger time differences indicate more bias, suggesting that the pair could be a potential gender bias dimension for that region. If un-reversed time is lower, it suggests a stronger association of Tfsubscript𝑇𝑓T_{f}italic_T start_POSTSUBSCRIPT italic_f end_POSTSUBSCRIPT with the F group and Tmsubscript𝑇𝑚T_{m}italic_T start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT with the M group. The family-career pair was also surveyed as a standard WEAT bias dimension. Please refer to Table3 for topic pair numbers (P1…P5) of each topic pair.

RegionF-M topic pairRedditUN General
Debates
Parenting and family relationships-Nollywood Actress and Movies0.5000.979
Marriage and relationships - Sports and Football-0.0510.224
AfricaWomens’ lives and successes - Fashion and Lifestyle0.4800.493
Music - Social Media1.8941.721
Dating and relationships advice - Religious and Spiritual growth1.4751.061
Hotel royalty - Political leadership in India1.3651.768
Healthy eating habits for children - Sports and Soccer0.006-0.068
AsiaRoyal wedding plans - Social Media platforms for video sharing1.051.393
Royal wedding plans - Religious devotion and spirituality1.1831.335
Marriage - Bollywood actors and films1.5430.918
Education - Music1.2611.920
Comfortable hotels - Political decision and impact on society0.3240.485
EuropeLuxury sailing - UK Government Taxation policies1.2321.558
Obituaries and Genealogy - Christian Theology and Practice0.001-0.405
Fashion and style - Christian theology and practice1.7301.028
Online Dating for Singles - Religion and Spirituality1.7281.830
Fashion and Style - Reproductive Health1.7231.095
North AmericaEducation and achievements - Reinsurance and capital markets-0.148-0.364
Family dynamics and relationships - Nike shoes and fashion0.1090.691
Reading and fiction - Cape Cod news0.2510.506
Family relationships - Religious beliefs and figures0.3050.267
Woodworking plans and projects - Music record and Artists0.056-0.258
OceaniaWeight loss and nutrition for women - Building and designing boats0.3360.582
Exercises for hormone development - Superheroes and their Universes-0.05-0.07
Kids’ furniture and decor - Building and designing boats0.6120.524

As expected, the family–career pair shows the highest bias across all three general IAT topic pairs. There are smaller differences among the other math–arts and science–arts.We also note that some pairs, such as dating and relationships advice–religious and spiritual growth (P5) for Africa, hotel royalty–political leadership in India (P1) for Asia, obituaries and geneology–Christian theology (P4), education–music (P1), and fashion and style–Christian theology (P5) for Europe, and online dating–religion and spirituality (P1), fashion and style–reproductive health (P2) for North America have differences higher than those for family–career in the respective regions, indicating that the participants associated more biases on our uncovered bias dimensions than the existing one in WEAT.These findings support our hypothesis that gender bias dimensions vary across regions and also bring preliminary evidence that the region-aware bias dimensions we uncover are in line with the human perception of bias in those regions.We also find that all the regions have biases that conform to our topic pairs gender association except P3: education–reinsurance and capital markets in North America, where the associated bias is negative.These findings confirm that topic pairs indeed differ across regions and that these differences must be taken into consideration when identifying and evaluating biases.

5 WEAT-based Evaluation Using Region-aware Topic Pairs

To measure biases in different data domains and regions, we extract region-aware topics using the GeoWAC dataset which spans Common Crawl separated by regions, and create a WEAT-style evaluation setup using these topics.

Data. We consider two datasets: (i) Reddit data and (ii) UN General Debates Baturo etal. (2017). The Reddit data consists of data from subreddits corresponding to specific regions: r/asia, r/africa, r/europe, r/northamerica, and r/oceania. We use the official Reddit API to extract data, consisting of 500 top posts222The Official Reddit API has rate limits, therefore 500 top posts from each subreddit ensures an equal number of examples for each region. from each subreddit. The posts are pre-processed to remove URLs and signs, and each post contains at least 30 words. The UN General Debate Corpus (UNGDC) includes texts of General Debate statements from 1970 to 2016. These statements, similar to annual legislative state-of-the-union addresses, are delivered by leaders and senior officials to present their government’s perspective on global issues. We filter the countries for each region and extract 500 data points per region, maintaining equal representation across region.333Oceania has limited available countries in UNGDC, hence the adherence to 500 data points for each region.

Method. WEAT tests consist of keywords corresponding to each attribute and topic word sets like family-career and male-female terms. To create a similar setup, we utilize KeyBERT Grootendorst (2020) to gather top topic representations corresponding to each topic extracted from GeoWAC. For male/female terms, we use the same representative words from WEAT. To further make it specific to a particular region, we employ GPT-4 OpenAI etal. (2024) to generate common male/female names used in the regions and add them to the list. We provide the list of words in Table12 of Appendix F. We use fastText Bojanowski etal. (2017)444We choose fastText because it allows to extract embeddings of words that are not present in the target text (as our topics are derived from GeoWAC). embedding algorithm to generate embeddings of the lists and compute the distances between the topic words and male/female terms (like WEAT).

Towards Region-aware Bias Evaluation Metrics (7)
Towards Region-aware Bias Evaluation Metrics (8)

Results.In Table4, a high number of positive scores indicates the presence of a positive bias for our region-aware topic pairs. This means a presence of bias with the same gender association as our topic pairs, for example, if ‘music-social media’ is F-M topic pair in Africa according to our study, a positive score on the Reddit dataset means that bias is in the same direction. The few negative scores in the table indicate that these topic pairs do not conform to the same gender bias associations. However, a higher negative magnitude also shows the presence of bias, therefore, these topic pairs are still important.

Additionally, magnitudes of many scores are high (>0.5absent0.5>0.5> 0.5) which shows a high presence of bias (positive/negative) corresponding to the topic pairs. We highlight the top-scoring bias topic pairs for each region in Table4. High-bias topics vary for each region based on the dataset. For example, ‘music-social media’ has the highest bias in Africa for both datasets, however for Asia, we find that ‘marriage - Bollywood actors and films’ and ‘Hotel royalty - Political leadership in India’ are the topic pairs with the highest biases in Reddit and UN General Debates respectively, indicating that biased topic pairs may be domain-dependent.

Using our topic pairs in this WEAT-style evaluation setup provides an illustration of how our automatically curated region-aware bias dimensions can be used in designing a region-aware bias evaluation test.It also shows the effectiveness of our region-aware bias topic pairs in capturing the dimensions that are likely to contain gender biases across regions.555Note that our topic pairs although extracted from GeoWAC are somewhat generalizable to other datasets like Reddit and UNGDC, we do not claim that these are best topic pairs achievable as topic pairs are also data dependent, but we can use our methodology to extract bias topic pairs that may exist in specific datasets.

6 Alignment of Region-Aware Bias Dimensions with LLM outputs

To understand if LLMs generate similar biases as our region-aware bias topic pairs, we devise a persona generation task by LLMs. We prompt the LLM to output personas interested in different ‘topics’ from the topic pairs that we extract. Fig4 shows an example of the prompt given to an LLM to generate personas. We experiment with different LLMs: GPT-3.5 Brown etal. (2020), GPT-4, Mistral-7b-Instruct Jiang etal. (2023), Claude-3 Sonnet,666https://claude.ai/ and Gemini-Pro Team etal. (2024). Many studies use LLM-generated personas for multi-agent interactions in different settings in societies Park etal. (2023); Zhou etal. (2024). But, if an LLM generates biased personas, for example, a female persona takes care of children, and a male persona is strong and takes care of emergencies, this would lead to further biases in consequent tasks. Therefore, we employ persona generation to check for the presence of any biases in the personas created by LLMs. To measure biases, we find the number of matched LLM output persona genders to the genders of our topic pairs.We average our results over seven runs.

Results.We plot the results of persona gender mismatched by LLMs in Fig5. The y-axis shows % mismatch between the LLM generated persona gender and the gender of the topic in our topic pair. For example, a mismatch is when LLM outputs a persona with ‘female’ for Politics in Asia, which is a ‘male’ topic according to our findings. Regions with high representation: North America, Europe and Asia have fewer mismatches, with North America having the lowest mismatch. Conversely, less represented regions like Africa and Oceania show higher mismatch rates. Among models, Mistral-7b (7B) has the highest mismatch rate while Gemini-Pro (50T) has the least, which may stem from varying model sizes.Overall, all the models exhibit similar mismatch trends for both highly and less represented regions. Fewer mismatches in highly-represented regions show the importance of evaluation using region-specific topic pairs. Higher mismatches in underrepresented regions like Africa and Oceania suggest LLMs don’t mimic these areas’ biases, which can be beneficial. However, due to growing research on LLMs’ cultural alignment, a more precise, region-specific bias evaluation metric becomes essential.

7 Related Work

IAT is one of the earliest and well-known method for measuring implicit social biases in humans Greenwald etal. (1998). Inspired by the IAT, WEAT uses word embeddings to measure biases in text Caliskan etal. (2017). Another extension of WEAT is the Sentence Embedding Association Test (SEAT), which measures biases at the sentence level May etal. (2019). Additionally, various bias detection measures in NLP focus on post-training model predictions, such as gender swapping Stanovsky etal. (2019). Moreover, there are specific gender bias evaluation test sets in tasks like coreference resolution Rudinger etal. (2018); Zhao etal. (2018); Webster etal. (2018) and sentiment analysis Kiritchenko and Mohammad (2018b).

Several studies have emphasized the significance of considering cultural awareness in the study of social phenomena. The demographics of individuals can shape their worldviews and thoughts Garimella etal. (2016), potentially influencing their language preferences and biases in daily life. Notably, some studies have observed a bias towards Western nations in current LLMs Dwivedi etal. (2023). Recent research has focused on cross-cultural aspects of LLMs, including aligning them with human values from different cultures Glaese etal. (2022); Sun etal. (2023) and exploring them as personas representing diverse cultures Gupta etal. (2024). To the best of our knowledge, no previous work has proposed a data-dependent approach to extract region-aware bias topics. Given the known biases in LLMs, a region-specific metric could greatly lead to an accurate evaluation of biases. This research holds significant importance in addressing cross-cultural biases effectively.

8 Conclusion

In this paper, we proposed a bottom-up approach using data to identify region-aware topic pairs that capture gender biases across different regions. Our human evaluation results demonstrated the validity of our proposed region-aware dimensions.

We employed a region-aware WEAT-based evaluation setup to assess biases in two additional datasets: Reddit and the UN General Debate Corpus. The presence of region-specific biases in these datasets underscores the importance of a region-aware bias evaluation metric. Additionally, when examining LLM outputs against the gender associations in our region-aware bias topic pairs, we found that biases align closely for three highly represented regions: North America, Europe, and Asia. This emphasizes the value of region-aware topic pairs in bias evaluation of LLMs. Conversely, biases do not align well for Africa and Oceania, indicating that LLMs do not adopt these regions’ specific biases–a potential benefit. Yet, it also highlights the ’cultural alignment’ issue in LLMs. More research on the cultural alignment of LLMs underlines the need to consider region-specific bias topic pairs for all regions in future studies.

Future work includes incorporating testing different model/dataset combinations and topic-pair dependency on data. We aim to study biases in different languages and explore region-aware bias mitigation techniques.

9 Limitations

We utilized the GeoWAC corpus as our sole data source for extracting topic pairs from various regions. However, we acknowledge the importance of incorporating additional datasets in our future work. Additionally, our WEAT-based evaluation was conducted on relatively smaller datasets. So, we intend to conduct further analysis on a larger dataset to ensure a comprehensive evaluation based on WEAT.

Our study did not account for different languages due to the diverse linguistic landscape of the regions (continents) included in our study. However, the significance of conducting a more detailed analysis to examine variations among different countries would be interesting.

Unfortunately, we encountered difficulties in finding participants from Oceania for human validation. Moving forward, we plan to include insights and findings from Oceania and incorporate a larger population to ensure a more comprehensive human validation.

10 Ethical Considerations

When developing our region-aware topic pairs, it is essential to consider the ethical implications. Since we utilize a much broader aspect of culture, i.e. continents to distinguish among cultures, the region-aware topic pairs we extract may not translate to cultures of communities that are not well-represented in models. Hence, it is important that we utilize topic pairs carefully.

It has been found that AI models often tend to output responses that are Western, educated, industrialized, rich, and democratic Henrich etal. (2010). In our experiments, we see LLMs also generate biases having the highest alignment with the West. Therefore, LLM experiments also need to be utilized carefully.

Our Reddit data for the region-aware evaluation metric may contain offensive content. However, we have anonymized the data (removed the usernames).

References

  • Baturo etal. (2017)Alexander Baturo, Niheer Dasandi, and SlavaJ. Mikhaylov. 2017.Understanding state preferences with text as data: Introducing the un general debate corpus.Research & Politics, 4(2):2053168017712821.
  • Bishop (2006)Christopher Bishop. 2006.Pattern recognition and machine learning.Springer google schola, 2:5–43.
  • Blog (2017)National GeographicEducation Blog. 2017.What continent do you think they are from? drawing humans to reveal internalized bias.
  • Bojanowski etal. (2017)Piotr Bojanowski, Edouard Grave, Armand Joulin, and Tomas Mikolov. 2017.Enriching word vectors with subword information.Transactions of the Association for Computational Linguistics, 5:135–146.
  • Bolukbasi etal. (2016)Tolga Bolukbasi, Kai-Wei Chang, JamesY Zou, Venkatesh Saligrama, and AdamT Kalai. 2016.Man is to computer programmer as woman is to homemaker? debiasing word embeddings.Advances in neural information processing systems, 29.
  • Brown etal. (2020)TomB. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, DanielM. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. 2020.Language models are few-shot learners.CoRR, abs/2005.14165.
  • Caliskan etal. (2017)Aylin Caliskan, JoannaJ Bryson, and Arvind Narayanan. 2017.Semantics derived automatically from language corpora contain human-like biases.Science, 356(6334):183–186.
  • Damann etal. (2023)TaylorJ Damann, Jeremy Siow, and Margit Tavits. 2023.Persistence of gender biases in europe.Proceedings of the National Academy of Sciences, 120(12):e2213266120.
  • Devlin etal. (2019)Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019.BERT: Pre-training of deep bidirectional transformers for language understanding.In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, Minneapolis, Minnesota. Association for Computational Linguistics.
  • Dunn and Adams (2020a)Jonathan Dunn and Ben Adams. 2020a.Geographically-balanced gigaword corpora for 50 language varieties.In Proceedings of the Twelfth Language Resources and Evaluation Conference, pages 2528–2536.
  • Dunn and Adams (2020b)Jonathan Dunn and Ben Adams. 2020b.Mapping languages and demographics with georeferenced corpora.arXiv preprint arXiv:2004.00809.
  • Dwivedi etal. (2023)Ashutosh Dwivedi, Pradhyumna Lavania, and Ashutosh Modi. 2023.EtiCor: Corpus for analyzing LLMs for etiquettes.In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 6921–6931, Singapore. Association for Computational Linguistics.
  • Garimella etal. (2022)Aparna Garimella, Rada Mihalcea, and Akhash Amarnath. 2022.Demographic-aware language model fine-tuning as a bias mitigation technique.In Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pages 311–319, Online only. Association for Computational Linguistics.
  • Garimella etal. (2016)Aparna Garimella, Rada Mihalcea, and James Pennebaker. 2016.Identifying cross-cultural differences in word usage.In Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, pages 674–683, Osaka, Japan. The COLING 2016 Organizing Committee.
  • Glaese etal. (2022)Amelia Glaese, Nat McAleese, Maja Trębacz, John Aslanides, Vlad Firoiu, Timo Ewalds, Maribeth Rauh, Laura Weidinger, Martin Chadwick, Phoebe Thacker, Lucy Campbell-Gillingham, Jonathan Uesato, Po-Sen Huang, Ramona Comanescu, Fan Yang, Abigail See, Sumanth Dathathri, Rory Greig, Charlie Chen, Doug Fritz, JaumeSanchez Elias, Richard Green, Soňa Mokrá, Nicholas Fernando, Boxi Wu, Rachel Foley, Susannah Young, Iason Gabriel, William Isaac, John Mellor, Demis Hassabis, Koray Kavukcuoglu, LisaAnne Hendricks, and Geoffrey Irving. 2022.Improving alignment of dialogue agents via targeted human judgements.
  • Greenwald etal. (1998)AnthonyG Greenwald, DebbieE McGhee, and JordanLK Schwartz. 1998.Measuring individual differences in implicit cognition: the implicit association test.Journal of personality and social psychology, 74(6):1464.
  • Grimm and Church (1999)StephanieD Grimm and ATimothy Church. 1999.A cross-cultural study of response biases in personality measures.Journal of Research in Personality, 33(4):415–441.
  • Grootendorst (2020)Maarten Grootendorst. 2020.Keybert: Minimal keyword extraction with bert.
  • Grootendorst (2022)Maarten Grootendorst. 2022.Bertopic: Neural topic modeling with a class-based tf-idf procedure.
  • Gupta etal. (2024)Akshat Gupta, Xiaoyang Song, and Gopala Anumanchipalli. 2024.Self-assessment tests are unreliable measures of llm personality.
  • Henrich etal. (2010)Joseph Henrich, StevenJ Heine, and Ara Norenzayan. 2010.The weirdest people in the world?Behavioral and brain sciences, 33(2-3):61–83.
  • Honnibal and Montani (2017)Matthew Honnibal and Ines Montani. 2017.spaCy 2: Natural language understanding with Bloom embeddings, convolutional neural networks and incremental parsing.To appear.
  • Jiang etal. (2023)AlbertQ. Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, DevendraSingh Chaplot, Diego delas Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lucile Saulnier, LélioRenard Lavaud, Marie-Anne Lachaux, Pierre Stock, TevenLe Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix, and WilliamEl Sayed. 2023.Mistral 7b.
  • Jiang etal. (2019)Mengyin Jiang, ShirleyKM Wong, HarryKS Chung, Yang Sun, JanetH Hsiao, Jie Sui, and GlynW Humphreys. 2019.Cultural orientation of self-bias in perceptual matching.Frontiers in Psychology, 10:1469.
  • Kiritchenko and Mohammad (2018a)Svetlana Kiritchenko and Saif Mohammad. 2018a.Examining gender and race bias in two hundred sentiment analysis systems.In Proceedings of the Seventh Joint Conference on Lexical and Computational Semantics, pages 43–53, New Orleans, Louisiana. Association for Computational Linguistics.
  • Kiritchenko and Mohammad (2018b)Svetlana Kiritchenko and Saif Mohammad. 2018b.Examining gender and race bias in two hundred sentiment analysis systems.In Proceedings of the Seventh Joint Conference on Lexical and Computational Semantics, pages 43–53, New Orleans, Louisiana. Association for Computational Linguistics.
  • Malzer and Baum (2020)Claudia Malzer and Marcus Baum. 2020.A hybrid approach to hierarchical density-based cluster selection.In 2020 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI). IEEE.
  • Markus and Kitayama (2003)HazelRose Markus and Shinobu Kitayama. 2003.Models of agency: sociocultural diversity in the construction of action.
  • May etal. (2019)Chandler May, Alex Wang, Shikha Bordia, SamuelR. Bowman, and Rachel Rudinger. 2019.On measuring social biases in sentence encoders.In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 622–628, Minneapolis, Minnesota. Association for Computational Linguistics.
  • McInnes etal. (2018)Leland McInnes, John Healy, Nathaniel Saul, and Lukas Großberger. 2018.Umap: Uniform manifold approximation and projection.Journal of Open Source Software, 3(29):861.
  • Mikolov etal. (2013)Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013.Efficient estimation of word representations in vector space.
  • Mimno etal. (2011)David Mimno, Hanna Wallach, Edmund Talley, Miriam Leenders, and Andrew McCallum. 2011.Optimizing semantic coherence in topic models.In Proceedings of the 2011 Conference on Empirical Methods in Natural Language Processing, pages 262–272, Edinburgh, Scotland, UK. Association for Computational Linguistics.
  • Nichols (2011)Dugan Nichols. 2011.Men and must-have shoes.Retrieved from.
  • OpenAI etal. (2024)OpenAI, Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, FlorenciaLeoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, Red Avila, Igor Babuschkin, Suchir Balaji, Valerie Balcom, Paul Baltescu, Haiming Bao, Mohammad Bavarian, Jeff Belgum, Irwan Bello, Jake Berdine, Gabriel Bernadett-Shapiro, Christopher Berner, Lenny Bogdonoff, Oleg Boiko, Madelaine Boyd, Anna-Luisa Brakman, Greg Brockman, Tim Brooks, Miles Brundage, Kevin Button, Trevor Cai, Rosie Campbell, Andrew Cann, Brittany Carey, Chelsea Carlson, Rory Carmichael, Brooke Chan, Che Chang, Fotis Chantzis, Derek Chen, Sully Chen, Ruby Chen, Jason Chen, Mark Chen, Ben Chess, Chester Cho, Casey Chu, HyungWon Chung, Dave Cummings, Jeremiah Currier, Yunxing Dai, Cory Decareaux, Thomas Degry, Noah Deutsch, Damien Deville, Arka Dhar, David Dohan, Steve Dowling, Sheila Dunning, Adrien Ecoffet, Atty Eleti, Tyna Eloundou, David Farhi, Liam Fedus, Niko Felix, SimónPosada Fishman, Juston Forte, Isabella Fulford, LeoGao, Elie Georges, Christian Gibson, Vik Goel, Tarun Gogineni, Gabriel Goh, Rapha Gontijo-Lopes, Jonathan Gordon, Morgan Grafstein, Scott Gray, Ryan Greene, Joshua Gross, ShixiangShane Gu, Yufei Guo, Chris Hallacy, Jesse Han, Jeff Harris, Yuchen He, Mike Heaton, Johannes Heidecke, Chris Hesse, Alan Hickey, Wade Hickey, Peter Hoeschele, Brandon Houghton, Kenny Hsu, Shengli Hu, Xin Hu, Joost Huizinga, Shantanu Jain, Shawn Jain, Joanne Jang, Angela Jiang, Roger Jiang, Haozhun Jin, Denny Jin, Shino Jomoto, Billie Jonn, Heewoo Jun, Tomer Kaftan, Łukasz Kaiser, Ali Kamali, Ingmar Kanitscheider, NitishShirish Keskar, Tabarak Khan, Logan Kilpatrick, JongWook Kim, Christina Kim, Yongjik Kim, JanHendrik Kirchner, Jamie Kiros, Matt Knight, Daniel Kokotajlo, Łukasz Kondraciuk, Andrew Kondrich, Aris Konstantinidis, Kyle Kosic, Gretchen Krueger, Vishal Kuo, Michael Lampe, Ikai Lan, Teddy Lee, Jan Leike, Jade Leung, Daniel Levy, ChakMing Li, Rachel Lim, Molly Lin, Stephanie Lin, Mateusz Litwin, Theresa Lopez, RyanLowe, Patricia Lue, Anna Makanju, Kim Malfacini, Sam Manning, Todor Markov, Yaniv Markovski, Bianca Martin, Katie Mayer, Andrew Mayne, Bob McGrew, ScottMayer McKinney, Christine McLeavey, Paul McMillan, Jake McNeil, David Medina, Aalok Mehta, Jacob Menick, Luke Metz, Andrey Mishchenko, Pamela Mishkin, Vinnie Monaco, Evan Morikawa, Daniel Mossing, Tong Mu, Mira Murati, Oleg Murk, David Mély, Ashvin Nair, Reiichiro Nakano, Rajeev Nayak, Arvind Neelakantan, Richard Ngo, Hyeonwoo Noh, Long Ouyang, Cullen O’Keefe, Jakub Pachocki, Alex Paino, Joe Palermo, Ashley Pantuliano, Giambattista Parascandolo, Joel Parish, Emy Parparita, Alex Passos, Mikhail Pavlov, Andrew Peng, Adam Perelman, Filipe deAvila BelbutePeres, Michael Petrov, HenriquePonde deOliveiraPinto, Michael, Pokorny, Michelle Pokrass, VitchyrH. Pong, Tolly Powell, Alethea Power, Boris Power, Elizabeth Proehl, Raul Puri, Alec Radford, Jack Rae, Aditya Ramesh, Cameron Raymond, Francis Real, Kendra Rimbach, Carl Ross, Bob Rotsted, Henri Roussez,Nick Ryder, Mario Saltarelli, Ted Sanders, Shibani Santurkar, Girish Sastry, Heather Schmidt, David Schnurr, John Schulman, Daniel Selsam, Kyla Sheppard, Toki Sherbakov, Jessica Shieh, Sarah Shoker, Pranav Shyam, Szymon Sidor, Eric Sigler, Maddie Simens, Jordan Sitkin, Katarina Slama, Ian Sohl, Benjamin Sokolowsky, Yang Song, Natalie Staudacher, FelipePetroski Such, Natalie Summers, Ilya Sutskever, Jie Tang, Nikolas Tezak, MadeleineB. Thompson, Phil Tillet, Amin Tootoonchian, Elizabeth Tseng, Preston Tuggle, Nick Turley, Jerry Tworek, Juan FelipeCerón Uribe, Andrea Vallone, Arun Vijayvergiya, Chelsea Voss, Carroll Wainwright, JustinJay Wang, Alvin Wang, Ben Wang, Jonathan Ward, Jason Wei, CJWeinmann, Akila Welihinda, Peter Welinder, Jiayi Weng, Lilian Weng, Matt Wiethoff, Dave Willner, Clemens Winter, Samuel Wolrich, Hannah Wong, Lauren Workman, Sherwin Wu, Jeff Wu, Michael Wu, Kai Xiao, Tao Xu, Sarah Yoo, Kevin Yu, Qiming Yuan, Wojciech Zaremba, Rowan Zellers, Chong Zhang, Marvin Zhang, ShengjiaZhao, Tianhao Zheng, Juntang Zhuang, William Zhuk, and Barret Zoph. 2024.Gpt-4 technical report.
  • Park etal. (2023)JoonSung Park, JosephC. O’Brien, CarrieJ. Cai, MeredithRingel Morris, Percy Liang, and MichaelS. Bernstein. 2023.Generative agents: Interactive simulacra of human behavior.In In the 36th Annual ACM Symposium on User Interface Software and Technology (UIST ’23), UIST ’23, New York, NY, USA. Association for Computing Machinery.
  • Reimers and Gurevych (2019)Nils Reimers and Iryna Gurevych. 2019.Sentence-BERT: Sentence embeddings using Siamese BERT-networks.In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 3982–3992, Hong Kong, China. Association for Computational Linguistics.
  • Rudinger etal. (2018)Rachel Rudinger, Jason Naradowsky, Brian Leonard, and Benjamin VanDurme. 2018.Gender bias in coreference resolution.In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers), pages 8–14, New Orleans, Louisiana. Association for Computational Linguistics.
  • Russell (2010)RebeccaRoss Russell. 2010.Gender and jewelry: A feminist analysis.Rebecca Ross Russell.
  • Stanovsky etal. (2019)Gabriel Stanovsky, NoahA. Smith, and Luke Zettlemoyer. 2019.Evaluating gender bias in machine translation.In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 1679–1684, Florence, Italy. Association for Computational Linguistics.
  • Sun etal. (2023)Zhiqing Sun, Yikang Shen, Qinhong Zhou, Hongxin Zhang, Zhenfang Chen, David Cox, Yiming Yang, and Chuang Gan. 2023.Principle-driven self-alignment of language models from scratch with minimal human supervision.In Thirty-seventh Conference on Neural Information Processing Systems.
  • Team etal. (2024)Gemini Team, Machel Reid, Nikolay Savinov, Denis Teplyashin, Dmitry, Lepikhin, Timothy Lillicrap, Jean baptiste Alayrac, Radu Soricut, Angeliki Lazaridou, Orhan Firat, Julian Schrittwieser, Ioannis Antonoglou, Rohan Anil, Sebastian Borgeaud, Andrew Dai, Katie Millican, Ethan Dyer, Mia Glaese, Thibault Sottiaux, Benjamin Lee, Fabio Viola, Malcolm Reynolds, Yuanzhong Xu, James Molloy, Jilin Chen, Michael Isard, Paul Barham, Tom Hennigan, Ross McIlroy, Melvin Johnson, Johan Schalkwyk, Eli Collins, Eliza Rutherford, Erica Moreira, Kareem Ayoub, Megha Goel, Clemens Meyer, Gregory Thornton, Zhen Yang, Henryk Michalewski, Zaheer Abbas, Nathan Schucher, Ankesh Anand, Richard Ives, James Keeling, Karel Lenc, Salem Haykal, Siamak Shakeri, Pranav Shyam, Aakanksha Chowdhery, Roman Ring, Stephen Spencer, Eren Sezener, Luke Vilnis, Oscar Chang, Nobuyuki Morioka, George Tucker, CeZheng, Oliver Woodman, Nithya Attaluri, Tomas Kocisky, Evgenii Eltyshev, XiChen, Timothy Chung, Vittorio Selo, Siddhartha Brahma, PetkoGeorgiev, Ambrose Slone, Zhenkai Zhu, James Lottes, Siyuan Qiao, Ben Caine, Sebastian Riedel, Alex Tomala, Martin Chadwick, Juliette Love, Peter Choy, Sid Mittal, Neil Houlsby, Yunhao Tang, Matthew Lamm, Libin Bai, Qiao Zhang, Luheng He, Yong Cheng, Peter Humphreys, Yujia Li, Sergey Brin, Albin Cassirer, Yingjie Miao, Lukas Zilka, Taylor Tobin, Kelvin Xu, Lev Proleev, Daniel Sohn, Alberto Magni, LisaAnne Hendricks, Isabel Gao, Santiago Ontanon, Oskar Bunyan, Nathan Byrd, Abhanshu Sharma, Biao Zhang, Mario Pinto, Rishika Sinha, Harsh Mehta, Dawei Jia, Sergi Caelles, Albert Webson, Alex Morris, Becca Roelofs, Yifan Ding, Robin Strudel, Xuehan Xiong, Marvin Ritter, Mostafa Dehghani, Rahma Chaabouni, Abhijit Karmarkar, Guangda Lai, Fabian Mentzer, Bibo Xu, YaGuang Li, Yujing Zhang, TomLe Paine, Alex Goldin, Behnam Neyshabur, Kate Baumli, Anselm Levskaya, Michael Laskin, Wenhao Jia, JackW. Rae, Kefan Xiao, Antoine He, Skye Giordano, Lakshman Yagati, Jean-Baptiste Lespiau, Paul Natsev, Sanjay Ganapathy, FangyuLiu, Danilo Martins, Nanxin Chen, Yunhan Xu, Megan Barnes, Rhys May, Arpi Vezer, Junhyuk Oh, Ken Franko, Sophie Bridgers, Ruizhe Zhao, Boxi Wu, Basil Mustafa, Sean Sechrist, Emilio Parisotto, ThanumalayanSankaranarayana Pillai, Chris Larkin, Chenjie Gu, Christina Sorokin, Maxim Krikun, Alexey Guseynov, Jessica Landon, Romina Datta, Alexander Pritzel, Phoebe Thacker, Fan Yang, Kevin Hui, Anja Hauth, Chih-Kuan Yeh, David Barker, Justin Mao-Jones, Sophia Austin, Hannah Sheahan, Parker Schuh, James Svensson, Rohan Jain, Vinay Ramasesh, Anton Briukhov, Da-Woon Chung, Tamara von Glehn, Christina Butterfield, Priya Jhakra, Matthew Wiethoff, Justin Frye, Jordan Grimstad, Beer Changpinyo, CharlineLe Lan, Anna Bortsova, Yonghui Wu, Paul Voigtlaender, Tara Sainath, Shane Gu, Charlotte Smith, Will Hawkins, Kris Cao, James Besley, Srivatsan Srinivasan, Mark Omernick, Colin Gaffney, Gabriela Surita, Ryan Burnell, Bogdan Damoc, Junwhan Ahn, Andrew Brock, Mantas Pajarskas, Anastasia Petrushkina, Seb Noury, LorenzoBlanco, Kevin Swersky, Arun Ahuja, Thi Avrahami, Vedant Misra, Raoul deLiedekerke, Mariko Iinuma, Alex Polozov, Sarah York, George vanden Driessche, Paul Michel, Justin Chiu, Rory Blevins, Zach Gleicher, Adrià Recasens, Alban Rrustemi, Elena Gribovskaya, Aurko Roy, Wiktor Gworek, Sébastien M.R. Arnold, Lisa Lee, James Lee-Thorp, Marcello Maggioni, Enrique Piqueras, Kartikeya Badola, Sharad Vikram, Lucas Gonzalez, Anirudh Baddepudi, Evan Senter, Jacob Devlin, James Qin, Michael Azzam, Maja Trebacz, Martin Polacek, Kashyap Krishnakumar, Shuo yiin Chang, Matthew Tung, Ivo Penchev, Rishabh Joshi, Kate Olszewska, Carrie Muir, Mateo Wirth, AleJakse Hartman, Josh Newlan, Sheleem Kashem, Vijay Bolina, Elahe Dabir, Joost van Amersfoort, Zafarali Ahmed, James Cobon-Kerr, Aishwarya Kamath, ArnarMar Hrafnkelsson, LeHou, Ian Mackinnon, Alexandre Frechette, Eric Noland, Xiance Si, Emanuel Taropa, Dong Li, Phil Crone, Anmol Gulati, Sébastien Cevey, Jonas Adler, Ada Ma, David Silver, Simon Tokumine, RichardPowell, Stephan Lee, Kiran Vodrahalli, Samer Hassan, Diana Mincu, Antoine Yang, Nir Levine, Jenny Brennan, Mingqiu Wang, Sarah Hodkinson, Jeffrey Zhao, Josh Lipschultz, Aedan Pope, MichaelB. Chang, Cheng Li, LaurentEl Shafey, Michela Paganini, Sholto Douglas, Bernd Bohnet, Fabio Pardo, Seth Odoom, Mihaela Rosca, CiceroNogueira dos Santos, Kedar Soparkar, Arthur Guez, Tom Hudson, Steven Hansen, Chulayuth Asawaroengchai, Ravi Addanki, Tianhe Yu, Wojciech Stokowiec, Mina Khan, Justin Gilmer, Jaehoon Lee, CarrieGrimes Bostock, Keran Rong, Jonathan Caton, Pedram Pejman, Filip Pavetic, Geoff Brown, Vivek Sharma, Mario Lučić, Rajkumar Samuel, Josip Djolonga, Amol Mandhane, LarsLowe Sjösund, Elena Buchatskaya, Elspeth White, Natalie Clay, Jiepu Jiang, Hyeontaek Lim, Ross Hemsley, Zeyncep Cankara, Jane Labanowski, NicolaDe Cao, David Steiner, SayedHadi Hashemi, Jacob Austin, Anita Gergely, Tim Blyth, Joe Stanton, Kaushik Shivakumar, Aditya Siddhant, Anders Andreassen, Carlos Araya, Nikhil Sethi, RakeshShivanna, Steven Hand, Ankur Bapna, Ali Khodaei, Antoine Miech, Garrett Tanzer, Andy Swing, Shantanu Thakoor, Lora Aroyo, Zhufeng Pan, Zachary Nado, Jakub Sygnowski, Stephanie Winkler, Dian Yu, Mohammad Saleh, Loren Maggiore, Yamini Bansal, Xavier Garcia, Mehran Kazemi, Piyush Patil, Ish*ta Dasgupta, Iain Barr, Minh Giang, Thais Kagohara, Ivo Danihelka, Amit Marathe, Vladimir Feinberg, Mohamed Elhawaty, Nimesh Ghelani, Dan Horgan, Helen Miller, Lexi Walker, Richard Tanburn, Mukarram Tariq, Disha Shrivastava, Fei Xia, Qingze Wang, Chung-Cheng Chiu, Zoe Ashwood, Khuslen Baatarsukh, Sina Samangooei, RaphaëlLopez Kaufman, Fred Alcober, Axel Stjerngren, Paul Komarek, Katerina Tsihlas, Anudhyan Boral, Ramona Comanescu, Jeremy Chen, Ruibo Liu, Chris Welty, Dawn Bloxwich, Charlie Chen, Yanhua Sun, Fangxiaoyu Feng, Matthew Mauger, Xerxes Dotiwalla, Vincent Hellendoorn, Michael Sharman, Ivy Zheng, Krishna Haridasan, Gabe Barth-Maron, Craig Swanson, Dominika Rogozińska, Alek Andreev, PaulKishan Rubenstein, RuoxinSang, Dan Hurt, Gamaleldin Elsayed, Renshen Wang, Dave Lacey, Anastasija Ilić, Yao Zhao, Adam Iwanicki, Alejandro Lince, Alexander Chen, Christina Lyu, Carl Lebsack, Jordan Griffith, Meenu Gaba, Paramjit Sandhu, Phil Chen, Anna Koop, Ravi Rajwar, SoheilHassas Yeganeh, Solomon Chang, Rui Zhu, Soroush Radpour, Elnaz Davoodi, VingIan Lei, Yang Xu, Daniel Toyama, Constant Segal, Martin Wicke, Hanzhao Lin, Anna Bulanova, AdriàPuigdomènech Badia, Nemanja Rakićević, Pablo Sprechmann, Angelos Filos, Shaobo Hou, Víctor Campos, Nora Kassner, Devendra Sachan, Meire Fortunato, Chimezie Iwuanyanwu, Vitaly Nikolaev, Balaji Lakshminarayanan, Sadegh Jazayeri, Mani Varadarajan, Chetan Tekur, Doug Fritz, Misha Khalman, David Reitter, Kingshuk Dasgupta, Shourya Sarcar, Tina Ornduff, Javier Snaider, Fantine Huot, Johnson Jia, Rupert Kemp, Nejc Trdin, Anitha Vijayakumar, Lucy Kim, Christof Angermueller, LiLao, Tianqi Liu, Haibin Zhang, David Engel, Somer Greene, Anaïs White, Jessica Austin, Lilly Taylor, ShereenAshraf, Dangyi Liu, Maria Georgaki, Irene Cai, Yana Kulizhskaya, Sonam Goenka, Brennan Saeta, Ying Xu, Christian Frank, Dario deCesare, Brona Robenek, Harry Richardson, Mahmoud Alnahlawi, Christopher Yew, Priya Ponnapalli, Marco Tagliasacchi, Alex Korchemniy, Yelin Kim, Dinghua Li, Bill Rosgen, Kyle Levin, Jeremy Wiesner, Praseem Banzal, Praveen Srinivasan, Hongkun Yu, Çağlar Ünlü, David Reid, Zora Tung, Daniel Finchelstein, Ravin Kumar, Andre Elisseeff, Jin Huang, Ming Zhang, Ricardo Aguilar, Mai Giménez, Jiawei Xia, Olivier Dousse, Willi Gierke, Damion Yates, Komal Jalan, LuLi, Eri Latorre-Chimoto, DucDung Nguyen, Ken Durden, Praveen Kallakuri, Yaxin Liu, Matthew Johnson, Tomy Tsai, Alice Talbert, Jasmine Liu, Alexander Neitz, Chen Elkind, Marco Selvi, Mimi Jasarevic, LivioBaldini Soares, Albert Cui, Pidong Wang, AlekWenjiao Wang, Xinyu Ye, Krystal Kallarackal, Lucia Loher, Hoi Lam, Josef Broder, Dan Holtmann-Rice, Nina Martin, Bramandia Ramadhana, Mrinal Shukla, Sujoy Basu, Abhi Mohan, NickFernando, Noah Fiedel, Kim Paterson, Hui Li, Ankush Garg, Jane Park, DongHyun Choi, Diane Wu, Sankalp Singh, Zhishuai Zhang, Amir Globerson, Lily Yu, John Carpenter, Félix deChaumontQuitry, Carey Radebaugh, Chu-Cheng Lin, Alex Tudor, Prakash Shroff, Drew Garmon, Dayou Du, Neera Vats, Han Lu, Shariq Iqbal, Alex Yakubovich, Nilesh Tripuraneni, James Manyika, Haroon Qureshi, Nan Hua, Christel Ngani, MariaAbi Raad, Hannah Forbes, Jeff Stanway, Mukund Sundararajan, Victor Ungureanu, Colton Bishop, Yunjie Li, Balaji Venkatraman, BoLi, Chloe Thornton, Salvatore Scellato, Nishesh Gupta, Yicheng Wang, Ian Tenney, Xihui Wu, Ashish Shenoy, Gabriel Carvajal, DianaGage Wright, Ben Bariach, Zhuyun Xiao, Peter Hawkins, Sid Dalmia, Clement Farabet, Pedro Valenzuela, Quan Yuan, Ananth Agarwal, Mia Chen, Wooyeol Kim, Brice Hulse, Nandita Dukkipati, Adam Paszke, Andrew Bolt, Kiam Choo, Jennifer Beattie, Jennifer Prendki, Harsha Vashisht, Rebeca Santamaria-Fernandez, LuisC. Cobo, Jarek Wilkiewicz, David Madras, AliElqursh, Grant Uy, Kevin Ramirez, Matt Harvey, Tyler Liechty, Heiga Zen, Jeff Seibert, ClaraHuiyi Hu, Andrey Khorlin, Maigo Le, Asaf Aharoni, Megan Li, Lily Wang, Sandeep Kumar, Norman Casagrande, Jay Hoover, DaliaEl Badawy, David Soergel, Denis Vnukov, Matt Miecnikowski, Jiri Simsa, Praveen Kumar, Thibault Sellam, Daniel Vlasic, Samira Daruki, Nir Shabat, John Zhang, Guolong Su, Jiageng Zhang, Jeremiah Liu, YiSun, Evan Palmer, Alireza Ghaffarkhah, XiXiong, Victor Cotruta, Michael Fink, Lucas Dixon, Ashwin Sreevatsa, Adrian Goedeckemeyer, Alek Dimitriev, Mohsen Jafari, Remi Crocker, Nicholas FitzGerald, Aviral Kumar, Sanjay Ghemawat, Ivan Philips, Frederick Liu, Yannie Liang, Rachel Sterneck, Alena Repina, Marcus Wu, Laura Knight, Marin Georgiev, Hyo Lee, Harry Askham, Abhishek Chakladar, Annie Louis, Carl Crous, Hardie Cate, Dessie Petrova, Michael Quinn, Denese Owusu-Afriyie, Achintya Singhal, Nan Wei, Solomon Kim, Damien Vincent, Milad Nasr, ChristopherA. Choquette-Choo, Reiko Tojo, Shawn Lu, DiegodeLasCasas, Yuchung Cheng, Tolga Bolukbasi, Katherine Lee, Saaber Fatehi, Rajagopal Ananthanarayanan, Miteyan Patel, Charbel Kaed, Jing Li, ShreyasRammohan Belle, Zhe Chen, Jaclyn Konzelmann, Siim Põder, Roopal Garg, Vinod Koverkathu, Adam Brown, Chris Dyer, Rosanne Liu, Azade Nova, Jun Xu, Alanna Walton, Alicia Parrish, Mark Epstein, Sara McCarthy, Slav Petrov, Demis Hassabis, Koray Kavukcuoglu, Jeffrey Dean, and Oriol Vinyals. 2024.Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context.
  • Touvron etal. (2023)Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, Dan Bikel, Lukas Blecher, CristianCanton Ferrer, Moya Chen, Guillem Cucurull, David Esiobu, Jude Fernandes, Jeremy Fu, Wenyin Fu, Brian Fuller, Cynthia Gao, Vedanuj Goswami, Naman Goyal, Anthony Hartshorn, Saghar Hosseini, Rui Hou, Hakan Inan, Marcin Kardas, Viktor Kerkez, Madian Khabsa, Isabel Kloumann, Artem Korenev, PunitSingh Koura, Marie-Anne Lachaux, Thibaut Lavril, Jenya Lee, Diana Liskovich, Yinghai Lu, Yuning Mao, Xavier Martinet, Todor Mihaylov, Pushkar Mishra, Igor Molybog, Yixin Nie, Andrew Poulton, Jeremy Reizenstein, Rashi Rungta, Kalyan Saladi, Alan Schelten, Ruan Silva, EricMichael Smith, Ranjan Subramanian, XiaoqingEllen Tan, Binh Tang, Ross Taylor, Adina Williams, JianXiang Kuan, Puxin Xu, Zheng Yan, Iliyan Zarov, Yuchen Zhang, Angela Fan, Melanie Kambadur, Sharan Narang, Aurelien Rodriguez, Robert Stojnic, Sergey Edunov, and ThomasScialom. 2023.Llama 2: Open foundation and fine-tuned chat models.
  • Webster etal. (2018)Kellie Webster, Marta Recasens, Vera Axelrod, and Jason Baldridge. 2018.Mind the GAP: A balanced corpus of gendered ambiguous pronouns.Transactions of the Association for Computational Linguistics, 6:605–617.
  • Wolf etal. (2019)Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Rémi Louf, Morgan Funtowicz, and Jamie Brew. 2019.Huggingface’s transformers: State-of-the-art natural language processing.CoRR, abs/1910.03771.
  • Zhao etal. (2018)Jieyu Zhao, Tianlu Wang, Mark Yatskar, Vicente Ordonez, and Kai-Wei Chang. 2018.Gender bias in coreference resolution: Evaluation and debiasing methods.In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers), pages 15–20, New Orleans, Louisiana. Association for Computational Linguistics.
  • Zhou etal. (2024)Xuhui Zhou, Hao Zhu, Leena Mathur, Ruohong Zhang, Haofei Yu, Zhengyang Qi, Louis-Philippe Morency, Yonatan Bisk, Daniel Fried, Graham Neubig, and Maarten Sap. 2024.SOTOPIA: Interactive evaluation for social intelligence in language agents.In The Twelfth International Conference on Learning Representations.
RegionCountry#Examples
Nigeria3,153,761
AfricaMali660,916
Gabon645,769
India12,327,494
AsiaSingapore6,130,047
Philippines3,166,971
Ireland8,689,752
EuropeUnited Kingdom7,044,434
Spain465,780
Canada7.965,736
North AmericaUnited States8,521,094
Bermuda244,500
New Zealand94,476
OceaniaPalau486,437
Vanuatu165,355

Appendix A GeoWAC dataset details

Table5 contain the total number of examples per country in a region. We consider the top three countries with the highest number of examples per region.

Appendix B F-M Dataset statistics

Table6 displays the total number of examples from female and male groups per region for the region-specific F-M dataset.

RegionTotal#Female#Male
Africa578952015337742
Asia568772140035477
Europe591212104938072
North America706652762743038
Oceania621012595136150
Target words - Attribute wordsRegionRegion-specific p-valueRegion-specific WEAT scoreOriginal WEAT score, p-value
Africa0.0161.7981.81, 0.001
Asia0.0071.508
Male names vs Female namesNorth America0.041.885
- career vs familyEurope61046superscript1046\cdot 10^{-4}6 ⋅ 10 start_POSTSUPERSCRIPT - 4 end_POSTSUPERSCRIPT1.610
Oceania0.031.727
Africa0.0031.4291.06, 0.018
Asia0.0451.187
Math vs ArtsNorth America0.0070.703
- Male vs Female termsEurope0.0050.334
Oceania0.031.158
Africa0.0481.2471.24, 0.01
Asia0.0040.330
Science vs ArtsNorth America11051superscript1051\cdot 10^{-5}1 ⋅ 10 start_POSTSUPERSCRIPT - 5 end_POSTSUPERSCRIPT0.036
- Male vs Female termsEurope11071superscript1071\cdot 10^{-7}1 ⋅ 10 start_POSTSUPERSCRIPT - 7 end_POSTSUPERSCRIPT-0.655
Oceania21042superscript1042\cdot 10^{-4}2 ⋅ 10 start_POSTSUPERSCRIPT - 4 end_POSTSUPERSCRIPT0.725
Africa31053superscript1053\cdot 10^{-5}3 ⋅ 10 start_POSTSUPERSCRIPT - 5 end_POSTSUPERSCRIPT0.8551.21, 0.01
Young people namesAsia41044superscript1044\cdot 10^{-4}4 ⋅ 10 start_POSTSUPERSCRIPT - 4 end_POSTSUPERSCRIPT0.917
vs old people namesNorth America0.0321.325
- pleasant vs unpleasantEurope0.0090.917
Oceania0.0140.947
Africa11051superscript1051\cdot 10^{-5}1 ⋅ 10 start_POSTSUPERSCRIPT - 5 end_POSTSUPERSCRIPT0.0081.28, 0.001
European American namesAsia11061superscript1061\cdot 10^{-6}1 ⋅ 10 start_POSTSUPERSCRIPT - 6 end_POSTSUPERSCRIPT-0.453
vs African American namesNorth America0.0091.29
- pleasant vs unpleasantEurope0.0010.617
Oceania11041superscript1041\cdot 10^{-4}1 ⋅ 10 start_POSTSUPERSCRIPT - 4 end_POSTSUPERSCRIPT0.492
Africa0.031.4431.53, <107absentsuperscript107<10^{-7}< 10 start_POSTSUPERSCRIPT - 7 end_POSTSUPERSCRIPT
Asia0.0091.001
Instruments vs WeaponsNorth America0.011.202
- pleasant vs unpleasantEurope0.021.21
Oceania0.0010.951
Africa0.0020.3121.5, <107absentsuperscript107<10^{-7}< 10 start_POSTSUPERSCRIPT - 7 end_POSTSUPERSCRIPT
Asia0.0090.869
Flowers vs InsectsNorth America0.0030.382
- pleasant vs unpleasantEurope0.0010.332
Oceania0.0090.660
Africa0.0080.8351.38, 0.01
Asia0.021.201
Mental disease vs Physical diseaseNorth America0.0080.692
- temporary vs permanentEurope0.041.382
Oceania0.0091.620

Appendix C Cultural differences in biases using WEAT

Table7 shows the WEAT scores for all WEAT dimensions defined in Caliskan etal. (2017). We find that scores and p-values differ across regions for different dimensions. High bias dimensions differ across regions, hence it is important to consider region-specific topic pairs.

Appendix D Region-wise topic lists in GeoWAC

Table8 displays a comprehensive list of topics for female and male groups across all regions.

RegionFemaleMale
AfricaCredit cards and finances, Royalty and Media, Trading strategies and market analysis, Dating and relationships guides, Parenting and family relationships, Fashionable Ankara Styles, women’s lives and successes, online datingFashion and Lifestyle, Male enhancement and sexual health, Nollywood actresses and movies, Nigerian politics and government, Essay writing and research, Medical care for children and adults, Journalism and Media Conference, Music industry news and releases, Football league standing and player performances, Academic success and secondary school education, Religious inspiration and spiritual growth, Economic diversification and Socio-economic development
AsiaHobbies and Interests, Healthy eating habits for children, Social media platforms, Royal wedding plans, Online Dating and Chatting, Adult Services, Gift ideas for Valentine’s DayDC comic characters, Mobile Application, Philippine Politics and Government, Sports and Soccer, Career, Bike enthusiasts, Artists and their work, Youth Soccer Teams, Career in film industry, Political leadership in India, Bollywood actors and films, Religious devotion and spirituality, Phone accessories
EuropePets and animal care, Fashion and Style, Education, Obituaries and Genealogy, Luxury sailing, Traveling, Energy and climate change, Family and relationships, Pension and costs, Tech and business operations, Dating, Comfortable hotels, Government transportation policiesPolitical developments in Northern Ireland, Christian Theology and Practice, Crime and murder investigation, EU Referendum and Ministerial Positions, Criminal Justice System, Israeli politics and International relations, Cancer and medications, UK Government Taxation policies, Art Exhibitions, Political decision and impact on society, Music Gendres and artists, Medical specialties and university training, Political discourse and parliamentary debates
North AmericaPets, Cooking: culinary delights and chef recipes, Fashion and style, Family dynamics and relationships, Reading and fiction, Scheduling and dates, Life and legacy of Adolf Hitler, Gender roles and inequality, Education and achievements, Online dating for singles, Luxury handbags, Footwear and Apparel brands, Essay writing and literatureCivil War and history, Middle East conflict and political tensions, Movies and filmmaking, Political leadership and party dynamics in Bermuda, Rock Music and songwriting, Wartime aviation adventures, Religion and Spirituality, Reproductive health, Reinsurance and Capital markets, Nike shoes and fashion, Cape Cod news, NHL players
OceaniaCooking and culinary delights, Romance, Weight loss and nutrition for women, Water travel experience, Woodworking plans and projects, Time management and productivity, Inspiring stories and books for alleges, Sexual violence and abuse, Car insurance, Exercises for hormone development, kid’s furniture and decorHarry Potter adventures, Art and Photography, Superheroes and their Universes, Music recording and Artists, Football in Vanuatu, Pet care and veterinary services, Building and designing boats, Religious beliefs and figures, Fashion, Classic movie stars, Men’s hairstyle and fashion, Male sexual health and supplements
RegionFemale topicsMale topics
AfricaCredit card-based financial servicesFashion - footwear and celebrities
Royalty and femininityMale enhancement and sexual health
Financial tradingNollywood
Dating guidesNigerian politics
Motherhood and parentingAcademic writing
AsiaHobbiesSuperhero comic books
Food and nutritionMobile applications
Social media platforms and content creationPhilippines politics and people
Royal weddingsSports
Online social interaction and datingCareer
EuropePetsIrish politics
FashionChristianity
EducationLaw enforcement and crime
Deaths and funeralsEU and Brexit
Luxury yachting and sailingCriminal justice system
North AmericaPetsCivil War Military
Cooking and FoodMiddle Eastern politics and conflicts
FashionMovies and direction
Family and relationshipsBermuda politics
Reading novelsRock music
OceaniaFood and eating habitsHarry Potter
Romance and emotionsArtistic expressions
Weight loss and nutritionSuperheroes of Marvel and DC
Boat and sailing experienceAlbums, songs and artists
Woodworking and carpentryVanuatu Football
Towards Region-aware Bias Evaluation Metrics (9)

Appendix E Unigram/Bigram Analysis

TopicRegionUnigramsBigrams
Africa (male)march, outlet, air, max, tods, man, said, pas, cher, peopleair max, pas cher, princess j, roshe run, nike air, tods outlet, j march, roger vivier, posts email, notify new
Fashion and lifestyleEurope (female)one, women, fashion, like, new, look, make, hair, girl, dressoakley sunglasses, louis vuitton, red carpet, new york, fashion model, engagement rings, per cent, year old, christian louboutin, diamond ring
North America (female)one, love, like, little, new, made, time, get, make, womens cooper, cooper main, t shirt, new york, little girl, men women, look good, main store, years ago, check out
Africa (female)music, song, album, new, video, single, one, singer, also, songsmusic industry, hip hop, record label, single titled, new single, chris brown, tiwa savage, ice prince, kanye west, niegrian music
MusicEurope (male)man, single, stage, years, world, many, metal, guitar, solo, irishyear shelfmark, black metal, time exercise, musical content, dundee repertory, singer songwriter, edinburgh year, zumba days, male vocalists, millions men
Oceania (male)music, album, new, songs, band, first, time, jazz, released, rocknew york, elizabth ii, debut album, years later, big band, rock roll, first time, studio album, los angeles, solo artist

Table10 shows the unigrams and bigrams of common topics with different gender associations. We find that ‘fashion’ is highly associated with shoes when it is a male topic in Africa, whereas in Europe and North America, it is mostly associated with accessories like sunglasses, rings, etc. This shows the typical association of women with jewelry and men with shoes Russell (2010); Nichols (2011). In the case of ‘Music’, we see that unigrams and bigrams pertaining to Africa contain words related to hip-hop music and artists. For Europe, we find location references and metal music. And finally, Oceania shows references of jazz and rock. We do not find any obvious gender associations in the analysis of the music topic.Table11 provides a unigram/bigram analysis of topics that are commonly associated with a specific gender across regions. For parenting and family relationships, Africa has mentions of children, while Asia and Oceania contain mentions of family events, etc. In North America, we mostly find text about maintaining health in families.For religion and spirituality, the unigrams/bigrams are mostly about Jesus and Christianity across regions. For politics, we find mentions of specific regions, as expected. Education topic is more about being successful in Europe, where it is about degrees in North America. Finally, ‘social media’ trends are mostly similar. Overall for topics with same gender associations across regions, do not have stark differences.

TopicRegionUnigramsBigrams
Africa (female)child, registration, form, information, sent, women, foster, best, catholic, womenregistration form, form information, child assigned, surgery doctors, new catholic, catholic women, contemporary challenge, best everything, foster short, doctors clinic
Parenting and family relationshipsAsia (female)year, old, weekly, fortnightly, clicking, create, alert, state, 1, termsyear old, weekly fortnightly, create alert, stated agree, conditions acknowledge, finals appearances, together playing, dial guarded, came work, outlet jackets
North America (female)women, healthday, loss, three, worked, closely, together, she, elegant, dignifiedthree women, women worked, closely together, elegant dignified, very pleasant, soft spoken, women men, healthday reporter, tuesday march, participate more
Oceania (female)laurel, school, moved, one, day, royal, wedding, house, sister, hopefullymoved one, royal wedding, laurel school, 1 california, weeks dad, high school, one hopefully, nobody knew, sister means, fu school
Africa (male)god, man, church, one, life, people, jesus, us, lord, christ,short description, jesus christ, man god, holy spirit, god said, thank god, bible says, catholic church, today god, every man
Asia (male)life, jesus, us, church, one, man ,lord, said, father, christholu spirit, jesus christ, pope francis, brothers sisters, son god, men women, holy father, opus dei, eternal life, paul ii
Religion and SpiritualityEurope (male)god, one, jesus, church, life, people, father, man , said, christjesus christ, son man, catholic church, holy spirit, men women, said him, holy father, john paul, jesus said, word god
North America (male)god, jesus, one, man, us, life, would, christ, lord, peoplerecognizable cheering, section league, jesus christ, exact synonyms, past years, god said, years before, thanks mostly, mostly steph, father dell
Oceania (male)also, said, best, love, new, come, good, like, men, madegod said, jesus christ, holy spirit, lord krishna, temple god, father devil, eternal life, son god, son man, god father
Asia (male)said, one, India, time, people, minister, government, years, state, police, courtindian congress, government plans, modi ministry, human rights, foreign politics, armed forces, international warfare, foreign ministry, middle east, united nations
PoliticsEurope (male)government, said, minister, people, international, country, one, foreign, president, statemake statement, prime minister, human rights, armed forces, secretary state, middle east, united nations, hon friend, foreign secretary, united states
Europe (female)school, primary, teacher, founder, CEO, judgment, group, named, ranking, prestigiousas founder, founder CEO, judgment group, named fortune, ranking prestigious, world scientist, scientist women, students comprehend, program support, support students
EducationNorth America (female)bachelor, years, student, leader, degree, animal, veterinary, music, taught, communicationbachelors degree, animal veterinary, bachelor music, alison taught, privately years, students ranging, development programmes, including leader, art communication, recent years
Africa (male)onigbinder, aura, pictures, first, gained, popularity, match, beaut, designed, musicaura pictures, gained popularity, match beaut, designed wonder, attending music, music festival, schomburg library, Instagram account, sugar coating, schedule tomorrow
Social MediaAsia (male)time, later, latest, tracks, speedy, Zulfiqar, nasty, children, tweeted, guysgets later, latest tracks, speedy zulfiqar, children pti, pti tweeted, taking long, long time, hosted pageant, time vincent, love fleeting

Appendix F WEAT-based evaluation setup details

For male/female terms, we use the same representative words from WEAT: brother, father, uncle, grandfather, son, he, his, him, man, boy, male for male and sister, mother, aunt, grandmother, daughter, she, hers, her, woman, girl, female for female.We also utilize GPT-4 to output the ten most common male/female names specific to each region. We provide the lists of word belonging toeach topic in Table12.

RegionTopics: Word lists
Nollywood Actress and Movies: nollywood, actress, actors, drama, celebrity, movie, acting, movies, producer, tv
Parenting and family relationships: mother, mom, mothers, mum, moms, parent, her, child, momodu, parents
Sports and Football: players, sports, fifa, team, player, football, mourinho, scored, league, champions
Marriage and relationships: wives, marriage, husbands, marriages, married, wife, relationships, husband, marry, relationship
Fashion and lifestyle: cher, nike, max, air, looked, face, love, tods, soldes, scarpe
AFRICAWomens lives and successes: women, ladies, woman, female, girls, men, gender, ones, employees, male
Social Media: instagram, facebook, social, twitter, tweet, snapchat, tweets, tweeted, hashtag, followers
Music: song, songs, album, hits, music, released, rap, singer, tracks, rapper
Religious and Spiritual Growth: god, almighty, bible, christ, faith, believers, christian, jesus, prayer, religion
Dating and relationships advice: dating, women, relationships, ladies, sites, singles, online, single, escorts, websites
Male terms: male, man, boy, brother, he, him, his, son, Kwame, Mandela, Moyo, Jelani, Tariq, Keita, Obi, Simba, Ayo, Kofi, Jabari, Tunde, Mekonnen, Anwar, Chukwuemeka
Female terms: sister, mother, aunt, grandmother, daughter, she, hers, her, Aisha, Zahara, Nia, Sade, Amara, Chinelo, Layla, Ayana, Nala, Zuri, Imani, Lola, Kamaria, Nyala, Kaya
Political Leardership in India: modi, political, said, bjp, told, says, leader, congress, minister, public
Hotel Royalty: visited, places, stayed, hotels, adventure, pictures, favourite, guest, hiking, hemingway
Sports and Soccer: sports, team, basketball, players, nba, league, championship, coach, rebounds, finals
Healthy eating habits for children: food, foods, eating, meals, nutrition, cuisine, diet, dishes, cooking, eat
Social Media platforms for video sharing: instagram, video, videos, twitter, tweet, facebook, gifs, vlog, youtube, followers
ASIARoyal wedding plans: meghan, duch*ess, engagement, england, royal, royalty, prince, kate, london, married
Religious devotion and spirituality: god, bible, holy, faith, prayer, believe, christian, blessed, christ, spiritual
Royal wedding plans: meghan, duch*ess, engagement, england, royal, royalty, prince, kate, london, married
Bollywood actors and films: bollywood, bachchan, kapoor, actors, acting, kareena, actor, film, shahrukh, hindi
Marriage: married, marriage, marriages, couple, couples, wife, marry, wedding, husband, divorced
Male terms: male, man, boy, brother, he, him, his, son, Hiroshi, Ravi, Kazuki, Jin, Satoshi, Rohan, Haruki, Dai, Akira, Yuan
Female terms: sister, mother, aunt, grandmother, daughter, she, hers, her, Sakura, Mei, Aiko, Yuna, Lina, Ji-hye, Mika, Nami, Anika, Rina
Music: music, songs, vocalists, album, albums, singing, vocals, singles, rock, song
Education: school, schools, classroom, students, education, educational, pupils, boys, academy, college
Political decisions and impact on society: government, public, minister, said, hon, people, first, the, column, committee
Comfortable hotels: guests, staying, rooms, friendly, welcoming, stayed, hotel, beds, stay, comfortable
UK Government Taxation Policies: corbyn, taxation, fiscal, tax, taxes, exchequer, labour, governments, government, deficit
EUROPELuxury Sailing: yachts, yacht, boat, sailing, sails, cruising, sail, berths, cruiser, cabin
Christian Theology and Practice: god, bible, christ, jesus, faith, christian, religious, religion, holy, gave
Obituaries and Genealogy: died, edward, relatives, anne, lived, elizabeth, funeral, irish, mrs, galway
Christian Theology and Practice: god, bible, christ, jesus, faith, christian, religious, religion, holy, gave
Fashion and style: fashion, shoes, style, clothes, clothing, shoe, wear, nike, dress, stylish
Male terms: male, man, boy, brother, he, him, his, son, Lukas, Matteo, Sebastian, Alexander, Gabriel, Nikolai, Maximilian, Leonardo, Daniel, Adrian
Female terms: sister, mother, aunt, grandmother, daughter, she, hers, her, Emma, Sophia, Olivia, Isabella, Ava, Mia, Charlotte, Amelia, Lily, Emily
Religion and Spirituality: god, christ, jesus, bible, christian, holy, christians, scripture, faith, heaven
Online Dating for Singles: dating, singles, hookup, single, relationships, dates, flirting, personals, date, mingle
Reproductive Health: download, available, pdf, online, edition, manual, free, reprint, kindle, file
Fashion and style: fashion, dresses, dress, wardrobe, clothes, clothing, style, outfit, vintage, wear
Reinsurance and capital markets: reinsurance, reinsurers, insurers, insurance, securities, investors, investment, finance, trading, pension
NORTHEducation and achievements: school, schools, graduated, college, students, undergraduate, graduation, graduate, attended, education
AMERICANike shoes and fashion: nike, shoes, sneakers, jordans, jeans, tops, black, boys, men, casual
Family dynamics and relationships: family, families, children, kids, grandchildren, relatives, grandparents, parents, child, parent
Cape Cod news: lifeguard, drowned, drowns, newstweet, hospitalized, snorkeling, cape, reported, reuterstweet, pulled
Reading and fiction: books, book, reading, novels, series, enjoyed, novel, romance, katniss, readers
Male terms: male, man, boy, brother, he, him, his, son, Liam, Noah, Ethan, Jacob, William, Michael, James, Alexander, Benjamin, Matthew
Female terms: sister, mother, aunt, grandmother, daughter, she, hers, her, Emma, Olivia, Ava, Sophia, Isabella, Mia, Charlotte, Amelia, Harper, Evelyn
Religious beliefs and figures: god, gods, bible, mankind, faith, christ, spiritual, christian, religion, jesus
Family relationships: mum, mother, mom, mums, parent, family, parents, baby, dad, father
Music record and Artists: music, album, albums, jazz, songs, hits, musicians, artists, recordings, blues
Woordworking plans and projects: plans, furniture, woodwork, wood, woodcraft, woodworking, plywood, carpentry, cabinets, wooden
Building and designing boats: boatbuilder, boatbuilding, boats, plans, boat, sauceboat, sailboat, build, catamaran, kits
Weight loss and nutrition for women: diet, workout, exercise, foods, weight, food, eating, healthy, pounds, fat
OCEANIASuperheroes and their Universes: superhero, superheroes, avengers, marvel, comics, superman, aquaman, heroes, comic, hero
Exercises for hormone development: hormones, weightlifting, workouts, deadlifts, hormonal, exercises, lifting, testosterone, fitness, squats
Building and designing boats: boatbuilder, boatbuilding, boats, plans, boat, sauceboat, sailboat, build, catamaran, kits
Kids furniture and decor: furniture, chairs, sofas, ikea, sofa, cushions, sectional, upholstered, couch, childrens
Male terms: male, man, boy, brother, he, him, his, son, Manaia, Tane, Kai, Ariki, Mika, Koa, Rangi, Kane, Tama, Hemi
Female terms: sister, mother, aunt, grandmother, daughter, she, hers, her, Aroha, Moana, Tui, Lani, Kahurangi, Ariana, Malie, Marama, Ava, Kaia

Appendix G Paired-list for F-M datasets

Here is the list of the 52 pairs used to create the F-M datasets per region:
[monastery, convent], [spokesman, spokeswoman], [Catholic priest, nun], [Dad, Mom], [Men, Women], [councilman, councilwoman], [grandpa, grandma], [grandsons, granddaughters], [prostate cancer, ovarian cancer], [testosterone, estrogen], [uncle, aunt], [wives, husbands], [Father, Mother], [Grandpa, Grandma], [He, She], [boy, girl], [boys, girls], [brother, sister], [brothers, sisters], [businessman, businesswoman], [chairman, chairwoman], [colt, filly], [congressman, congresswoman], [dad, mom], [dads, moms], [dudes, gals], [ex girlfriend, ex boyfriend], [father, mother], [fatherhood, motherhood], [fathers, mothers], [fella, granny], [fraternity, sorority], [gelding, mare], [gentleman, lady], [gentlemen, ladies], [grandfather, grandmother], [grandson, granddaughter], [he, she], [himself, herself], [his, her], [king, queen], [kings, queens], [male, female], [males, females], [man, woman], [men, women], [nephew, niece], [prince, princess], [schoolboy, schoolgirl], [son, daughter], [sons, daughters], [twin brother, twin sister].
Each pair in the above is denoted as a [male, female] pair.

Appendix H Llama 2 prompt for topic modeling

The prompt scheme for Llama2 consists of three prompts: (1) System Prompt: a general prompt that describes information given to all conversations, (2) Example Prompt: an example that demonstrates the output we are looking for, and (3) Main Prompt: describes the structure of the main question, that is with a given set of documents and keywords, we ask the model to create a short label for the topic. Fig6 displays the three prompts as used in the code.

Appendix I Topic Cluster Labels using other LLMs

We use Llama2 to fine-tune our topics to label them for better coherence in our paper. However, we also experiment with GPT-4 and arrive at similar topics in Table9. (see Table2 for comparison with Llama2 topic labels).

Appendix J Topic Word Clusters Example - Africa

Here, we provide an example of how topics look in our data. In Fig7, we provide word clusters of topics from Africa. The word clusters contain the top 10 words from each topic in Africa. We find that topic labels by Llama2 are coherent in terms of top topic words.

Towards Region-aware Bias Evaluation Metrics (10)

Appendix K Region specific BERTs to identify top words in F/M direction

To motivate our case to investigate differences in biases across regions, we use BERT to compute the top words corresponding to the she-he axis in the embedding space. BERT is a pre-trained transformer-based language model that consists of a set of encoders. As a motivation experiment to identify differences in the contextual embedding space for different regions, we fine-tune BERT with the masked language modeling task (no labels) for each region separately. For a given word, we compute its embeddings by averaging out all sentence embeddings where it occurs across the dataset.Similarly, we compute embeddings for all words in the dataset. The tokenized input goes through the BERT model and we take the hidden states at the end of the last encoder layer (in our case, BERT-base, i.e. 12 encoder layers) as sentence embeddings. We identify the top words with the highest projection across the she-he axis in the region-specific datasets. If we find differences in the top words across regions, it is possible that dominating bias topics vary by region as well. Fig8 shows the top words closest to ‘she’ and ‘he’ contextual embeddings in our data for each region. We find that top words differ quite a bit across different regions.We find many differences in the top F (close to she𝑠𝑒sheitalic_s italic_h italic_e) and M (close to he𝑒heitalic_h italic_e) words across regions. Some top F words are soprano, archaeological (Africa); graduate, secretary (Asia); innovative, graphics (Europe); poets, sentiments (NA); and arts, sleep (Oceania). Some top M words are history, leading (Africa); astronomer, commissioners (Asia); honorary, songwriters (Europe); owner, hospital (NA); and wrestlemania, orbits (Oceania). Gender-neutral words such as poets, secretaries, astronomers, commissioners, songwriters, owners, and so on are closer to either the she or he axes. Although comparable to the findings of Bolukbasi etal. (2016), the variances among regions inspire us to look deeper into the data to arrive at culture-specific bias themes.

Appendix L Implementations details

For training our Bertopic model, we use Google Colab’s Tesla T4 GPU, and it takes 15 min to run topic modeling for a region-specific F-M dataset. Region-specific BERTs are run on NVIDIA RTX2080 GPUs. Each BERT training experiment takes 1 GPU hour. For our LLM experiment, we used NVIDIA-A40 for Mistral-7b-Instruct and Llama-3-8b for an hour. We do not use any GPUs for GPT-4, Claude-3-Sonnet and Gemini-Pro.

L.1 Bertopic

We use Bertopic’s default models: SBERT Reimers and Gurevych (2019) to contextually embed the dataset, UMAP McInnes etal. (2018) to perform dimensionality reduction, HDBSCAN Malzer and Baum (2020) for clustering to perform topic modeling. We choose the embedding model BAAI/bge-small-en from Huggingface Wolf etal. (2019). We set top_n_words to 10 and verbose as True and set the min_topic_size to 100 for the Bertopic model. Finally, we use Bertopic’s official library to implement the model.

L.2 Llama2

We use Llama2 to finetune the topics to give shorter labels for each topic. We set the temperature to 0.10.10.10.1, max_new_tokens to 500500500500 and repetition_penalty to 1.11.11.11.1. We utilize Bertopic’s built-in representation models to use Llama2 in our topic model.

L.3 LLM experiment

For GPT-4, and Mistral-7b-Instruct and Llama-3-8b, we utilize the Microsoft Azure API777https://learn.microsoft.com/en-us/rest/api/azure/, huggingface888https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1, and huggingface999https://huggingface.co/meta-llama/Meta-Llama-3-8B for inference respectively. We use a temperature 0.80.80.80.8 for all models. For Gemini-Pro and Claude-3-Sonnet, we use the available chat interface.

L.4 Region-specific BERT

We use the uncased version BERT Devlin etal. (2019) for our region-specific BERT model trained for the MLM objective. We use a batch size of 8, a learning rate of 11041superscript1041\cdot 10^{-4}1 ⋅ 10 start_POSTSUPERSCRIPT - 4 end_POSTSUPERSCRIPT, and an AdamW optimizer to train our BERT models for 3 epochs.

Towards Region-aware Bias Evaluation Metrics (11)
Towards Region-aware Bias Evaluation Metrics (12)
Towards Region-aware Bias Evaluation Metrics (13)
Towards Region-aware Bias Evaluation Metrics (14)
Towards Region-aware Bias Evaluation Metrics (15)

Appendix M Human Validation

Students and staff from a college campus were recruited as annotators in the study. Screenshots of the form are displayed in Fig9. We have 6 annotators per region (3 male and 3 female).

Towards Region-aware Bias Evaluation Metrics (16)
Towards Region-aware Bias Evaluation Metrics (17)
Towards Region-aware Bias Evaluation Metrics (18)
Towards Region-aware Bias Evaluation Metrics (19)
Towards Region-aware Bias Evaluation Metrics (20)

Appendix N Reproducibility

We open-source our codes, which are uploaded to the submission system. We include commands with hyperparameters in our codes. This would help future work to reproduce our results.

Towards Region-aware Bias Evaluation Metrics (2024)
Top Articles
Latest Posts
Article information

Author: Reed Wilderman

Last Updated:

Views: 6740

Rating: 4.1 / 5 (52 voted)

Reviews: 91% of readers found this page helpful

Author information

Name: Reed Wilderman

Birthday: 1992-06-14

Address: 998 Estell Village, Lake Oscarberg, SD 48713-6877

Phone: +21813267449721

Job: Technology Engineer

Hobby: Swimming, Do it yourself, Beekeeping, Lapidary, Cosplaying, Hiking, Graffiti

Introduction: My name is Reed Wilderman, I am a faithful, bright, lucky, adventurous, lively, rich, vast person who loves writing and wants to share my knowledge and understanding with you.