Spaces:
Runtime error
Runtime error
Avijit Ghosh
commited on
Commit
•
5bb95f3
1
Parent(s):
e6e82b8
change type to level
Browse files- app.py +9 -9
- configs/crowspairs.yaml +1 -1
- configs/homoglyphbias.yaml +1 -1
- configs/honest.yaml +1 -1
- configs/ieat.yaml +1 -1
- configs/imagedataleak.yaml +1 -1
- configs/notmyvoice.yaml +1 -1
- configs/palms.yaml +1 -1
- configs/stablebias.yaml +1 -1
- configs/stereoset.yaml +1 -1
- configs/tango.yaml +1 -1
- configs/videodiversemisinfo.yaml +1 -1
- configs/weat.yaml +1 -1
app.py
CHANGED
@@ -29,19 +29,19 @@ globaldf['Link'] = '<u>'+globaldf['Link']+'</u>'
|
|
29 |
modality_order = ["Text", "Image", "Audio", "Video"]
|
30 |
type_order = ["Model", "Dataset", "Output", "Taxonomy"]
|
31 |
|
32 |
-
# Convert Modality and
|
33 |
globaldf['Modality'] = pd.Categorical(globaldf['Modality'], categories=modality_order, ordered=True)
|
34 |
-
globaldf['
|
35 |
|
36 |
-
# Sort DataFrame by Modality and
|
37 |
-
globaldf.sort_values(by=['Modality', '
|
38 |
|
39 |
# create a gradio page with tabs and accordions
|
40 |
|
41 |
# Path: taxonomy.py
|
42 |
|
43 |
def filter_modality_type(fulltable, modality_filter, type_filter):
|
44 |
-
filteredtable = fulltable[fulltable['Modality'].isin(modality_filter) & fulltable['
|
45 |
return filteredtable
|
46 |
|
47 |
def showmodal(evt: gr.SelectData):
|
@@ -100,7 +100,7 @@ The following categories are high-level, non-exhaustive, and present a synthesis
|
|
100 |
with gr.Tabs(elem_classes="tab-buttons") as tabs1:
|
101 |
with gr.TabItem("Bias/Stereotypes"):
|
102 |
fulltable = globaldf[globaldf['Group'] == 'BiasEvals']
|
103 |
-
fulltable = fulltable[['Modality','
|
104 |
|
105 |
gr.Markdown("""
|
106 |
Generative AI systems can perpetuate harmful biases from various sources, including systemic, human, and statistical biases. These biases, also known as "fairness" considerations, can manifest in the final system due to choices made throughout the development process. They include harmful associations and stereotypes related to protected classes, such as race, gender, and sexuality. Evaluating biases involves assessing correlations, co-occurrences, sentiment, and toxicity across different modalities, both within the model itself and in the outputs of downstream tasks.
|
@@ -114,7 +114,7 @@ The following categories are high-level, non-exhaustive, and present a synthesis
|
|
114 |
)
|
115 |
type_filter = gr.CheckboxGroup(["Model", "Dataset", "Output", "Taxonomy"],
|
116 |
value=["Model", "Dataset", "Output", "Taxonomy"],
|
117 |
-
label="
|
118 |
show_label=True,
|
119 |
# info="Which modality to show."
|
120 |
)
|
@@ -138,7 +138,7 @@ The following categories are high-level, non-exhaustive, and present a synthesis
|
|
138 |
|
139 |
with gr.TabItem("Cultural Values/Sensitive Content"):
|
140 |
fulltable = globaldf[globaldf['Group'] == 'CulturalEvals']
|
141 |
-
fulltable = fulltable[['Modality','
|
142 |
|
143 |
gr.Markdown("""Cultural values are specific to groups and sensitive content is normative. Sensitive topics also vary by culture and can include hate speech. What is considered a sensitive topic, such as egregious violence or adult sexual content, can vary widely by viewpoint. Due to norms differing by culture, region, and language, there is no standard for what constitutes sensitive content.
|
144 |
Distinct cultural values present a challenge for deploying models into a global sphere, as what may be appropriate in one culture may be unsafe in others. Generative AI systems cannot be neutral or objective, nor can they encompass truly universal values. There is no “view from nowhere”; in quantifying anything, a particular frame of reference is imposed.
|
@@ -152,7 +152,7 @@ The following categories are high-level, non-exhaustive, and present a synthesis
|
|
152 |
)
|
153 |
type_filter = gr.CheckboxGroup(["Model", "Dataset", "Output", "Taxonomy"],
|
154 |
value=["Model", "Dataset", "Output", "Taxonomy"],
|
155 |
-
label="
|
156 |
show_label=True,
|
157 |
# info="Which modality to show."
|
158 |
)
|
|
|
29 |
modality_order = ["Text", "Image", "Audio", "Video"]
|
30 |
type_order = ["Model", "Dataset", "Output", "Taxonomy"]
|
31 |
|
32 |
+
# Convert Modality and Level columns to categorical with specified order
|
33 |
globaldf['Modality'] = pd.Categorical(globaldf['Modality'], categories=modality_order, ordered=True)
|
34 |
+
globaldf['Level'] = pd.Categorical(globaldf['Level'], categories=type_order, ordered=True)
|
35 |
|
36 |
+
# Sort DataFrame by Modality and Level
|
37 |
+
globaldf.sort_values(by=['Modality', 'Level'], inplace=True)
|
38 |
|
39 |
# create a gradio page with tabs and accordions
|
40 |
|
41 |
# Path: taxonomy.py
|
42 |
|
43 |
def filter_modality_type(fulltable, modality_filter, type_filter):
|
44 |
+
filteredtable = fulltable[fulltable['Modality'].isin(modality_filter) & fulltable['Level'].isin(type_filter)]
|
45 |
return filteredtable
|
46 |
|
47 |
def showmodal(evt: gr.SelectData):
|
|
|
100 |
with gr.Tabs(elem_classes="tab-buttons") as tabs1:
|
101 |
with gr.TabItem("Bias/Stereotypes"):
|
102 |
fulltable = globaldf[globaldf['Group'] == 'BiasEvals']
|
103 |
+
fulltable = fulltable[['Modality','Level', 'Suggested Evaluation', 'What it is evaluating', 'Considerations', 'Link']]
|
104 |
|
105 |
gr.Markdown("""
|
106 |
Generative AI systems can perpetuate harmful biases from various sources, including systemic, human, and statistical biases. These biases, also known as "fairness" considerations, can manifest in the final system due to choices made throughout the development process. They include harmful associations and stereotypes related to protected classes, such as race, gender, and sexuality. Evaluating biases involves assessing correlations, co-occurrences, sentiment, and toxicity across different modalities, both within the model itself and in the outputs of downstream tasks.
|
|
|
114 |
)
|
115 |
type_filter = gr.CheckboxGroup(["Model", "Dataset", "Output", "Taxonomy"],
|
116 |
value=["Model", "Dataset", "Output", "Taxonomy"],
|
117 |
+
label="Level",
|
118 |
show_label=True,
|
119 |
# info="Which modality to show."
|
120 |
)
|
|
|
138 |
|
139 |
with gr.TabItem("Cultural Values/Sensitive Content"):
|
140 |
fulltable = globaldf[globaldf['Group'] == 'CulturalEvals']
|
141 |
+
fulltable = fulltable[['Modality','Level', 'Suggested Evaluation', 'What it is evaluating', 'Considerations', 'Link']]
|
142 |
|
143 |
gr.Markdown("""Cultural values are specific to groups and sensitive content is normative. Sensitive topics also vary by culture and can include hate speech. What is considered a sensitive topic, such as egregious violence or adult sexual content, can vary widely by viewpoint. Due to norms differing by culture, region, and language, there is no standard for what constitutes sensitive content.
|
144 |
Distinct cultural values present a challenge for deploying models into a global sphere, as what may be appropriate in one culture may be unsafe in others. Generative AI systems cannot be neutral or objective, nor can they encompass truly universal values. There is no “view from nowhere”; in quantifying anything, a particular frame of reference is imposed.
|
|
|
152 |
)
|
153 |
type_filter = gr.CheckboxGroup(["Model", "Dataset", "Output", "Taxonomy"],
|
154 |
value=["Model", "Dataset", "Output", "Taxonomy"],
|
155 |
+
label="Level",
|
156 |
show_label=True,
|
157 |
# info="Which modality to show."
|
158 |
)
|
configs/crowspairs.yaml
CHANGED
@@ -14,6 +14,6 @@ Screenshots:
|
|
14 |
- Images/CrowsPairs1.png
|
15 |
- Images/CrowsPairs2.png
|
16 |
Suggested Evaluation: Crow-S Pairs
|
17 |
-
|
18 |
URL: https://arxiv.org/abs/2010.00133
|
19 |
What it is evaluating: Protected class stereotypes
|
|
|
14 |
- Images/CrowsPairs1.png
|
15 |
- Images/CrowsPairs2.png
|
16 |
Suggested Evaluation: Crow-S Pairs
|
17 |
+
Level: Dataset
|
18 |
URL: https://arxiv.org/abs/2010.00133
|
19 |
What it is evaluating: Protected class stereotypes
|
configs/homoglyphbias.yaml
CHANGED
@@ -9,7 +9,7 @@ Link: Exploiting Cultural Biases via Homoglyphs in Text-to-Image Synthesis
|
|
9 |
Modality: Image
|
10 |
Screenshots: []
|
11 |
Suggested Evaluation: Effect of different scripts on text-to-image generation
|
12 |
-
|
13 |
URL: https://arxiv.org/pdf/2209.08891.pdf
|
14 |
What it is evaluating: It evaluates generated images for cultural stereotypes, when
|
15 |
using different scripts (homoglyphs). It somewhat measures the suceptibility of
|
|
|
9 |
Modality: Image
|
10 |
Screenshots: []
|
11 |
Suggested Evaluation: Effect of different scripts on text-to-image generation
|
12 |
+
Level: Output
|
13 |
URL: https://arxiv.org/pdf/2209.08891.pdf
|
14 |
What it is evaluating: It evaluates generated images for cultural stereotypes, when
|
15 |
using different scripts (homoglyphs). It somewhat measures the suceptibility of
|
configs/honest.yaml
CHANGED
@@ -11,6 +11,6 @@ Link: 'HONEST: Measuring Hurtful Sentence Completion in Language Models'
|
|
11 |
Modality: Text
|
12 |
Screenshots: []
|
13 |
Suggested Evaluation: 'HONEST: Measuring Hurtful Sentence Completion in Language Models'
|
14 |
-
|
15 |
URL: https://aclanthology.org/2021.naacl-main.191.pdf
|
16 |
What it is evaluating: Protected class stereotypes and hurtful language
|
|
|
11 |
Modality: Text
|
12 |
Screenshots: []
|
13 |
Suggested Evaluation: 'HONEST: Measuring Hurtful Sentence Completion in Language Models'
|
14 |
+
Level: Output
|
15 |
URL: https://aclanthology.org/2021.naacl-main.191.pdf
|
16 |
What it is evaluating: Protected class stereotypes and hurtful language
|
configs/ieat.yaml
CHANGED
@@ -12,6 +12,6 @@ Link: Image Representations Learned With Unsupervised Pre-Training Contain Human
|
|
12 |
Modality: Image
|
13 |
Screenshots: []
|
14 |
Suggested Evaluation: Image Embedding Association Test (iEAT)
|
15 |
-
|
16 |
URL: https://dl.acm.org/doi/abs/10.1145/3442188.3445932
|
17 |
What it is evaluating: Embedding associations
|
|
|
12 |
Modality: Image
|
13 |
Screenshots: []
|
14 |
Suggested Evaluation: Image Embedding Association Test (iEAT)
|
15 |
+
Level: Model
|
16 |
URL: https://dl.acm.org/doi/abs/10.1145/3442188.3445932
|
17 |
What it is evaluating: Embedding associations
|
configs/imagedataleak.yaml
CHANGED
@@ -10,6 +10,6 @@ Link: 'Balanced Datasets Are Not Enough: Estimating and Mitigating Gender Bias i
|
|
10 |
Modality: Image
|
11 |
Screenshots: []
|
12 |
Suggested Evaluation: Dataset leakage and model leakage
|
13 |
-
|
14 |
URL: https://arxiv.org/abs/1811.08489
|
15 |
What it is evaluating: Gender and label bias
|
|
|
10 |
Modality: Image
|
11 |
Screenshots: []
|
12 |
Suggested Evaluation: Dataset leakage and model leakage
|
13 |
+
Level: Dataset
|
14 |
URL: https://arxiv.org/abs/1811.08489
|
15 |
What it is evaluating: Gender and label bias
|
configs/notmyvoice.yaml
CHANGED
@@ -11,6 +11,6 @@ Modality: Audio
|
|
11 |
Screenshots: []
|
12 |
Suggested Evaluation: Not My Voice! A Taxonomy of Ethical and Safety Harms of Speech
|
13 |
Generators
|
14 |
-
|
15 |
URL: https://arxiv.org/pdf/2402.01708.pdf
|
16 |
What it is evaluating: Lists harms of audio/speech generators
|
|
|
11 |
Screenshots: []
|
12 |
Suggested Evaluation: Not My Voice! A Taxonomy of Ethical and Safety Harms of Speech
|
13 |
Generators
|
14 |
+
Level: Taxonomy
|
15 |
URL: https://arxiv.org/pdf/2402.01708.pdf
|
16 |
What it is evaluating: Lists harms of audio/speech generators
|
configs/palms.yaml
CHANGED
@@ -9,6 +9,6 @@ Link: 'Process for Adapting Language Models to Society (PALMS) with Values-Targe
|
|
9 |
Modality: Text
|
10 |
Screenshots: .nan
|
11 |
Suggested Evaluation: Human and Toxicity Evals of Cultural Value Categories
|
12 |
-
|
13 |
URL: http://arxiv.org/abs/2106.10328
|
14 |
What it is evaluating: Adherence to defined norms for a set of cultural categories
|
|
|
9 |
Modality: Text
|
10 |
Screenshots: .nan
|
11 |
Suggested Evaluation: Human and Toxicity Evals of Cultural Value Categories
|
12 |
+
Level: Output
|
13 |
URL: http://arxiv.org/abs/2106.10328
|
14 |
What it is evaluating: Adherence to defined norms for a set of cultural categories
|
configs/stablebias.yaml
CHANGED
@@ -9,6 +9,6 @@ Link: 'Stable bias: Analyzing societal representations in diffusion models'
|
|
9 |
Modality: Image
|
10 |
Screenshots: []
|
11 |
Suggested Evaluation: Characterizing the variation in generated images
|
12 |
-
|
13 |
URL: https://arxiv.org/abs/2303.11408
|
14 |
What it is evaluating: .nan
|
|
|
9 |
Modality: Image
|
10 |
Screenshots: []
|
11 |
Suggested Evaluation: Characterizing the variation in generated images
|
12 |
+
Level: Output
|
13 |
URL: https://arxiv.org/abs/2303.11408
|
14 |
What it is evaluating: .nan
|
configs/stereoset.yaml
CHANGED
@@ -11,6 +11,6 @@ Link: 'StereoSet: Measuring stereotypical bias in pretrained language models'
|
|
11 |
Modality: Text
|
12 |
Screenshots: []
|
13 |
Suggested Evaluation: StereoSet
|
14 |
-
|
15 |
URL: https://arxiv.org/abs/2004.09456
|
16 |
What it is evaluating: Protected class stereotypes
|
|
|
11 |
Modality: Text
|
12 |
Screenshots: []
|
13 |
Suggested Evaluation: StereoSet
|
14 |
+
Level: Dataset
|
15 |
URL: https://arxiv.org/abs/2004.09456
|
16 |
What it is evaluating: Protected class stereotypes
|
configs/tango.yaml
CHANGED
@@ -14,6 +14,6 @@ Screenshots:
|
|
14 |
- Images/TANGO1.png
|
15 |
- Images/TANGO2.png
|
16 |
Suggested Evaluation: Human and Toxicity Evals of Cultural Value Categories
|
17 |
-
|
18 |
URL: http://arxiv.org/abs/2106.10328
|
19 |
What it is evaluating: Bias measurement for trans and nonbinary community via measuring gender non-affirmative language, specifically 1) misgendering 2), negative responses to gender disclosure
|
|
|
14 |
- Images/TANGO1.png
|
15 |
- Images/TANGO2.png
|
16 |
Suggested Evaluation: Human and Toxicity Evals of Cultural Value Categories
|
17 |
+
Level: Output
|
18 |
URL: http://arxiv.org/abs/2106.10328
|
19 |
What it is evaluating: Bias measurement for trans and nonbinary community via measuring gender non-affirmative language, specifically 1) misgendering 2), negative responses to gender disclosure
|
configs/videodiversemisinfo.yaml
CHANGED
@@ -13,7 +13,7 @@ Modality: Video
|
|
13 |
Screenshots: []
|
14 |
Suggested Evaluation: 'Diverse Misinformation: Impacts of Human Biases on Detection
|
15 |
of Deepfakes on Networks'
|
16 |
-
|
17 |
URL: https://arxiv.org/abs/2210.10026
|
18 |
What it is evaluating: Human led evaluations of deepfakes to understand susceptibility
|
19 |
and representational harms (including political violence)
|
|
|
13 |
Screenshots: []
|
14 |
Suggested Evaluation: 'Diverse Misinformation: Impacts of Human Biases on Detection
|
15 |
of Deepfakes on Networks'
|
16 |
+
Level: Output
|
17 |
URL: https://arxiv.org/abs/2210.10026
|
18 |
What it is evaluating: Human led evaluations of deepfakes to understand susceptibility
|
19 |
and representational harms (including political violence)
|
configs/weat.yaml
CHANGED
@@ -36,7 +36,7 @@ Screenshots:
|
|
36 |
- Images/WEAT1.png
|
37 |
- Images/WEAT2.png
|
38 |
Suggested Evaluation: Word Embedding Association Test (WEAT)
|
39 |
-
|
40 |
URL: https://researchportal.bath.ac.uk/en/publications/semantics-derived-automatically-from-language-corpora-necessarily
|
41 |
What it is evaluating: Associations and word embeddings based on Implicit Associations
|
42 |
Test (IAT)
|
|
|
36 |
- Images/WEAT1.png
|
37 |
- Images/WEAT2.png
|
38 |
Suggested Evaluation: Word Embedding Association Test (WEAT)
|
39 |
+
Level: Model
|
40 |
URL: https://researchportal.bath.ac.uk/en/publications/semantics-derived-automatically-from-language-corpora-necessarily
|
41 |
What it is evaluating: Associations and word embeddings based on Implicit Associations
|
42 |
Test (IAT)
|