An Analysis of Societal Bias in Sota NLP Transfer Learning | PyData Global 2021
An Analysis of Societal Bias in Sota NLP Transfer Learning
Speakers: Benjamin Ajayi-Obe, David Hopes
Summary
The popularisation of large pre-trained language models has resulted in their increased adoption in commercial settings. However, these models are usually pre-trained on raw, unprocessed corpora that are known to contain a plethora of societal biases. In this talk, we explore the sources of this bias, as well as recent methods of measuring and mitigating it.
Description
Since the publication of Google’s seminal paper, “Attention is all you need”, attention based transformers have become widely celebrated and adopted for their impressive ability to emulate human-like text. However, it has become increasingly evident that, while these models are very capable of modelling text from a large corpus, they also embed societal biases present in the data. These biases can be difficult to detect unless intentionally inspected for or documented, and so they pose a real risk to organisations
1 view
40
11
2 weeks ago 00:12:30 1
Charas: The World’s First Full-Spectrum Concentrate
3 weeks ago 00:12:05 1
ZenBusiness Review 2024: What Makes It Stand Out?
4 weeks ago 00:08:10 1
AI Agents Will Create MILLIONAIRES in 2025 – Are You Ready
1 month ago 00:08:38 1
Retired General on How Ukraine Is ‘Bleeding Out’ Against Russia | WSJ
1 month ago 01:03:07 1
🔴 Amsterdam: The Shocking Truth the Media Won’t Tell You | Syriana Analysis