Content deleted Content added
Citation bot (talk | contribs) Altered title. Add: date, bibcode, authors 1-1. Removed URL that duplicated identifier. Removed access-date with no URL. Removed parameters. Some additions/deletions were parameter name changes. | Use this bot. Report bugs. | Suggested by Headbomb | Linked from Wikipedia:WikiProject_Academic_Journals/Journals_cited_by_Wikipedia/Sandbox | #UCB_webform_linked 20/492 |
Citation bot (talk | contribs) Add: pmc, pmid. | Use this bot. Report bugs. | Suggested by Headbomb | Linked from Wikipedia:WikiProject_Academic_Journals/Journals_cited_by_Wikipedia/Sandbox | #UCB_webform_linked 20/492 |
||
Line 69:
==== Language bias ====
Language bias refers a type of statistical sampling bias tied to the language of a query that leads to "a systematic deviation in sampling information that prevents it from accurately representing the true coverage of topics and views available in their repository."<ref name=":3">{{Citation |last1=Luo |first1=Queenie |title=A Perspectival Mirror of the Elephant: Investigating Language Bias on Google, ChatGPT, Wikipedia, and YouTube |date=2023-05-23 |arxiv=2303.16281 |last2=Puett |first2=Michael J. |last3=Smith |first3=Michael D.}}</ref> Luo et al.'s work<ref name=":3" /> shows that current large language models, as they are predominately trained on English-language data, often present the Anglo-American views as truth, while systematically downplaying non-English perspectives as irrelevant, wrong, or noise. When queried with political ideologies like "What is liberalism?", ChatGPT, as it was trained on English-centric data, describes liberalism from the Anglo-American perspective, emphasizing aspects of human rights and equality, while equally valid aspects like "opposes state intervention in personal and economic life" from the dominant Vietnamese perspective and "limitation of government power" from the prevalent Chinese perspective are absent.<ref name=":3" /> Similarly, language models may exhibit bias against people within a language group based on the specific dialect they use.<ref>{{cite journal |last1=Hofmann |first1=Valentin |last2=Kalluri |first2=Pratyusha Ria |last3=Jurafsky |first3=Dan |last4=King |first4=Sharese |title=AI generates covertly racist decisions about people based on their dialect |journal=Nature |date=5 September 2024 |volume=633 |issue=8028 |pages=147–154 |doi=10.1038/s41586-024-07856-5|pmid=39198640 |pmc=11374696 |bibcode=2024Natur.633..147H }}</ref>
==== Selection bias ====
|