Which Humans? (2023)

45 pointsposted 3 days ago
by surprisetalk

24 Comments

mncharity

3 days ago

Since the page didn't load for me several times, and the title is ambiguous, here's the Abstract: Large language models (LLMs) have recently made vast advances in both generating and analyzing textual data. Technical reports often compare LLMs’ outputs with “human” performance on various tests. Here, we ask, “Which humans?” Much of the existing literature largely ignores the fact that humans are a cultural species with substantial psychological diversity around the globe that is not fully captured by the textual data on which current LLMs have been trained. We show that LLMs’ responses to psychological measures are an outlier compared with large-scale cross-cultural data, and that their performance on cognitive psychological tasks most resembles that of people from Western, Educated, Industrialized, Rich, and Democratic (WEIRD) societies but declines rapidly as we move away from these populations (r = -.70). Ignoring cross-cultural diversity in both human and machine psychology raises numerous scientific and ethical issues. We close by discussing ways to mitigate the WEIRD bias in future generations of generative language models.

memoriuaysj

3 days ago

[flagged]

observationist

3 days ago

AWFL is my recent favorite - affluent white female liberal. Western would work as well.

catigula

3 days ago

The implicit subtext of 'WEIRD' is "these people are amazing and that's weird" tbh.

jaapz

3 days ago

Is it? Just sounds like fun acronym to me, nothing more

didgetmaster

3 days ago

Surprise, Surprise. LLMs will respond according to the set of data that their model was trained on!

While just about every LLM is trained on data that far surpasses the output of just one person, or even a decent sized group; it will still reflect the average sentiment of the corpus of data fed into it.

If the bulk of the training data was scraped from websites created in 'WEIRD' countries, then it's responses will largely mimic their culture.

user

3 days ago

[deleted]

rokizero

3 days ago

This was submitted 30 months ago. Still interesting. I would be interested if this got 'worse' or 'better' with newer models.

jdkee

3 days ago

As an aside: Last year a student of mine (we're at a U.S. college) told me that his teenage cousins back in Mongolia were all learning English in order to use ChatGPT.

MengerSponge

3 days ago

/usr/bin/humans, presumably

dwa3592

3 days ago

/usr/bin/human/weird, to be precise.

Timwi

3 days ago

Homo peculiaris

levocardia

3 days ago

I think it is mostly a good thing that LLMs have "WEIRD" values. We are at a very fortuitous point in history, where the modal position in extant written text is classically liberal and believes in respecting human rights. Virtually no other point in history would be that way, and a true modal position among values and moral beliefs held among all 8 billion people currently alive on earth -- much less the modal position among all ~100 billion humans ever -- would, I'd hazard to guess, not be a very nice place to end up.

rexpop

2 days ago

You're foolishly myopic if you think your ideology coincides with the global historical best. Have a little humility and consider that there are concepts you haven't even heard of that might be better.

daymanstep

3 days ago

If you go back 200 years, you'll be smack bang in the middle of the Enlightenment with thinkers like Kant and Jeremy Bentham.

You could argue that if you trained LLMs on only texts from that time period, you would get something even more "classically liberal" or human-rights-respecting

g8oz

3 days ago

The extant text also encodes Western hypocrisies and blind spots.

cathyreisenwitz

3 days ago

I wonder whether it might be useful to the continued existence of humanity to correct for the individualism bias in WEIRD countries with some collectivism

andy99

3 days ago

2023, and using some kind of in-page pdf reader from 2002

OrionNox

3 days ago

They deliberately picked "weird". That tells you enought about their ideological position and bias.

user

3 days ago

[deleted]