(no title)
jari_mustonen | 9 months ago
The bias toward the first presented candidate is interesting. The effect size for this bias is larger, and while it is generally consistent across models, there is an exception: Gemini 2.0.
If things in the beginning of the prompt are considered "better", does this affect chat like interface where LLM would "weight" first messages to be more important? For example, I have some experience with Aider, where LLM seems to prefer the first version of a file that it has seen.
h2zizzle|9 months ago
As for gender bias being a reflection of training data, LLMs being likely to reproduce existing biases without being able to go back to a human who made the decision to correct it is a danger that was warned of years ago. Timnit Gebru was right, and now it seems that the increasing use of these systems will mean that the only way to counteract bias will be to measure and correct for disparate impact.
nottorp|9 months ago
You can clearly cut off the name, gender, marital status.
You can eliminate their age, but older candidates will possibly have more work experience listed and how do you eliminate that without being biased in other ways?
You should eliminate any free form description of their job responsabilities because the way they phrase it can trigger biases.
You also need to cut off the work place names. Maybe they worked at a controversial place because it was the only job available in their area.
So what are you left with? Last 3 jobs, and only the keywords for them?
jari_mustonen|9 months ago
Similar examples can also be made for name and gender.
soerxpso|9 months ago
empath75|9 months ago
It seems weird to even include identifying material like that in the input.