this post was submitted on 11 Jan 2024
289 points (96.8% liked)

Technology

59666 readers
3616 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

At a Senate hearing on AI’s impact on journalism, lawmakers backed media industry calls to make OpenAI and other tech companies pay to license news articles and other data used to train algorithms.

you are viewing a single comment's thread
view the rest of the comments
[–] Motavader 3 points 10 months ago (1 children)

Thanks for the link to Common Crawl; I didn't know about that project but it looks interesting.

That's also an interesting point about heavily curated data sets. Would something like that be able to overcome some of the bias in current models? For example, if you were training a facial recognition model, access a curated, open source dataset that has representative samples of all races and genders to try and reduce the racial bias. Anyone training a facial recognition model for any purpose could have a training set that can be peer reviewed for accuracy.

[–] General_Effort 3 points 10 months ago

Face recognition is probably dead as an open endeavor. The surveillance aspect makes it too controversial. I mean that not only will we not see open source work on this, but any work is behind closed doors.

In general, a major problem is that it is often not clear what reducing bias means. With face recognition, it is clear that we just want it to work for everyone. With genAI it is unclear. EG you type "US president" into an image generator. The historical fact is that all US presidents were male, and all but one were white. What's the unbiased output?

One answer is that it should reflect who is eligible for the US presidency. But in the future, one would expect far more people to be of "mixed race". So would that perhaps be biased against "interracial marriage"? In either case, one could accuse the makers of covering up historical injustice. I think in practice, people want image generators that just give them what they want with minimum fuss; wants which are probably biased by social expectations.

In any case, such curated datasets are used to fine-tune models trained on uncurated data. I don't think that is known how such a dataset should look like exactly, to yield an unbiased model (however defined).