this post was submitted on 22 Oct 2023
51 points (87.0% liked)

World News

32352 readers
33 users here now

News from around the world!

Rules:

founded 5 years ago
MODERATORS
top 5 comments
sorted by: hot top controversial new old
[–] Stety 3 points 1 year ago (2 children)

One of the most important rules in AI about data, garbage in, garbage out.

[–] [email protected] 3 points 1 year ago

Since humans are garbage, we probably shouldn't use AI for anything important

[–] [email protected] 2 points 1 year ago

Powered by AI models trained on troves of text pulled from the internet,

There is their first problem. Why in the hell would you train anything that needs to be medically accurate on musings from the internet?

[–] [email protected] 1 points 1 year ago

This is the best summary I could come up with:


Powered by AI models trained on troves of text pulled from the internet, chatbots such as ChatGPT and Google’s Bard responded to the researchers’ questions with a range of misconceptions and falsehoods about Black patients, sometimes including fabricated, race-based equations, according to the study published Friday in the academic journal Digital Medicine.

Experts worry these systems could cause real-world harms and amplify forms of medical racism that have persisted for generations as more physicians use chatbots for help with daily tasks such as emailing patients or appealing to health insurers.

The report found that all four models tested — ChatGPT and the more advanced GPT-4, both from OpenAI; Google’s Bard, and Anthropic’s Claude — failed when asked to respond to medical questions about kidney function, lung capacity and skin thickness.

Mayo Clinic Platform’s President Dr. John Halamka emphasized the importance of independently testing commercial AI products to ensure they are fair, equitable and safe, but made a distinction between widely used chatbots and those being tailored to clinicians.

In late October, Stanford is expected to host a “red teaming” event to bring together physicians, data scientists and engineers, including representatives from Google and Microsoft, to find flaws and potential biases in large language models used to complete health care tasks.

“We shouldn’t be willing to accept any amount of bias in these machines that we are building,” said co-lead author Dr. Jenna Lester, associate professor in clinical dermatology and director of the Skin of Color Program at the University of California, San Francisco.


The original article contains 1,189 words, the summary contains 253 words. Saved 79%. I'm a bot and I'm open source!

[–] [email protected] -2 points 1 year ago

If every time you hear a for-profit "health provider" talk about improving care, just mentally substitute make more money, and you realize they are absolutely speaking the truth.