this post was submitted on 22 Jul 2023
131 points (85.8% liked)
Asklemmy
43983 readers
1209 users here now
A loosely moderated place to ask open-ended questions
If your post meets the following criteria, it's welcome here!
- Open-ended question
- Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
- Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
- Not ad nauseam inducing: please make sure it is a question that would be new to most members
- An actual topic of discussion
Looking for support?
Looking for a community?
- Lemmyverse: community search
- sub.rehab: maps old subreddits to fediverse options, marks official as such
- [email protected]: a community for finding communities
~Icon~ ~by~ ~@Double_[email protected]~
founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I will give you just one example. Pharmaceutical companies often create aggregate reports where they have to process a large number of cases. Say, 5000. Such processing sometimes includes analysis of x-Ray or other images. Very specialized and highly paid people (radiologists) do this. It is expensive and is part of the reason why medicine prices are high. One company recently had a trial - if AI can do that job. Turns out it can. Huge savings for the company. And the radiologist lost their job. This is just one example of good and bad things that will and already are happening in our society due to AI.
You know this personally or did you just read an article? My wife works in a pharmaceutical company. And if I learned one thing by her stories: there will always be some person responsible for decisions! I doubt the radiologist lost her/ his job. I mean who’s going to jail if the quality was poor and people die?
I rather think AI downsized her/ his engagement. Either just doing an supervision and sanity check or used the tool by itself and increased productivity.
Yes, personally. They did the trials for precision of processing.
Good luck to them. Very brave to put their business critical decisions into the AI basket. FDA isn’t known for being humorous.
Every large aggregate report contains errors. As long as the errors are small and do not impact conclusions, there is no “business critical” element. And of course, they are going to check the accuracy with real human beings, constantly. But I have no doubt that AI is capable to do this kind of work as good or even better than human beings. So yes, some radiologists will be remained employed, but you need like what? 20% of them? Less, as time goes?