this post was submitted on 17 Feb 2024
85 points (75.4% liked)
Technology
59607 readers
3405 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I pasted your question verbatim into Bing Chat. Here's what it responded with:
Seems like a pretty comprehensive list of the things I'm aware of myself. There's also tons of interesting future applications being worked on that, if they pan out, will be hugely beneficial in all sorts of ways. From what I've seen of what the tech is capable of we're looking at a revolution here.
Seems a bit biased to ask an AI for the benefits of AI......
Not saying anything specific is wrong, just that appearances matter
Was thinking the same.. let's ask Honest Joe the car seller which one is the best mean of transport.
I think implying that it has a bias is giving the Advanced Auto Prediction Engine a bit too much credit.
Oh I am in fact giving the giant auto complete function little credit. But just like any computer system, an AI can reflect the biases of it's creators and dataset. Similarly, the computer can only give an answer to the question it has been asked.
Dataset wise, we don't know exactly what the bot was trained on, other than "a lot". I would like to hope it's creators acted in good judgement, but as creators/maintainers of the AI, there may be an inherent (even if unintentional) bias towards the creation and adoption of AI. Just like how some speech recognition models have issues with some dialects or image recognition has issues with some skin tones - both based on the datasets they ingested.
The question itself invites at least some bias and only asks for benefits. I work in IT, and I see this situation all the time with the questions some people have in tickets: the question will be "how do I do x", and while x is a perfectly reasonable thing for someone to want to do, it's not really the final answer. As reasoning humans, we can also take the context of a question to provide additional details without blindly reciting information from the first few lmgtfy results.
(Stop reading here if you don't want a ramble)
AI is growing yes and it's getting better, but it's still a very immature field. Many of its beneficial cases have serious drawbacks that mean it should NOT be "given full control of a starship", so to speak.
While OP's question is about the benefits, I think it's also important to talk about the drawbacks at the same time. All that information could be inadvertently filtered out. Would you blindly trust the health of you child or significant other to a chatbot that may or may not be hallucinating? Would you want your boss to fire you because the computer determined your recorded task time to resolution was low? What about all those dozens of people you helped in side chats that don't have tickets?
There's a great saying about not letting progress get in the way of perfection, meaning that we shouldn't get too caught on getting the last 10-20% of completion. But with decision making that can affect peoples' lives and livelihoods, we need to be damn sure the computer is going to make the right decision every time or not trust it to have full controls at all.
As the future currently stands, we still need humans constantly auditing the decisions of our computers (both standard procedural and AI) for safely's sake. All of those examples above could have been solved by a trained human gating the result. In the powershell case, my coworker was that person. If we're trusting the computers with at much decision making as that Bing answer proposes, the AI models need to be MUCH better trained at how to do their jobs than they currently are. Am I saying we should stop using and researching AI? No, but not enough people currently understand that these tools have incredibly rough edges and the ability for a human to verify answers is absolutely critical.
Lastly, are humans biased? Yes absolutely. You can probably see my own bias in the construction of this answer.
👏👏👏
Yes, dystopia already arrived and we will all going to suffer. Here are just a few simple examples of blind trust of algorithms which ruined people's lives. And day by day more are coming.
Before AI: https://sg.finance.yahoo.com/news/prison-bankruptcy-suicide-software-glitch-080025767.html
After AI: https://news.yahoo.com/man-raped-jail-ai-technology-210846029.html
https://www.washingtonpost.com/technology/2019/10/22/ai-hiring-face-scanning-algorithm-increasingly-decides-whether-you-deserve-job/
It was in part a demonstration. I see a huge number of questions posted these days that could be trivially answered by an AI.
Try asking Bing Chat for negative aspects of AI, it'll give you those too.