You could ask AI to find antibiotics to kill antibiotic resistant bacteria. The bonus would be to give it a lab and drones to conduct actual tests.
Asklemmy
A loosely moderated place to ask open-ended questions
If your post meets the following criteria, it's welcome here!
- Open-ended question
- Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
- Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
- Not ad nauseam inducing: please make sure it is a question that would be new to most members
- An actual topic of discussion
Looking for support?
Looking for a community?
- Lemmyverse: community search
- sub.rehab: maps old subreddits to fediverse options, marks official as such
- [email protected]: a community for finding communities
~Icon~ ~by~ ~@Double_[email protected]~
Isn't that more of a usecase for genetic/evolutionary algorithms? Those are anything but new, however. I don't really see much use of LLMs here, which is what the current "AI" trend is about.
We’ve been using it at my day job to help us outline ideas for our content writers. It writes garbage content on its own, but it is a decent tool for organizing ideas.
At least that is what we use it for. I’m sure there are other valuable uses, but it is not as valuable (to me at least) as it has been made out to be.
As a programmer, I think it’s scary how AI is now able to write functioning programs out of natural language input now. Sure, it’s not perfect. It’s still pretty mediocre at the task. But a few years ago this was way outside the realm of possibility.
It can even correct the code it has written if there’s any error (with varying results).
What will happen in five years time? Ten years? My fear is that it will only need to be “good enough” to replace most of the programmer’s work. Unlike self driving cars, where “good enough” isn’t good enough.