this post was submitted on 05 Jan 2025
265 points (97.8% liked)
Technology
60306 readers
5577 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Search is eventually going to be so enshitified that the way to actually find out things is going to fall back on "ask someone you trust who knows things you don't". At least by that point those trustworthy people should be better informed than in the past..
Maybe we will see the return of lists like what Yahoo was.
It's ultimately self-defeating as well because any future AI is going to be polluted by past AI's garbage content. Making it even harder to develop intelligent AI systems.
It can survive well where there's editorial control. I'd talk to an AI if it had only read encyclopedias for example..
I tried doing some of this. I trained on a corpus of data I wanted it to read, with such a small amount of training data, I found it was overall too lossy. If I asked it a question about something that was in there and it responded there was a really good chance that it was in there. But there was a lot of not knowing something that was definitely in there. It wasn't completely useless but I wouldn't say that it was at the level of being truly helpful.
I worry that there's not enough verified data out there to set up for proper training.
I suspect such a model would have to be far more attuned to its data being smaller but trustworthy. Something like chatGPT for example requires a huge volume because it's weakly affected by any particular datum going in. It's designed to adapt to general conversation norms, rather than specific facts. If you could take a generalist like chatGPT and combine it with an expert model that's been told everything it's told has a huge weighting then that would probably be a big step forward.