this post was submitted on 27 Jan 2024
119 points (91.0% liked)

Arch Linux

7175 readers
2 users here now

The beloved lightweight distro

founded 4 years ago
MODERATORS
 

In my case, there are 95 packages that depend on zlib, so removing it is absolutely the last thing you want to do. Fortunately though, GPT also suggested refreshing the gpg keys, which did solve the update problem I was having.

You gotta be careful with that psycho!

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 5 points 9 months ago (2 children)

I have had some luck asking it follow-up questions to explain what each line does. LLMs are decent at that and might even discover bugs.

You could also copy the conversation and paste it to another instance. It is much easier to critique than to come up with something, and this holds true for AI as well, so the other instance can give feedback like "I would have suggested x" or "be careful with commands like y"

[–] throwwyacc 3 points 9 months ago

This feels like a lot of hoops to avoid reading a wiki page thoroughly But if you want to use gpt this may work

[–] [email protected] 2 points 9 months ago

I’ve also tried that, but with mixed results. Generally speaking, GPT is too proud to admit its mistakes. Occasionally I’ve managed successfully point out a mistake, but usually it just thinks I’m trying to gaslight it.

Asking follow up questions works really well as long as you avoid turning it into a debate. When I notice that GPT is contradicting itself, I just keep that information to myself and make a mental note about not trusting it. Trying to argue with someone like GPT is usually just an exercise in futility.

When you have some background knowledge in the topic you’re discussing, you can usually tell when GPT is going totally off the rails. However, you can’t dive into every topic out there, so using GPT as a shortcut is very tempting. That’s when you end up playing with fire, because you can’t really tell if GPT is pulling random nonsense out of its ass or if what it’s saying is actually based on something real.