this post was submitted on 27 Jul 2023
182 points (96.9% liked)
Programming
17313 readers
251 users here now
Welcome to the main community in programming.dev! Feel free to post anything relating to programming here!
Cross posting is strongly encouraged in the instance. If you feel your post or another person's post makes sense in another community cross post into it.
Hope you enjoy the instance!
Rules
Rules
- Follow the programming.dev instance rules
- Keep content related to programming in some way
- If you're posting long videos try to add in some form of tldr for those who don't want to watch videos
Wormhole
Follow the wormhole through a path of communities [email protected]
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I use ChatGPT frequently for programming and I've found it to be pretty good.
The key is using it conversational nature as this gets better results.
Start simple and expand. You can't just ask it wrote huge chunks of code.
Yeah works well, as long as the code is rather simple and it occurred rather often in the training set. But I seldom use it currently (got a little bit more complex stuff going on). It's good though to find new stuff (as it often introduces a new library I haven't known yet). But actual code... I'm writing myself (tried it often, and the quality just isn't there... and I think it even got worse over the last couple of months as also studies suggest)
Agreed. I got ChatGPT to convert python code to JavaScript and I got a buggy code sample back with new bugs.
I've found it great for asking documentation questions. It saves me a ton of time having to search through documentation myself. The problem is when it encounters something it doesn't have information on, it'll just confidently make shit up, and if you're not enough of an expert to recognize when that happens, you can be mislead. It still saves me time, but I use it as a recall tool to get me started when I'm learning to do something new, I'd never use the code it puts out without reading through it line by line. I'm also experienced enough to know when it's wrong and how to refactor its examples. People new to programming could get set down the wrong path by over relying on gpt to teach them.
I've gotten really good results asking chat gpt for programming help. Problem is that it's wrong like 10% of the time, and when it's wrong it's very confidently incorrect. That wasn't a problem for me because I knew when it was wrong and could course correct it and get the correct solution and it still saved me time and helped me eventually get to the right solution. But if someone who's still getting started is trying to use chat gpt to learn, they could easily be mislead because they won't know when its output is wrong.
Definitely depends on the type of question. I find for documentation type questions I get the 90% good answers, like how do I do something with this library, it's good, which makes sense because that libraries documentation is probably in the training data. But for more open ended questions, like how do I solve this problem, I see similar performance to what you're saying. I think it's a good retrieval and synthesises tool which can really save a ton of time if you already have a high level plan of action and just use it to fill in some specific details.
The code it gives me generally just throws me into the debug stage, skipping right over the me writing buggy code stage.