this post was submitted on 11 Sep 2023
104 points (68.6% liked)

Technology

59665 readers
3602 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 10 points 1 year ago (1 children)

If you just let it do a full rewrite again and again, what protects against breaking changes in the API? Software doesn't exist in a vacuum, there might be other businesses or people using a certain API and relying on it. A breaking change could be as simple as the same endpoint now being named slightly differently.

So if you now start to mark every API method as "please no breaking changes for this" at what point do you need a full software developer again to take care of the AI?

I've also never seen AI modify an existing code base, it's always new code getting spit out (80% correct or so, it likes to hallucinate functions that don't even exist). Sure, for run of the mill templates you can use it, but even a developer who told me on here they rely heavily on ChatGPT said they need to verify all the code it spits out, because sometimes it's garbage.

In the end it's a damn language model that uses probability on what the next word should be. It's fantastic for what it does, but it has no consistent internal logic and the way it works it never will.