There may be exceptions but everything I’ve seen from AI programming is next level trash. It’s like copy pasting from Stack Overflow without the thousand comments all around it saying DO NOT DO THIS!
When ChatGPT was just released to the general public I wanted to try it out. I had it write a script to handle some simple parsing of network log files. I was having some intermittent issue with my home network I couldn’t figure out, so I had logged a lot of data and was hoping to figure out the issue. But I needed to filter out all the routine stuff that would be just noise in the background. I could have written it myself in about an hour, but figured hey maybe ChatGPT can help me bang it out in a couple of minutes.
The code it wrote looked at a glance to be very good and I was impressed. However as I read it, it turned out to be total nonsense. It was using variables and declaring them after. Halfway the script it seemed to have switched to a completely different approach leaving some sort of weird hybrid between the two. At one point it had just inserted pseudo code instead of actual functional code. Every attempt to get it to fix it’s issues just made it worse. In the end I just wrote the script myself.
I’ve seen examples from other people who attempted to use it and it’s just bad. It’s like having a junior programmer high on weed writing your code, checking it and fixing it takes more time than just writing the code itself.
Then there’s the issue of copyright, a lot of the training data wasn’t licensed and stuff like Github Copilot want to add your data to it’s training set if you want to use it. That’s not OK on many levels and not even possible for people working on corporate codebases.
A lot of programmers work on big code bases, with things like best practices and code standards. Not only does the AI not know the codebase and thus wouldn’t know how to do a lot of stuff in that codebase, it also doesn’t know about the best practices and code standards. So for those kinds of situations it isn’t useful.
I feel like people ask it to do some first year student programming tutorial tasks and the result looks somewhat like what one would expect and conclude the thing can actually write code. It really can’t in reality and probably shouldn’t even if it could.
That’s what I mean though. It helps give you different ideas, maybe looking at the problem a different way, but I don’t trust the garbage it spits out. At least half the time it makes something up, or it gives a solution that just won’t work, and even then it will double or even triple down on it.
Yeah but it’s missing the discussion around the code, why it’s bad, why it may have been once correct but wrong now etc. It strips the context from the search result, which in my opinion makes it useless.
There may be exceptions but everything I’ve seen from AI programming is next level trash. It’s like copy pasting from Stack Overflow without the thousand comments all around it saying DO NOT DO THIS!
When ChatGPT was just released to the general public I wanted to try it out. I had it write a script to handle some simple parsing of network log files. I was having some intermittent issue with my home network I couldn’t figure out, so I had logged a lot of data and was hoping to figure out the issue. But I needed to filter out all the routine stuff that would be just noise in the background. I could have written it myself in about an hour, but figured hey maybe ChatGPT can help me bang it out in a couple of minutes.
The code it wrote looked at a glance to be very good and I was impressed. However as I read it, it turned out to be total nonsense. It was using variables and declaring them after. Halfway the script it seemed to have switched to a completely different approach leaving some sort of weird hybrid between the two. At one point it had just inserted pseudo code instead of actual functional code. Every attempt to get it to fix it’s issues just made it worse. In the end I just wrote the script myself.
I’ve seen examples from other people who attempted to use it and it’s just bad. It’s like having a junior programmer high on weed writing your code, checking it and fixing it takes more time than just writing the code itself.
Then there’s the issue of copyright, a lot of the training data wasn’t licensed and stuff like Github Copilot want to add your data to it’s training set if you want to use it. That’s not OK on many levels and not even possible for people working on corporate codebases.
A lot of programmers work on big code bases, with things like best practices and code standards. Not only does the AI not know the codebase and thus wouldn’t know how to do a lot of stuff in that codebase, it also doesn’t know about the best practices and code standards. So for those kinds of situations it isn’t useful.
I feel like people ask it to do some first year student programming tutorial tasks and the result looks somewhat like what one would expect and conclude the thing can actually write code. It really can’t in reality and probably shouldn’t even if it could.
That’s what I mean though. It helps give you different ideas, maybe looking at the problem a different way, but I don’t trust the garbage it spits out. At least half the time it makes something up, or it gives a solution that just won’t work, and even then it will double or even triple down on it.
Yeah but it’s missing the discussion around the code, why it’s bad, why it may have been once correct but wrong now etc. It strips the context from the search result, which in my opinion makes it useless.
Well said… Thanks for spelling it out!