The research from Purdue University, first spotted by news outlet Futurism, was presented earlier this month at the Computer-Human Interaction Conference in Hawaii and looked at 517 programming questions on Stack Overflow that were then fed to ChatGPT.
“Our analysis shows that 52% of ChatGPT answers contain incorrect information and 77% are verbose,” the new study explained. “Nonetheless, our user study participants still preferred ChatGPT answers 35% of the time due to their comprehensiveness and well-articulated language style.”
Disturbingly, programmers in the study didn’t always catch the mistakes being produced by the AI chatbot.
“However, they also overlooked the misinformation in the ChatGPT answers 39% of the time,” according to the study. “This implies the need to counter misinformation in ChatGPT answers to programming questions and raise awareness of the risks associated with seemingly correct answers.”
Sure does, but even when wrong it still gives a good start. Meaning in writing less syntax.
Particularly for boring stuff.
Example: My boss is a fan of useMemo in react, not bothered about the overhead, so I just write a comment for the repetitive stuff like sorting easier to write
// Sort members by last name ascending
And then pressing return a few times. Plus with integration in to Visual Studio Professional it will learn from your other files so if you have coding standards it’s great for that.
Is it perfect? No. Does it same time and allow us to actually solve complex problems? Yes.
Agreed and i have the exact same approach. It’s like having a colleague next to you who’s not very good but who’s super patient and always willing to help. It’s like having a rubber duck on Adderall who has read all the documentation that exists.
It seems people are in such a hurry to reject this technology that they fall into the age old trap of forming completely unrealistic expectations then being disappointed when they don’t pan out.