Chatgpt is good for code reviews. I wouldn’t trust it for making code though.
Comment on Only thing getting developed is a headache
virku@lemmy.world 1 year agoChatGPT has taught me so much basic programming I should have learned years ago. Don’t know if I’ll remember any of it since Ill just end up pasting my almost working stuff into ChatGPT again once I’m stuck.
ChiefSinner@lemm.ee 1 year ago
pirat@lemmy.world 1 year ago
Are there any other LLMs or co-pilot tools you’d recommend over ChatGPT, or wouldn’t you trust LLMs in general?
As I see it, since we can keep iterating, reviewing, rewriting prompts, and even use tools like autogen to let multiple bots interact to solve a problem on our behalf, your take is somehow only diggin’ the surface.
ChiefSinner@lemm.ee 1 year ago
Mostly, its my own personal choice / preference.
When I see chatgpt spun up code, sometimes its rock solid. Sometimes it uses weird logic that is hard to follow. I prefer it to review my code, rather than review its.
I’m kind of partial to how military concepts use cases for ai. Like anything that can do damage or complex tasks must be done by a human. Mediocre tasks, I can see a use for it.
Like for instance, write a code to automate scheduling a jobs to backup multiple systems using this fileset to backup or skip I’d feel OK to let ai do. They should all be basically the same. But to script code that is critical to infrastructure and/or complex I feel it is not the right tool to use.
pirat@lemmy.world 1 year ago
Don’t you think it’s just a question of time before it’s becoming good enough or better than us at some of these tasks? I do. With better training data, more parameters, better quantization, and now even memory management with memGPT, I see some very interesting evolutions becoming possible.
To me, saying “all LLMs are the same”, sounds like saying “all humans are the same”, to which I strongly disagree. Different people are good at different things. Otherwise, there’d be no need to hire any specialized workers.
Military usecases is a big ethical discussion. Autonomous drones sound pretty scary, at least as long as the technologies are not fully understood. Many people would probably agree that humans should always stay in control when serious damage could be done. However, many are ok with self-driving cars etc. There are many answers to be found in this regard, but I’m not sure this thread is the place to find them… However, I’d still like to hear if you have any thoughts to add on autonomous machines.
KSPAtlas@sopuli.xyz 1 year ago
In my experience, it tends to have difficulties with restrictions, and will attempt to use the normal method for things no matter what
VantaBrandon@lemmy.world 1 year ago
^^^ How skynet was born, ChatGPT injecting subroutines copied by lackadaisical programmers copying and pasting without checking