Also, LLM doesn’t usually have memory or experience. It’s the first page of Google search every time you put in your tokens. A forever trainee that would never leave that stage in their career.
Human’s abilities like pattern recognition, intuition, acummulation of proven knowledge in combination makes us become more and more effective at finding the right solution to anything.
The LLM bubble can’t replace it and also actively hurts it as people get distanced from actual knowledge by the code door of LLM. They learn how to formulate their requests instead of learning how to do stuff they actually need. This outsourcing makes sense when you need a cookie recipe once a year, it doesn’t when you work in a bakery. What makes the doug behave each way? You don’t need to ask so you wouldn’t know.
And the difference between asking like Lemmy and asking a chatbot is the ultimative convincing manner in which it tells you things, while forums, Q&A boards, blogs handled by people usually have some of these humane qualities behind replies and also an option for someone else to throw a bag of dicks at the suggestion of formating your system partition or turning stuff off and on.
heavydust@sh.itjust.works 1 year ago
Yesterday the test team asked me for 3 new features to help them. I thought about it for a few minutes and unederstood that these features are all incompatible. You can get one and only one. Good luck one finding an AI that understands this.
DrFistington@lemmy.world 1 year ago
Exactly. And if AI somehow finds a way to do it, end users will find even more ways to do it wrong