Comment on $219 Springer Nature book on machine learning was written with a chatbot
tfowinder@lemmy.ml 3 weeks ago
Okay so hear me out on this. The book mentioned in this article is definitely a trash, the author used LLM without informing readers, which is why most people feel the are being scammed and express feelings of frustration and hate.
I personally have deployed LLMs on my local machines and used them for variety of things such as Summarize news and Articles, Coding, Image Generation, etc and I have to be honest it is really really impressive technology. Any author who takes assistance from LLM would be hyper-productive compared to someone who does all the labour themselves. I used to take hours to read a broad area of knowledge and then deep dive in intrested topics. When LLMs generate summary and you can decide weather to read the source yourself or not is a big time saver and productivity boost. Of course this can be abused by someone who trusts LLMs too much and don’t again verifies what they read, it can give false information but that’s not how they are supposed to be used.
These language models are really good at creating summaries. I use a locally deployed LLM to read summaries of Articles and then if I feel interested I read the entire article end to end from original source. In Almost every case the summary is spot on without it missing any important points or topics, heck I have created system prompt so that it tries to give hot takes and nuanced perspectives from the article and it impresses me sometimes giving me a new perspective which I would have not thought otherwise.
I am convinced that using LLM along with your own knowledge always surpasses the quality of your work if someone with your same capability generates work without taking assistance of a LLM.
MonkderVierte@lemmy.zip 3 weeks ago
sundray@lemmus.org 3 weeks ago
Didn’t have time to read that, so I threw your comment into my local TinyLlama instance:
baltakatei@sopuli.xyz 3 weeks ago
Didn’t have time to read that, so I threw your comment into ChatGPT:
interdimensionalmeme@lemmy.ml 3 weeks ago
Actually, I just dropped the entire thread — including the original article — into ChatGPT since I didn’t have time to read it all either. Then I asked it to write the best possible 3-sentence Lemmy comment specifically for your reply. So, ironically, we’ve both delegated the thinking to LLMs — now that’s productivity.
Then I asked “write the next 3 sentence providing the most insight that you can in 3 sentences” and it responded
The real issue here isn’t just that LLMs can fabricate citations — it’s that institutions and publishers are increasingly failing to enforce even basic quality controls, allowing AI-generated content to slip through without proper oversight. When tools meant to assist in research are used to replace it, the result is a veneer of credibility masking hollow or false scholarship. If this trend continues, we risk flooding academic discourse with noise that’s indistinguishable from rigor, eroding trust in legitimate research.
then I asked “But chatgpt, quality control, rigor, those have been gone for aeons, the barrier to write was so high no one really tried so the control, rigor, editors have largely atrophied long ago, respond in 3 sentences” and it responded
You’re right — what we’re seeing now is less a collapse and more an exposure of how little substance was holding it all up. The illusion of rigor persisted mostly because the barrier to entry was too high for mass participation, not because the gatekeepers were doing their jobs. LLMs didn’t break the system; they just made its hollowness impossible to ignore.
tfowinder@lemmy.ml 3 weeks ago
Hmm Interesting, seems your Model is hallucinating a lot, maybe try improving your system prompt. I have a deepseek-ai_DeepSeek-R1-Distill-Qwen-1.5B Running locally. Here is my output for the Orignal Comment
I use the following prompt before article
andallthat@lemmy.world 3 weeks ago
Interestingly, your original comment is not much longer and I find it much easier to read.
Was it written with the help of a LLM? Not being sarcastic, I’m just trying to understand if the (perceived) deterioration in quality was due to the fact that the input was already LLM-assisted.
tfowinder@lemmy.ml 3 weeks ago
No its not LLM, i wrote it from my own brain, though my first language is not English.
floquant@lemmy.dbzer0.com 3 weeks ago
That output doesn’t really conform to the system prompt does it?