canihasaccount
@canihasaccount@lemmy.world
- Comment on She-Ra Lives! 2 months ago:
A bit of an exaggeration, sure. But only a bit. The lay summary of the article I referenced states the following:
Venkataraman et al. find that the paper commits every error that it was possible to make in the paper: leaving out important papers, including irrelevant papers, using duplicate papers, mis-coding their societies, getting the wrong values for “big” versus “small” game, and many others.
“commits every error that it was possible to make in the paper,” and, “completely incorrect,” aren’t very different.
- Comment on She-Ra Lives! 2 months ago:
This study this meme is based on is completely incorrect and should be retracted. Here’s a lay summary of its issues:
whyevolutionistrue.com/…/new-paper-debunks-the-pr…
And the published article detailing the problems with that study’s issues:
- Comment on Amazon, Tesla and Meta among world’s top companies undermining democracy – report 3 months ago:
Examples? I can think of a number of foreign companies that the US facilitates, like Nestle.
- Comment on Oregonian driving 4 months ago:
Oregonians almost take pleasure in driving slowly in front of you. Maybe they’ve just gotten used to going slow because the entire state freeway system is always under construction. People driving crazily is infuriating for a completely different reason.
- Comment on China’s AI overload: Baidu CEO warns of too many models, too few applications 5 months ago:
I actually took that bit out because LLMs are pro climate and against everything that makes the environment worse. That’s a result of being trained on a lot of scientific literature. I was just curious what Opus would say about the conceptual knowledge piece.
- Comment on China’s AI overload: Baidu CEO warns of too many models, too few applications 5 months ago:
Claude Opus disagrees, lol:
I respectfully disagree with the claim that current language models are logically incapable of mapping language to actual concepts. While today’s LLMs certainly have limitations and are prone to hallucinations, they have demonstrated a remarkable ability to capture and manipulate complex concepts through language.
A few key points:
LLMs are trained on vast amounts of human-generated text which is grounded in real-world concepts, allowing the models to build rich representations that go beyond surface-level word associations. Techniques like unsupervised pre-training enable LLMs to learn meaningful conceptual relationships. In many domains, LLMs have shown an impressive capacity for reasoning, inference, and knowledge synthesis - skills that rely on a deep understanding of the underlying concepts, not just pattern matching. Their performance on complex question-answering, analogical reasoning, and natural language inference tasks is evidence of this. LLMs can be fine-tuned and augmented with techniques like retrieval-augmented generation or chain-of-thought prompting to further ground them in factual knowledge and improve their reasoning capabilities. This allows more robust conceptual mappings. Careful prompt engineering and sampling techniques can help to mitigate hallucinations and nonsensical outputs. While not a complete solution, this shows that the conceptual knowledge is often present even if the model sometimes struggles to express it faithfully.
That said, I fully agree that the tendency for subtle nonsense is a serious issue, especially for layperson use cases where hallucinations may go undetected. Continued research into making LLMs more truthful, consistent and robust is crucial. Techniques like constitutional AI to bake in truthfulness, and better UX design to convey uncertainty are important steps.
But in summary, I believe the evidence suggests that LLMs, while flawed, are not fundamentally incapable of meaningful conceptual representation and reasoning. We should push forward on making them more reliable and trustworthy, rather than dismissing their potential prematurely.
- Comment on We all know it's true 6 months ago:
I’m back on my BS is also a solid contributor
- Comment on A New York Times copyright lawsuit could kill OpenAI 11 months ago:
Would you, after devoting full years of your adult life to the unpaid work of learning the requisite advanced math and computer science needed to develop such a model, like to spend years more of your life to develop a generative AI model without compensation? Within the US, it is legal to use public text for commercial purposes without any need to obtain a permit. Developers of such models deserve to be paid, just like any other workers, and that doesn’t happen unless either we make AI a utility (or something similar) and funnel tax dollars into it or the company charges for the product so it can pay its employees.
I wholeheartedly agree that AI shouldn’t be trained on copyrighted, private, or any other works outside of the public domain. I think that OpenAI’s use of nonpublic material was illegal and unethical, and that they should be legally obligated to scrap their entire model and train another one from legal material. But developers deserve to be paid for their labor and time, and that requires the company that employs them to make money somehow.
- Comment on Today's web is the opposite of what early Internet utopians had in mind. Now the situation is somewhat similar climate change: even committed activists can no longer turn the tide for the better. 11 months ago:
- Comment on China Tries To Censor Data About Nearly 1 Billion People in Poverty 1 year ago:
Wow, a real, live, tankie!
- Comment on China Tries To Censor Data About Nearly 1 Billion People in Poverty 1 year ago:
Wow, a real, live tankie!
- Comment on Study finds that Chat GPT will cheat when given the opportunity and lie to cover it up later. 1 year ago:
GPT-4 will. For example, I asked it the following:
What is the neighborhood stranger model of fluid mechanics?
It responded:
The “neighborhood stranger model” of fluid mechanics is not a recognized term or concept within the field of fluid mechanics, as of my last update in April 2023.
Now, obviously, this is a made-up term, but GPT-4 didn’t confidently give incorrect answer. Other LLMs will. For example, Bard says,
The neighborhood stranger model of fluid mechanics is a simplified model that describes the behavior of fluids at a very small scale. In this model, fluid particles are represented as points, and their interactions are only considered with other particles that are within a certain “neighborhood” of them. This neighborhood is typically assumed to be a sphere or a cube, and the size of the neighborhood is determined by the length scale of the phenomena being studied.
- Comment on I want to talk in an American accent but how can I transition into it slowly for people who know me without them noticing a sudden change? 1 year ago:
There are still people who have terrible American accents in media. Lucifer’s twin, for example, was so ridiculously bad. The only person without an American accent who I’ve ever seen pull one off in media was Hugh Laurie in later seasons of House. I still find most attempts amusing, even with coaching.
- Comment on Average Lemmy Active Users by Month 1 year ago:
Linux is a hell of a drug
- Comment on Hobbyte 1 year ago:
I interpreted it as showing that 8 hobbytes were equivalent to a hobbit. I didn’t see that it could be interpreted as saying each little frodo picture under the hobbyte was a hobbit until your comment.
- Comment on Hobbyte 1 year ago:
But a byte is 8 bits, not the other way around
- Comment on The pirates are back - Anew study from the European Union’s Intellectual Property Office (EUIPO) suggest that online piracy has increased for the first time in years. In fact, piracy rates have bee... 1 year ago:
Glad to hear this classic may make a comeback
- Comment on 5 feet apart = Not gay 1 year ago:
You can choose to ignore things you don’t understand on Lemmy. I don’t go into the Risa posts because I know they’re not for me. If something doesn’t make sense and you want to ask about it, go for it, but don’t get upset when the explanation isn’t one that makes sense to you. For a lot of people, popular Vines were everywhere for a while. Not everything on the internet is for you.
- Comment on Google User Data Has Become a Favorite Police Shortcut 1 year ago:
What works as a phone app instead of Google Maps? The only thing I’ve seen is Magic Maps, but that shares location data with third parties, so that seems like an awful solution.
- Comment on Authors Are Furious After Finding Their Works on List of Books Used To Train AI 1 year ago:
Yeah, accurately simulating a single pyramidal neuron requires an eight-layer deep neural network:
- Comment on Free raisins! 1 year ago:
Nick (fine print) learned that by choosing violence, he gets more of the limited resources.
- Comment on Pizza 1 year ago:
So close, and yet so concussed
- Comment on Unity Claims PlayStation, Xbox & Nintendo Will Pay Its New Runtime Fee On Behalf Of Devs 1 year ago:
Nintendo looooves to sue people over so many things though
- Comment on A disturbing number of TikTok videos about autism include claims that are “patently false,” study finds 1 year ago:
Adderall makes nearly everyone work-motivated. That’s why college students abuse it when studying/writing, and why the old meth commercials used to describe people cleaning their spotless houses.
- Comment on Imagine 1 year ago:
Good lord, I hope no one employed at Microsoft reads this. I would bet they institute it if they think of it.
- Comment on [deleted] 1 year ago:
Popularize the apps that exist. I couldn’t figure out how to browse it in a Reddit-like way until I tried an app. That was all I needed to make the switch.
- Comment on OpenAI confirms that AI writing detectors don’t work 1 year ago:
They tried training an AI to detect AI, too, and failed
- Comment on It shouldn't be called ADHD; it should be called restless brain syndrome. 1 year ago:
We’ve known that this isn’t true since 2005:
www.sciencedirect.com/…/S000632230500171X
In addition, plenty of other disorders show worse executive function than ADHD:
www.frontiersin.org/articles/10.3389/…/full
If executive dysfunction is your primary issue, that is not indicative of ADHD. ADHD is driven by reward processing dysfunction and slower information processing:
journals.sagepub.com/doi/full/…/1087054714558872
www.nature.com/articles/s41398-021-01758-0
Note that both poorer information processing and reward processing dysfunction can produce poorer performance on executive function tasks, too.
- Comment on It shouldn't be called ADHD; it should be called restless brain syndrome. 1 year ago:
I would be surprised if he said that, because that’s blatantly incorrect:
www.frontiersin.org/articles/10.3389/…/full (Table 2, for example).
Schizophrenia, bipolar disorder, and major depressive disorder all produce or are associated with executive function impairments that are more severe than ADHD. ADHD is the only one of those with its specific pattern of attentional and reward-related abnormalities, but broad EF deficits are common across forms of psychopathology.