Hackworth
@Hackworth@piefed.ca
- Comment on Google sues web scraper for sucking up search results ‘at an astonishing scale’ 6 hours ago:
<Three Spidermen Point>
- Comment on Larian CEO Responds to Divinity Gen AI Backlash: 'We Are Neither Releasing a Game With Any AI Components, Nor Are We Looking at Trimming Down Teams to Replace Them With AI' 9 hours ago:
As I understand it, CLIP (and other text encoders in diffusion models) aren’t trained like LLMs, exactly. They’re trained on image/text pairing, which ya get from the metadata creators upload with their photos in Adobe Stock. That said, Adobe hasn’t published their entire architecture.
- Comment on Larian CEO Responds to Divinity Gen AI Backlash: 'We Are Neither Releasing a Game With Any AI Components, Nor Are We Looking at Trimming Down Teams to Replace Them With AI' 1 day ago:
The Firefly image generator is a diffusion model, and the Firefly video generator is a diffusion transformer. LLMs aren’t involved in either process. I believe there are some ChatGPT integrations with Reader and Acrobat, but that’s unrelated to Firefly.
- Comment on New Ways to Corrupt LLMs: The wacky things statistical-correlation machines like LLMs do – and how they might get us killed
2 days ago:
Here’s a metaphor/framework I’ve found useful but am trying to refine, so feedback welcome.
Visualize the deforming rubber sheet model commonly used to depict masses distorting spacetime. Your goal is to roll a ball onto the sheet from one side such that it rolls into a stable or slowly decaying orbit of a specific mass. You begin aiming for a mass on the outer perimeter of the sheet. But with each roll, you must aim for a mass further toward the center. The longer you roll, the more masses sit between you and your goal, to be rolled past or slingshot-ed around. As soon as you fail to hit a goal, you lose. But you can continue to play indefinitely.
The model’s latent space is the sheet. The prompt is your rolling of the ball. The response is the path the ball takes. And the good (useful, correct, original, whatever your goal was) response is the orbit of the mass you’re aiming for. As the context window grows, there are more pitfalls the model can fall into. Until you lose, there’s a phase transition, and the model starts going way off the rails. This phase transition was formalized mathematically in this paper from August.
The masses are attractors that have been studied at different levels of abstraction. And the metaphor/framework seems to work at different levels as well, as if the deformed rubber sheet is a fractal with self-similarity across scale.
One level up: the sheet becomes the trained alignment, the masses become potential roles the LLM can play, and the rolled ball is the RLHF or fine-tuning. So we see the same kind of phase transition in both prompting (from useful to hallucinatory) and in training.
Two levels down: the sheet becomes the neuron architecture, the masses become potential next words, and the rolled ball is the transformer process.
In reality, the rubber sheet has like 40,000 dimensions, and I’m sure a ton is lost in the reduction.
- Comment on Larian CEO Responds to Divinity Gen AI Backlash: 'We Are Neither Releasing a Game With Any AI Components, Nor Are We Looking at Trimming Down Teams to Replace Them With AI' 2 days ago:
Adobe’s image generator (Firefly) is trained only on images from Adobe Stock.
- Comment on Larian CEO Responds to Divinity Gen AI Backlash: 'We Are Neither Releasing a Game With Any AI Components, Nor Are We Looking at Trimming Down Teams to Replace Them With AI' 3 days ago:
Coincidentally, this paper published yesterday indicates that LLMs are worse at coding the closer you get to the low level like assembly or binary. Or more precisely, ya stop seeing improvements pretty early on in scaling up the models. If I’m reading it right, which I’m probably not.
- Comment on Larian CEO Responds to Divinity Gen AI Backlash: 'We Are Neither Releasing a Game With Any AI Components, Nor Are We Looking at Trimming Down Teams to Replace Them With AI' 3 days ago:
There are AI’s that are ethically trained. There are AI’s that run on local hardware. We’ll eventually need AI ratings to distinguish use types, I suppose.
- Comment on Larian CEO Responds to Divinity Gen AI Backlash: 'We Are Neither Releasing a Game With Any AI Components, Nor Are We Looking at Trimming Down Teams to Replace Them With AI' 3 days ago:
Yup! Certifying a workflow as AI-free would be a monumental task now. First, you’d have to designate exactly what kinds of AI you mean, which is a harder task than I think people realize. Then, you’d have to identify every instance of that kind of AI in every tool you might use. And just looking at Adobe, there’s a lot. Then you, what, forbid your team from using them, sure, but how do you monitor that? Ya can’t uninstall generative fill from Photoshop. Anyway, that’s why anything with a complicated design process marked “AI-Free” is going to be the equivalent of greenwashing, at least for a while. But they should be able to prevent obvious slop from being in the final product just in regular testing.
- Comment on It Only Takes A Handful Of Samples To Poison Any Size LLM, Anthropic Finds 4 days ago:
There’s a lot of research around this. So, LLM’s go through phase transitions when they reach the thresholds described in Multispin Physics of AI Tipping Points and Hallucinations. That’s more about predicting the transitions between helpful and hallucination within regular prompting contexts. But we see similar phase transitions between roles and behaviors in fine-tuning presented in Weird Generalization and Inductive Backdoors: New Ways to Corrupt LLMs.
This may be related to attractor states that we’re starting to catalog in the LLM’s latent/semantic space. It seems like the underlying topology contains semi-stable “roles” (attractors) that the LLM generations fall into (or are pushed into in the case of the previous papers).
Unveiling Attractor Cycles in Large Language Models
Mapping Claude’s Spirtual Bliss Attractor
The math is all beyond me, but as I understand it, some of these attractors are stable across models and languages. We do, at least, know that there are some shared dynamics that arise from the nature of compressing and communicating information.
Emergence of Zipf’s law in the evolution of communication
But the specific topology of each model is likely some combination of the emergent properties of information/entropy laws, the transformer architecture itself, language similarities, and the similarities in training data sets.
- Comment on Oracle made a $300 billion bet on OpenAI. It's paying the price. 6 days ago:
Copilot is just an implementation of GPT. Claude’s the other main one.
- Comment on Biblically accurate tree angel 1 week ago:
- Comment on Do you ever feel like your life is "scripted"? Like everything is written by some entity controlling your life? Like you live in a fictional universe? Is this feeling normal/common? 1 week ago:
- Comment on Make me feel like a man 1 week ago:
Æsahættr has entered the chat.
- Comment on Why don't compasses have just two Cardinal directions (North, East, -North, -East)? 1 week ago:
Double plus ungood
- Comment on AI Slop Is Ruining Reddit for Everyone 1 week ago:
Westworld
- Comment on Google’s AI model is getting really good at spoofing phone photos 2 weeks ago:
Yeah, a more honest take would discuss the strengths & weakness of the model. Flux is still better at text than Nano Banana, for instance. There’s no “one model to rule them all,” as much as tech journalism seems to wants to write like that.
- Comment on Google’s AI model is getting really good at spoofing phone photos 2 weeks ago:
Directly, generating higher res stuff requires way more compute. But there are plenty of AI upscalers out there, some better, some worse. These are also built into Photoshop now. The difference between an AI image that is easy to spot and hard to spot is using good models. The difference between an AI image that is hard to spot and nearly impossible to spot is another 20 min of work in post.
- Comment on Google’s AI model is getting really good at spoofing phone photos 2 weeks ago:
Nano Banana Pro’s built into Photoshop now.
- Comment on I'm glad psychedelics entered my life before entheogens became popular. Today's psychonauts are mostly imposters looking for a free ride and social status. Psychedelics aren't meant to be frivolous. 2 weeks ago:
Hail Eris!
- Comment on Absolutely nothing 2 weeks ago:
I was told ignorance would be bliss. I would like a refund.
- Comment on What did I forget? 2 weeks ago:
Present day. Present time!
- Comment on Christina Chong Has a Wild Idea for a 'Doctor Who' and 'Star Trek: Strange New Worlds' Crossover 3 weeks ago:
Borg Daleks?
- Comment on What is your favorite Metroidvania? 4 weeks ago:
Rogue Legacy 2 if it counts. If not, Axiom Verge.
- Comment on OpenAI Introduces 'ChatGPT for Teachers' to Further Destroy the Minds of Our Youth 4 weeks ago:
Khan Academy’s had ChatGPT (Khanmigo) baked into it for nearly three years.
- Comment on "Fed-iverse" can be rephrased as "Fe-diverse", which suits Lemmy more. 4 weeks ago:
I can dig it. I love layered wordplay in any language.
- Comment on Russia’s first AI-powered humanoid robot AIDOL collapses during its onstage debut 5 weeks ago:
I guess those scientist guys all working on A.I. never gave cocaine and Monster Energy a try.
- Comment on Why do all text LLMs, no matter how censored they are or what company made them, all have the same quirks and use the slop names and expressions? 5 weeks ago:
Ctrl+f "attractor state" to find the section. They named it "spiritual bliss."
- Comment on Why do all text LLMs, no matter how censored they are or what company made them, all have the same quirks and use the slop names and expressions? 5 weeks ago:
DeepMind keeps trying to build a model architecture that can continue to learn after training, first with the Titans paper and most recently with Nested Learning. It's promising research, but they have yet to scale their "HOPE" model to larger sizes. And with as much incentive as there is to hype this stuff, I'll believe it when I see it.
- Comment on Why do all text LLMs, no matter how censored they are or what company made them, all have the same quirks and use the slop names and expressions? 5 weeks ago:
Everyone seems to be tracking on the causes of similarity in training sets (and that’s the main reason), so I’ll offer a couple of other factors. System prompts use similar sections for post-training alignment. Once something has proven useful, some version of it ends up in every model’s system prompt.
Another possibility is that there are features of the semantic space of language itself that act as attractors. They demonstrated and poorly named an ontological attractor state in the Claude model card that is commonly reported in other models.
- Comment on Are you going to take the chance that he's kidding? 1 month ago:
I'm super curious about the small text underneath.