Comment on Squiggly Boie

<- View Parent
NotANumber@lemmy.dbzer0.com ⁨1⁩ ⁨day⁩ ago

To be more specific this is an MLP (Multi-Layer Perceptron). Neural Network is a catch all term that includes other things such as Convolutional Neural Networks (CNNs), Recurrent Neural Networks, Diffusion models and of course Transformers.

What you are arguing online is some variant of a Generative Pre-Trained Transformer, which do have MLP or MoE layers but that’s only one part of what they are. They also have multi-headed attention mechanisms and embedding + unembedding vectors.

I know all this and wouldn’t call myself a machine learning expert. I just use the things. Though I did once train a simple MLP like the one in the picture. I think it’s quite bad calling yourself a machine learning expert and not knowing all of this stuff and more.

source
Sort:hotnewtop