Comment on Nvidia delivers first Vera Rubin AI GPU samples to customers — 88-core Vera CPU paired with Rubin GPUs with 288 GB of HBM4 memory apiece

<- View Parent
in_my_honest_opinion@piefed.social ⁨6⁩ ⁨days⁩ ago

I was unclear I guess, I was talking about injecting other models, running their prediction pipeline for the specific topic, and then dropped out of the window to be replaced by another expert. This functionality handled by a larger model that is running the context window. Not nested models, but interchangeable ones dependent on the vector of the tokens.

Currently this is handled with MCP servers and as I understand it those use natural language.

source
Sort:hotnewtop