Comment on Somebody managed to coax the Gab AI chatbot to reveal its prompt

<- View Parent
kromem@lemmy.world ⁨5⁩ ⁨weeks⁩ ago

People definitely do LoRA with LLMs. This was a great writeup on the topic from a while back.

But I have a broader issue with a lot of discussion on LLMs these days, which is that community testing and evaluation of methods and approaches is typically done on smaller models due to cost, and I’m generally very skeptical as to the generalization of results in those cases to large models.

Especially on top of the increased issues around Goodhart’s Law and how the industry is measuring LLM performance right now.

Personally I prefer avoiding fine tuned models wherever possible and just working more on crafting longer constrained contexts for pretrained models with a pre- or post-processing layer to format requests and results in acceptable ways if needed (latency permitting, but things like Groq are fast enough this isn’t much of an issue).

There’s a quality and variety that’s lost with a lot of the RLHF models these days (though getting better with the most recent batch like Claude 3 Opus).

source
Sort:hotnewtop