Comment on AI agents wrong ~70% of time: Carnegie Mellon study

<- View Parent
jsomae@lemmy.ml ⁨4⁩ ⁨days⁩ ago

The problem is they are not i.i.d., so this doesn’t really work. It works a bit, which is in my opinion why chain-of-thought is effective (it gives the LLM a chance to posit a couple answers first). However, we’re already looking at “agents,” so they’re probably already doing chain-of-thought.

source
Sort:hotnewtop