I wish I could, but it would kinda be PII for me. Though, to clarify some things:
- I’m mostly not taking about vibe coding. Vibe coding might be okay for quickly exploring or (in)validating some concept/idea, but they tend to pile up a lot of tech debt if you let them.
- I don’t think “more efficient” (in terms of energy and pricing) models are more efficient for work. I haven’t measured it, but the smaller/“dumber” models tend to require more cycles before they reach their goals, as they have to debug their code more among the way. However, with the right workflow (using subagents, etc.), you can often still reach the goals with smaller models.
There’s a difference between efficiency and effectiveness. The hardware is becoming more efficient, while models and tooling are becoming more effective. The tooling/techniques to use LLMs more effectively also tend to burn a LOT of tokens.
- Hardware is getting more efficient.
- Models, tools, and techniques are getting more effective.