No benchmarks and evals present, how do you know it produces better result than /create-skill ? Naive testing doesn't provide any confidence
No benchmarks and evals present, how do you know it produces better result than /create-skill ? Naive testing doesn't provide any confidence
I think it means human skill development. It offers learning opportunities to the user.
> When you complete architectural work (new files, schema changes, refactors), Claude offers optional 10-15 minute learning exercises grounded in evidence-based learning science. The exercises use techniques like prediction, generation, retrieval practice, and spaced repetition to provide you with semi-worked examples from across your own project work.
Confusing name though.
When your brain is so cooked on LLMs that mentioning any related terminology triggers Pavlovian response.
Hey, it's awesome that you mention evals. May I ask what you currently use, or look for? Do you roll your own or use an existing framework?