My experience has been the opposite — a shell prompt is too many degrees of freedom for an LLM, and it consistently misses important information.

I’ve had much better luck with constrained, structure tools that give me control over exactly how the tools behave and what context is visible to the LLM.

It seems to be all about making doing the correct thing easy, the hard things possible, and the wrong things very difficult.