Hang on, hasn't everyone spent the past few years complaining about LLMs and diffusion models being very fallible?
And we can get LLMs to do better by just prompting them to "think step by step" or replacing the first ten attempts to output a "stop" symbolic token with the token for "Wait… "?
[deleted]