Seems like in this new AI world that the word sandbox is used to describe a system that asks "are you sure".
I'm used to a different usage of that word: from malware analysis, a sandbox is a contained system that is difficult to impossible to break out of so that the malware can be observed safely.
Applying this to AI, I think there are many companies trying to build technical boundaries stronger than just "are you sure" prompts. Interesting space to watch.
Yeah, this is also a group of people who refer to gentle suggestions as “guardrails.” It’s not clear they’ve ever read a single security paper.