> The hard part is making an LLM that reliably ignores instructions that aren't delineated by those special tokens.

That's the part that's both fundamentally impossible and actually undesired to do completely. Some degree of prioritization is desirable, too much will give the model an LLM equivalent of strong cognitive dissonance / detachment from reality, but complete separation just makes no sense in a general system.