> In the same way, LLMs should speak to us in our favored format - in images, infographics, slides, whiteboards, animations/videos, web apps, etc.

You think every Electron app out there re-inventing application UX from scratch is bad, wait until LLMs are generating their own custom UX for every single action for every user for every device. What does command-W do in this app? It's literally impossible to predict, try it and see!

On the other side of the spectrum, I see some of the latest agents, like Codex, take care to get accessibility right -- something not even many humans bother to do.

It's an extension of how I've noticed that AIs will generally write very buttoned-down, cross-the-ts-and-dot-the-is code. Everything gets commented, every method has a try-catch with a log statement, every return type is checked, etc. I think it's a consequence of them not feeling fatigue. These things (accessibility included) are all things humans generally know they 'should' do, but there never seems to be enough time in the day; we'll get to it later when we're less tired. But the ghost in the machine doesn't care. It operates at the same level all the time

But that's exactly what an LLM solved.

It's the best ui ever.

It understands a lot of languages and abstract concepts.

It will not be necessary at all to let LLM generate random uis.

I'm not a native English speaker. I sometimes just throw in a German word and it just works.

>our favored format - in images, infographics, slides, whiteboards, animations/videos, web apps, etc

If you look at how humans actually communicate I'd guess #1 is text/speech, #2 pictures