Yeah it's still hard to deal with LLM gaps (fwiw Study mode would also be prone to this). I do try to catch the super obvious stuff and put up a disclaimer but it's far from perfect.
I had some prototypes basing the generations in websearch but the APIs are still super expensive on that front + the models tend to overindex on the results.