Do you have examples of recent models hallucinating when asked to summarize a text?