How do images affect context? Does an image run separately on another model and returns a text description of it that ends up smaller than the accessibility text tree?