I work in media and frequently use a custom Gemini “Gem” to generate screen-reader-friendly alt text for web illustrations, photos, and charts. For the past year, this workflow has been seamless: I’d upload an image, and Gemini would provide high-quality alt text that needed only minimal editing. But over the last few days, performance has gotten much worse. Gemini now constantly hallucinates, providing detailed descriptions of nonexistent images. After several prompts of basically “that’s wrong, here’s why, try again,” it eventually apologizes and produces accurate alt text like it used to. When I ask how the mistake happened, the model says it’s generating descriptions based on image titles and metadata, despite my explicit instructions to ignore file names and metadata. I’ve tried updating the Gem’s instructions to prioritize accuracy and prevent hallucination, but that hasn’t helped. Has anyone else noticed a sudden decline in image analysis performance? Are there specific prompt adjustments you’d suggest that might fix this? submitted by /u/kbb824
Originally posted by u/kbb824 on r/ArtificialInteligence
