Gemini 2.5 Pro Preview (May 2025) is a multimodal reasoning model developed by Google, released as a specialized update ahead of the 2025 Google I/O conference. Identified as gemini-2.5-pro-preview-05-06, this iteration focused on significant enhancements for coding performance and agentic workflows. It was specifically tuned to improve front-end and UI development tasks, allowing users to generate functional web applications and aesthetically refined interfaces from visual or text-based prompts.
The model features a natively multimodal architecture capable of processing text, image, audio, video, and code within a single reasoning cycle. It supports a context window of 1 million tokens, enabling the analysis of extensive datasets such as deep code repositories or long-form video content. During its release, the model demonstrated strong multimodal benchmarks, including a score of 84.8% on the VideoMME evaluation for video understanding.
Technical Improvements
This preview version refined the model's ability to handle complex tool-use and multi-step logical planning. Improvements included higher trigger rates for function calling and a reduction in reasoning errors during sophisticated programming tasks. While it maintained the 'capability over latency' ratio established by previous Pro models, the May 2025 update specifically prioritized reliability in agentic scenarios and better instruction-following for UI-centric development.