What to know
- xAI’s Grok chatbot now has advanced visual processing abilities, allowing it to analyze images and visual data.
- Users can upload photos, charts, and screenshots for Grok to interpret, expanding its practical applications.
- Grok combines visual analysis with real-time data, enhancing tasks like news analysis and social media interpretation.
- This update marks a significant step forward for multimodal AI, making Grok more versatile and context-aware.
xAI has released a major update to its Grok chatbot, introducing visual processing capabilities that allow the AI to "see" and interpret the world around it. This enhancement means you can now upload images—such as documents, charts, screenshots, or everyday photos—and Grok will analyze and explain the visual content.
The chatbot can identify objects, interpret data from charts, and even convert visual information into executable code, making it a practical tool for a wide range of tasks.
Introducing Grok Vision, multilingual audio, and realtime search in Voice Mode. Available now.
— Ebby Amir (@ebbyamir) April 22, 2025
Grok habla español
Grok parle français
Grok Türkçe konuşuyor
グロクは日本語を話す
ग्रोक हिंदी बोलता है pic.twitter.com/lcaSyty2n5
Previously, xAI had announced Grok-1.5Vision, but the visual features were not available to the public. With this update, Grok’s visual functionality is now officially online.
The new capabilities are not limited to static analysis; Grok can combine visual input with real-time data from the internet, including social media platforms like X. For example, if you upload a news image, Grok can provide background analysis and interpret events by referencing current information.
GROK CAN SEE WHAT YOU SEE—LITERALLY
— Mario Nawfal (@MarioNawfal) April 20, 2025
Grok’s voice mode comes with camera access, letting users point their phone at something and ask, “What am I looking at?”
The Vision feature on iOS allows the chatbot to analyze real-world objects, text, and environments through your… https://t.co/cmtINP8yp6 pic.twitter.com/N1b6pcYZOi
Grok’s visual intelligence is powered by advanced models that excel in spatial understanding and visual reasoning. In benchmark tests such as RealWorldQA, Grok demonstrates a leading advantage, reflecting its ability to handle complex visual tasks.
This positions Grok as a strong competitor in the rapidly evolving field of multimodal AI, where systems are expected to process and integrate information from multiple sources, including text, images, and audio.
Alongside visual processing, the update introduces multilingual audio capabilities and real-time search in voice mode, further broadening Grok’s utility. These features are currently available on the iOS Grok app, with additional support for Android users who subscribe to the SuperGrok tier.
With these advancements, Grok becomes a more versatile and context-aware assistant, capable of helping you interpret the world visually and textually. This update marks a significant milestone for xAI and the broader AI community, as it pushes the boundaries of what conversational agents can achieve.
Via: TechCrunch
Discussion