Users can find the perfect pair of pants, furniture, or kitchen decor simply by describing what they need. Here's how it works
After having expanded AI Mode Google announced a launch in more than 180 Spanish-speaking countries. update in the tool's functionalities: it will now allow explore visually through conversational descriptionsIn other words, users will be able to describe what they imagine—as if they were telling a friend—and naturally obtain related images and products as a result.
For example, if you're looking for inspiration to decorate a bedroom in a "maximalist style," AI Mode will show you images that fit that "visual feel," and you can keep asking "more dark tones?", "more contrast?", or "ideas with bold patterns?", without having to apply manual filters.
Even more: you can start the search by providing an image (uploading it or taking a photo) and then "converse" with AI Mode about what you see.

It makes it easier to find products to buy
If you want to buy something you've seen, simply describe it. For example: "barrel jeans that aren't too baggy," and AI Mode will present you with purchasable visual options. If you like an item, you can go directly to the seller's website.
To do this, Google has a Shopping chart of over 50 billion product listings, which helps you view products from stores around the world, from large retailers to local shops, each with details such as reviews, latest offers, colors and availability. "You'll only see the most recent purchase results, as more than 2 billion of those product listings are updated on Google every hour." the company explains.
Lens technology and the power of Gemini 2.5
To achieve this new functionality, Google combines its expertise in visual search —with Lens and Image Search— with the multimodal capabilities of the model Gemini 2.5 to interpret both language and images.
The most relevant thing is that it uses a technique called “visual search fan-out”: Instead of simply recognizing the main subject of an image, it generates multiple queries about secondary details, context, additional objects, and visual nuances to give more complete answers.
This allows not only the main object to matter, but also its surroundings and visual relationships.
Furthermore, the new modality is integrated in a multimodal way: You can combine text and images, Keep asking questions about what you see in an image, and gradually refine your search.

Google AI Mode operating model
For now, this conversational visual experience is being rolled out in English in the US for AI Mode users.
Original article by Noelia Fraguela | October 02 2025 | Marketing4eCommerce











