Google Lens Evolves: AI-Powered Visual Search Expands to Video and Voice

2 Sources

Share

Google enhances its Lens visual search tool with multimodal capabilities, including video and voice inputs, while improving shopping features and processing 20 billion visual searches monthly.

News article

Google Lens: A Revolutionary Visual Search Tool

Google Lens, the AI-powered visual search tool launched in 2017, has become an integral part of the search giant's ecosystem, processing a staggering 20 billion visual searches every month

1

2

. This innovative technology allows users to search for information about objects they see in the real world simply by pointing their smartphone camera at them.

Expanding Capabilities: Multimodal Search

In a significant update, Google has announced that Lens is evolving to support multimodal searches, combining video, images, and voice inputs

2

. This enhancement allows users to interact with Lens in more natural and intuitive ways. For example, users can now point their camera at an object and simultaneously use voice commands like "What kind of clouds are those?" or "What brand of sneakers are those and where can I buy them?"

Video Understanding: A New Frontier

Google is also introducing an experimental feature called "video understanding"

2

. This capability enables Lens to work with real-time video capture, taking the tool beyond static image identification. Users can potentially use this feature to troubleshoot malfunctioning appliances or seek repair tips by capturing a quick video through Lens.

Enhanced Shopping Experience

With 20% of all Google Lens searches being shopping-related, the company is doubling down on improving the shopping experience

1

. The latest update prominently displays key product information when Lens identifies an item in a photo. Users can now see details such as price comparisons across retailers, current deals, product reviews, and purchase options – all powered by Google's Shopping Graph, which contains over 45 billion products

1

.

AI-Driven Context and Information

Google Lens leverages the company's advanced machine learning and AI tools to provide rich context and information about identified objects

2

. This aligns with Google's broader strategy of using generative AI models to enhance search results and provide more comprehensive summaries of information.

Potential Future Applications

The evolution of Google Lens, particularly its video understanding feature, opens up intriguing possibilities for the future. There's potential for making vast video repositories taggable and shoppable

2

. Additionally, the multimodal capabilities of Lens share similarities with Google's Project Astra, hinting at possible applications in future augmented reality devices

2

.

Impact on Search and E-commerce

As Google Lens continues to evolve, it's reshaping how people interact with the world around them and how they shop. The tool's ability to seamlessly blend visual, audio, and now video inputs is pushing the boundaries of search technology and creating new opportunities for e-commerce integration

1

2

. With major players like Amazon and Pinterest also investing in visual search tools, this technology is poised to play an increasingly important role in the future of online shopping and information discovery.

TheOutpost.ai

Your Daily Dose of Curated AI News

Don’t drown in AI news. We cut through the noise - filtering, ranking and summarizing the most important AI news, breakthroughs and research daily. Spend less time searching for the latest in AI and get straight to action.

© 2025 Triveous Technologies Private Limited
Instagram logo
LinkedIn logo