Curated by THEOUTPOST
On Mon, 3 Mar, 7:03 PM UTC
16 Sources
[1]
Forget ChatGPT -- Google Gemini can now see the world with live video and screen-sharing
Google's AI assistant, Gemini, is set to introduce exciting features to give Android users new ways to interact more intuitively with their devices. Leveraging advanced capabilities, Gemini will soon allow users to ask questions about content on their screens, much like the screen sharing feature currently available in Gemini 2.0 on desktop. In a recent announcement, Google unveiled these Gemini functionalities, which focus on real-time interaction and on-screen inquiries. These features are part of Google's Project Astra. The screen-sharing function allows users to share their screens with Gemini and ask questions based on displayed content. For instance, while viewing an image of a jacket, a user might ask for shoe recommendations to complement the attire. The live video interactions, which are undoubtedly a direct response to ChatGPT's Voice and Vision option, let users engage in real-time conversations about their surroundings by enabling the camera within the Gemini app. This functionality allows Gemini to provide insights based on live video feeds, similar to a video call experience. These enhancements position Gemini as a versatile AI assistant capable of understanding and interacting with visual content to deliver support that is more personalized and context-aware. Gemini's new features are designed to integrate seamlessly with various applications such as YouTube. Now, while watching a video, users can activate Gemini to ask questions about the content. For example, a user could inquire about a specific muscle or fitness technique during an exercise tutorial. In addition, when viewing a PDF, the "Ask about this PDF" option allows users to obtain summaries or clarifications, streamlining the research process without moving to a desktop. These features aim to make on-the-go information retrieval more efficient, reducing the need for manual searches and enhancing user productivity. The development of these features falls under Google's Project Astra, a multimodal AI assistant initiative. Project Astra seeks to create an assistant to perceive and understand its environment, facilitating more natural interactions. By enabling Gemini to interpret visual inputs and engage in contextual conversations, Google is advancing toward a more immersive AI experience. Google plans to roll out these features to Gemini Advanced subscribers on Android devices later this month. Google's introduction of screen-aware capabilities in Gemini marks a pivotal moment in AI assistant development. By allowing users to ask questions about on-screen content, Gemini is moving beyond passive viewing into interactive experiences, enhancing AI's utility in everyday life. As these features become widely available, they hold the potential to redefine user expectations and set new benchmarks for what AI assistants can achieve.
[2]
Google Unveils Gemini Live AI: Real-Time Video and Screen Sharing for Smarter Assistance
Google's Gemini AI Gets Live Video and Screen Sharing for Real-Time Assistance Google is revolutionizing AI-powered assistance with live video and screen-sharing features for its Gemini AI. Announced at Mobile World Congress (MWC) 2025 in Barcelona, this update allows users to share their phone screens and surroundings with Gemini AI for instant, context-aware responses. With these new capabilities, users can now display webpages, apps, or images while seeking AI-driven insights, making interactions more dynamic and intuitive. This marks a major leap beyond text-based prompts, bringing real-time AI support closer to everyday needs. screen-sharing feature lets users display webpages, apps, or images while asking questions. The live video function enables AI-powered assistance using a smartphone camera. Google shared a demonstration showing an online shopper using Gemini to get outfit suggestions. The update is a great shift in AI technology. Instead of relying only on text-based prompts, users can now interact through voice and visuals. Google says these features will make AI more responsive and context-aware. will start rolling out to Gemini Advanced subscribers in March. The live video feature was first teased in August 2024. Google also released Gemini 2.0 Pro and an experimental "Flash Thinking" model last month. The latest updates aim to enhance AI's reasoning and explainability. A promotional video released by Google shows a ceramicist who obtains glaze suggestions by using Google Gemini Live. In real-time, the AI provides suggestions as the camera moves around. Technical experts from Google indicate that this technology enables users to make decisions efficiently. Multiple technology announcements from different corporations will take place at Many other technology companies will also present their latest innovations at MWC 2025. The industry considers these developments as part of the roadmap for creating more interactive AI experiences. The evolution of the AI sector demands real-time abilities from companies active within this industry. Users could experience disruptive changes in their digital assistant interactions with Google's current product features. The new tools from Gemini Live will help deliver additional understanding about in AI.
[3]
Google, Oppo, Moto and Honor are finally giving us the AI we deserve
Table of Contents Table of Contents Gemini Live Moto AI and Smart Connect 2.0 Real-time translation from Oppo, Honor, and Tecno Honor's AI Agent Welcome to the era of helpful AI I don't know about you, but I'm tired of the term AI. When every company in the industry is using the term, it loses its meaning, and I'm tired of more AI use cases that I barely use. Artificial intelligence was meant to make our lives easier, but the first era of AI was all about generative uses. It has led to millions of means, can help anyone become an artist, and has led to a deepfake era that phone makers are also hoping to use AI to solve. Recommended Videos Yet, the lack of a use case that everyone desires has also meant we're about to enter the next era of agentic AI. The difference is key: the first era was about creating new content, and the second era is all about AI being used to make your life significantly easier. I've just experienced five different examples of the future of AI, and it's clear that we're finally about to get the AI that we have been dreaming about. Gemini Live Google is pushing AI on smartphones forward for the hundreds of millions of users who use Gemini or OEM partner AI solutions. The future of Google Gemini was on show at the Android Avenue at MWC 2025, and what I saw made me super excited. First, AI is finally addressing one of the biggest issues since the advent of the smart home: getting things to work together properly. Instead of needing to build complex routines yourself, you can now prompt Gemini to do so for you. The example used to demonstrate was of a dog sneaking a cookie from a kitchen counter. The demo showed the user first asking Gemini to review footage and find out whether it was the pet or the child. Then the demo asked Gemini to build a routine so the lights would come on, and a preset broadcast would go out to all speakers to stop this. This is the type of AI feature that I can see my mother using, as is the preview of the new Gemini Live. In the coming weeks, all Gemini Advanced customers will be able to share a screen with Gemini, and mobile users will also be able to use their phone camera to ask questions about their surroundings. The possibilities are endless with these new features, and I love that Google is also making them incredibly user-friendly. They're simple enough that anyone can use them. Moto AI and Smart Connect 2.0 Motorola didn't unveil any new hardware at MWC, but that didn't stop it from showing off a suite of new AI features that will transform how you use your phone. The biggest of these is Smart Connect, which will transform the integration of all Android and Windows devices, not just those made by Motorola and its parent company, Lenovo. It also provides the best ecosystem integration I've seen, aside from Apple's. Smart Connect allows you to link a laptop, tablet, and phone together to open a range of new features. These include easily transferring files and mirroring the screen. There are also some nifty AI tricks, such as being able to search images, documents, and files across all your linked devices and launching apps on a device using Moto AI on another device. In particular, Lenovo and Motorola are taking a different approach to the one deployed by many tech companies. Instead of limiting the feature to just their devices, it's instead available to help any Android or Windows user. It's refreshing, and I hope it's a sign of the things to come in the industry. Real-time translation from Oppo, Honor, and Tecno One of the AI use cases I've been most excited about is real-time translation. I used to be great at learning new languages, but I need to refresh this skill. Yet, I travel a lot and have many meetings where translators are required. I've longed for a setup that made conversations seamless on the go, and Oppo, Honor, and Tecno revealed new solutions that could make this a reality. Oppo launched the world's thinnest foldable -- the Oppo Find N5 -- two weeks before MWC, meaning its presence during the show was focused on its AI efforts in partnership with Google Gemini. The most exciting part of the AI Summit was the reveal of real-time translation during phone calls, which should roll out to users in the coming months. It's unclear whether this is just for Oppo phones, but it's a huge benefit for travel and work. Honor also tackled the same problem using an entirely different approach. While Oppo's translation feature is only for phone calls, Honor's is designed to tackle the problem of live translation during conversations. Its solution? Its first open-ear earbuds, the Earbuds Open, have ANC and offer real-time translation; pass one to the other person, place one in your ear, and you can have a conversation with no one else needed. Tecno took a third approach to solving the same problem but focused more on business meetings and less on portability. Its new MegaBook S14 computer features the ability to translate a conversation or business meeting in real time, allowing you to conduct business without the need for human translators. Particularly interesting is that it runs entirely on-device and offline, so it should even work if you need to have an impromptu conversation in the skies. Each of these features has the ability to make your interactions with other humans much easier, but the next generation of AI is also designed to be the assistant that gets things done for you. Honor's AI Agent There's a lot of growing apps and services that offer some form of agentic artificial intelligence, or in simpler terms, the ability to do things like a human would. Consider the difference between AI right now and a human assistant: if you ask the latter to book a table, flight, or hotel with your set preferences, you're not likely to have to use multiple prompts to do so. Alongside several new products and announcements -- including its new AI-focused Alpha plan corporate strategy, 7 years of updates for its flagship smartphones, and several AI demos -- Honor showed off its new AI Agent. It's a GUI-based mobile AI agent that can read your screen and perform task,s and the example above was used in the company's demo. The actual user experience won't display all the steps, but the demo showed each of the steps it takes. For security reasons, you also have to confirm before the final step is executed. It can work with other third-party applications as well, and it can adapt to use your preferred app instead of the pre-programmed options. Welcome to the era of helpful AI Yes, it doesn't sound as cool as Agentic AI, but the future of AI seems to be about helping people solve complex problems or perform mundane or repetitive tasks. It's designed to help you live life on your terms, and each of these products will help you do that. I've been in plenty of situations where real-time translation is needed, and recording to translate later hasn't always been feasible. Many people would like an assistant for tasks like booking hotels or restaurants just to save them time, while Motorola Smart Connect could have a huge impact on all Android and Windows users. The biggest impact will be from Google Gemini. Google is at the forefront of Android efforts around AI, but it has thankfully recognised that supporting partners is the way to get AI in the hands of the masses. A partnership that started with Samsung Galaxy AI has extended to also working closely with phone makers -- including Motorola, Honor, and Oppo -- to develop and build new AI experiences. I am a big fan of Apple Intelligence Notification Summaries, but in the same week that Apple delayed the next generation of Siri, it's telling that we're getting the AI we deserve from multiple Android phone makers working together.
[4]
I saw the new Gemini and Project Astra, here's why it's the future
Table of Contents Table of Contents Astra features: Gemini Live Video and screen sharing Gemini Live can now read files, documents, and images New features for Google Home: Gemini Routines The future of AI is Google Gemini We're quickly entering the realm of AI that is useful, and key to this is Project Astra, Google's new universal AI agent that is helpful in everyday tasks. Oppo, Honor, Motorola and Tecno have all developed new ways for AI to help you in your daily life, but key to the next generation of artificial intelligence is Astra's multimodal approach. The premise is simple: point your phone camera at something and have a live conversation with Google Gemini, where you can ask it questions and have it provide suggestions based on what it is seeing. Recommended Videos The technology behind complex, and as you can imagine, the rollout of features is happening on a more piecemeal basis. The first two features are finally ready, and ahead of their launch later this month, I got to experience a preview of them alongside other Gemini announcements. What I saw is the future of AI, and I'm super excited: Astra features: Gemini Live Video and screen sharing The big update to Gemini is the new Gemini Live, which gains new visual abilities powered by Project Astra. It makes logical sense that Astra features will help build the next generation of Google Live in more ways than one. If you've been waiting for an AI that can help you understand the world around you, the new video-sharing feature will change your life. The demo involved asking questions related to a pottery business, and Gemini Live was about understanding colors, shapes, and context without needing multiple prompts. As you'll see in the video above, it's absolutely exhilarating, and the possibilities are endless. I have no idea if it's possible, but could Gemini help you change a tire or fix a common engine problem if you've never had to do it before? What about asking it for fashion advice, looking for a medical diagnosis, or live translation of your surroundings while traveling? Of course, there's also the professional use case for this, and the new Gemini Live also supports screen sharing. This will allow you to share your screen, ask questions, and have Gemini guide you through it. I can see this being particularly impactful when performing complex tasks like filing paperwork, learning an advanced subject, or completing financial and tax documentation. These aren't the only advancements in this new Agentic AI, as Google showed off other new Gemini-powered features for products in its ecosystem. Gemini Live can now read files, documents, and images Alongside the screen sharing feature on Gemini Live, Google showcased the ability for Gemini to read and understand a wide variety of images, files, and documents. This feature extends the core premise of Gemini Live to include a variety of different file types. This feature will likely be a huge boon for students, as Google demonstrated how a student may use it. Consider a textbook page on DNA. As shown in the video, Gemini Live can explain the topic in much more detail, search its knowledge base for additional relevant information, and even create a rhyme to help you remember the key facts. The addition of these features will elevate Gemini Live to the next level and will hopefully usher in the era of the next Google Glasses sooner rather than later. The demo took place with the Gemini app on the Galaxy S25 Ultra, so this should be available to all Gemini Advanced users. New features for Google Home: Gemini Routines This demo was firmly designed to show how Gemini AI is evolving the smart home. In many ways, Gemini will be used to deliver the long-awaited autonomous smart home dream. The demo involves a relatable scenario of missing cookies. If you have children, a partner with a sweet tooth, or even a sneaky pet, the new Google Home and Gemini integration will catch them in the act. The demo showed how Gemini can be used to scroll through the footage from a Nest Cam, find the specific moment where the cookies disappeared, and analyze the scene. All from a simple prompt around who ate the missing cookie? With a single command, Gemini can also then set up a new routine that will automatically execute the next time the culprit is spotted on that camera. I can't wait to test out routines further, especially with more complex prompts and routines. The future of AI is Google Gemini I'm impressed with Google's Gemini rollout, at least for its smartphone efforts. The widespread rollout across hundreds of millions of Android devices and the partnerships with different phone makers to develop new features are key drivers of the growth in users and features. Google is the consummate middleman here, helping to fuse different ideas and needs from phone makers as part of its feature roadmap. There will come a time when some features remain exclusive to a specific phone maker, but for now, it's great that all Gemini users can test and experience these advancements. That is, if you're paying for Gemini Advanced. As expected, the Video and Screen-sharing features in Gemini Live are limited to Gemini Advanced users, while it's unclear whether all or some part of the other features will be available without a paid subscription. If you haven't already purchased one, now might be a good time. If you need a new phone, you can also still get 1 year of the Google One AI plan -- which includes Gemini Advanced -- for free with a purchase of either the Pixel 9 Pro, the Pixel 9 Pro XL, or the Pixel 9 Pro Fold. The latest updates to Gemini have me extremely excited about AI on smartphones in the future. The early Gemini features were less useful to me as they focused on more creative endeavors, whereas I am more interested in productivity hacks, but this has changed quite rapidly. For iPhone users, perhaps Gemini can help fill the hole following the delay to the new AI-powered Siri announced earlier this week.
[5]
Gemini Live can now 'see' with live video and screen sharing
Google is rolling out new features to its continuous conversational AI Gemini Live, including live video and screen sharing abilities to really go real-time with your chats. Announced during Mobile World Congress (MWC) in Barcelona, the company's ChatGPT Voice Mode competitor will now let you share your screen or switch to live video on your phone. This will allow you to ask the AI chatbot questions about your immediate surrounds or the website you're currently on with real-time visuals. Google teased Gemini Live's ability to "see" when it launched in August 2024. Now, the live video features will start rolling out to Gemini Advanced subscribers on Android devices later in March. The company posted a video (via The Verge) showing the new live video feature in action, with a ceramicist asking Gemini Live which glazes it would recommend for newly fired vases, aiming for a "mid-century modern look." Yes, it's a marketing video so take it with a grain of salt, but the contextual continuity here is pretty impressive as the camera moves around. The news comes a month after Google released a whole slew of updates including Gemini 2.0 Pro and "enhanced reasoning model" Gemini 2.0 Flash Thinking Experimental which is "capable of showing its thoughts to improve performance and explainability," according to the company. Google's not the only tech giant showcasing at MWC this week, and Mashable's on the ground covering the best tech products that have Barcelona abuzz.
[6]
Google's Gemini now lets you ask questions using videos or what's on your screen | TechCrunch
Google is adding new features to its AI assistant, Gemini, that let users ask it questions using video or content on the screen in real time. At the Mobile World Congress (MWC) 2025 in Barcelona, the company showed off a new "Screenshare" feature, which lets users share what's on their phone's screen with Gemini and ask questions about the company. As an example, the company showed a video of a user shopping for a pair of baggy jeans and asking Gemini what other clothing would pair well with it. As for the video search feature, Google had teased it at Google I/O last year. The feature lets you take a video and ask Gemini questions about it as you're filming. Google said these features will roll out to Gemini Advanced users on the Google One AI Premium plan on Android later this month.
[7]
Google Will Soon Let You Have Live Video Conversations With Gemini
Gemini interface will soon show a "Share screen with Live" button Google is showcasing two new Gemini Live features at the Mobile World Congress (MWC) 2025 in Barcelona. The Mountain View-based tech giant will also be rolling out the Live Video and screen-sharing features to the paid subscribers later this month. Live Video will let users have a video conversation while interacting with the artificial intelligence (AI) chatbot in the Gemini Live interface. The screen-sharing feature, on the other hand, will let users share their screen with Gemini and the AI can answer queries based on what it sees. In a blog post, the tech giant stated that it is hosting interactive demonstrations for the new Gemini features during the MWC 2025. Apart from these, attendees can also check out other Gemini Live features as well as Circle to Search's language translation functionality. Among the two new Gemini features, the first is Live Videos. This was unveiled at Google I/O 2024 and is developed by DeepMind, the AI division of the company, as a part of Project Astra. Live Videos can be accessed within Gemini Live, the AI-powered real-time two-way voice communication feature. With this feature, users can show Gemini a video feed from their device's camera and the AI can process the visual information in real-time and answer queries about them. The company also posted a video on YouTube showcasing the Live Video feature. In the video, a user shows Gemini their pottery collection and asks which colour they should choose for their next vase. The colour options were displayed on the table. Gemini was able to analyse the colours of the pots and the colour options to suggest an option that would complement the rest of the collection almost instantly. Notably, once the feature rolls out, eligible users will be able to see a new video icon in Gemini Live interface to access the feature. The second feature is screen-sharing with Gemini. Currently, the Gemini AI assistant comes with an "Ask about screen" button above the text field. Tapping the button captures a screenshot that users can share with the AI and ask queries about the image. The new feature adds another "Share screen with Live" button. Tapping this will automatically enable the AI to record the screen, and users can ask queries about what they see in real-time. Gemini can not only answer queries about the current screen, but can also follow the user's moves as they switch from one page to another or from one app to the next. Apart from answering general queries, the feature can also help users navigate to certain settings or find a web page. Google confirmed that both of these features will roll out to Gemini Advanced subscribers later in March. These are Android-specific features, and are currently not available on other platforms.
[8]
Gemini Live will learn to peer through your camera lens in a few weeks
At Mobile World Congress, Google confirmed that a long-awaited Gemini AI feature it first teased nearly a year ago is ready for launch. The company's conversational Gemini Live will soon be able to view live video and screen sharing, a feature Google previously demoed as Project Astra. When Gemini's video capabilities arrive, you'll be able to simply show the robot something instead of telling it. Right now, Google's multimodal AI can process text, images, and various kinds of documents. However, its ability to accept video as an input is spotty at best -- sometimes it can summarize a YouTube video, and sometimes it can't, for unknown reasons. Later in March, the Gemini app on Android will get a major update to its video functionality. You'll be able to open your camera to provide Gemini Live a video stream or share your screen as a live video, thus allowing you to pepper Gemini with questions about what it sees. It can be hard to keep track of which Google AI project is which -- the 2024 Google I/O was largely a celebration of all things Gemini AI. The Astra demo made waves as it demonstrated a more natural way to interact with the AI. In the original video, which you can see below, Google showed how Gemini Live could answer questions in real time as the user swept a phone around a room. It had things to say about code on a computer screen, how speakers work, and a network diagram on a whiteboard. It even remembered where the user left their glasses from an earlier part of the video. We don't yet know if the version of this technology that's coming to the Gemini app will be as capable as Google's clearly staged I/O demo. However, Google is talking a big game in advance of Gemini Live's upgrade. The company claims this update to the Gemini 2.0 platform will finally make Gemini a "true assistant." Google suggests you could use Gemini Live's video chops to have an informative conversation with the robot while you explore new places or get help with piecing together an outfit by sharing your screen while online shopping. The more powerful Gemini Live will arrive on Android phones in the Gemini app later this month. Piping a continuous stream of video into the model is undoubtedly going to consume much more processing than churning through some text. The video features will be part of Gemini Advanced, so you'll need to be on the $20 per month AI Premium plan. That subscription also provides access to Google's largest and most expensive AI models. Even with the subscription requirements, Gemini Live will probably lose Google even more money with this update. No company has cracked the code on making money from generative AI just yet. Adding video to the mix will only deepen Google's losses on each user, but this is the kind of feature that could get people to use Gemini more. Securing market share versus the likes of OpenAI is seemingly worth the short-term loss. Despite Google's enormous mobile footprint, its monthly Gemini usage numbers are only in the tens of millions, which is an order of magnitude lower than OpenAI's tools.
[9]
You'll soon be able to show Gemini Live what you see in real time
These features will be available to Gemini Advanced subscribers. It's already been an eventful Mobile World Congress (MWC) with announcements from brands like Samsung, Xiaomi, HMD Global, and more. Google is also at the event where it's previewing two new features for Gemini that will come to Advanced users later this month. Today, Google announced it is showing off Gemini Live's new live video capabilities to MWC attendees. First teased last year during Google I/O, live video allows Gemini to see what your phone's camera sees. For those who can't check out the demo at MWC, the company has also shared a video of the feature in action. The video features a ceramicist who asks Gemini Live to help her pick a glaze for "a mid-century modern look." She then pans the camera around to show Gemini a selection of colors to choose from. Keep in mind that this video should be taken with a grain of salt as its a marketing video. The fine print at the bottom also notes "Results for illustrative purposes and may vary." The other feature Google is introducing is screen-sharing with Gemini Live. Just as it sounds, you'll now be able to share whatever is on your screen with Gemini and ask questions about it. The tech giant also created a video for this feature where the user asks Gemini to come up with outfit ideas for the pants on her screen. As mentioned earlier, these two features are coming to Gemini later this month. However, they will be exclusive to "Gemini Advanced subscribers as part of the Google One AI Premium plan on Android devices." Although they are exclusive now, there's hope these features could eventually come to free users just like Saved Info did a few days ago.
[10]
Google's Gemini Live will now be able to access your camera and screen
Nvidia, Dell, and Super Micro stock tumble as new China probe sparks fears of tighter scrutiny Going beyond just text and image input, select users will be able to use their Android device's camera or share their screen with Gemini Live to ask the chatbot questions about what appears in the live video feed. In the video demonstrations shared by the company, the new and advanced Gemini Live is shown giving style advice via screen share while online shopping, and interior design advice as the camera moves around the room. The new features will only be available for Gemini Advanced subscribers with the $20-a-month Google One AI Premium plan. Google debuted the updates on Monday at the Mobile World Congress 2025 in Barcelona, Spain, but the tech giant has been teasing the new features for almost a year now. Google first started hinting at advanced video capabilities at the Google I/O conference in May 2024, where the tech giant demoed Project Astra, a research prototype of a multimodal AI assistant that could process video input to give you answers on the environment around you, even going as far as to remind you where you last placed something. At the conference in May 2024, Google executives shared a vision for this universal AI assistant that also included the ability to see through smart glasses, although Monday's announcement has not detailed any additional product launches. The unveiling of the latest capabilities comes on the heels of the December introduction of Gemini 2.0, Google's most capable AI agent model, and a new Gemini feature called Deep Research, which can compile research reports on behalf of users.
[11]
You'll Soon Be Able to Send Video and Screenshares to Google's Gemini AI
Originally teased as Project Astra in May of last year, Google's Gemini Live with Video and Gemini Live with Screenshare features each have a new trailer, showing how the search giant plans to catch up with ChatGPT's Advanced Voice Mode. Soon, Google One AI Premium subscribers will be able to share their phone screens and real-time video with Gemini, which the chatbot will then be able to use to answer questions. ChatGPT's Advanced Voice Mode got these features late last year, but it also requires you to shell out at least $20/month for a ChatGPT Plus subscription. While a Google One AI Premium plan is priced identically, it also throws in 2TB of storage for your Drive and Gmail accounts, so depending on your needs, it could be the better pick for you. In the trailers, Google shows users accessing Gemini Live via the Gemini phone app to open a real-time conversation with the AI, then tapping either a video or screen sharing button at the bottom of the screen to get started. In the Gemini Live with Video demo, we see someone showing the chatbot some vases they've just fired as well as some glaze swatches and asking for advice on which one to choose. In the Screenshare demo, the chatbot instead looks at a store listing for a pair of jeans and gives advice on which type of clothes to pair with them. I'd love to meet the type of person who would go through all of the effort of spinning and firing a vase without knowing what color to glaze it, but the point is clear. Gemini will soon be able to use real-time video and screen captures as input when answering prompts. Unfortunately, Google hasn't yet said much more than that. When Project Astra was originally teased, it boasted such impressive (and creepy) abilities as "being able to tell where you live just by looking out the window." These trailers seem significantly scaled back in comparison, but they're also clearly just short demos. I'm curious to see how the feature actually functions once users get their hands on it, which Google says will happen "later this month."
[12]
Google unveils ChatGPT competitor that sees your screen
Google's AI engine introduces a new feature that can "see" the screen and provide relevant information. Initially, this update will be available exclusively for Android users. A year after ChatGPT's breakthrough, Google is joining the race. At the Mobile World Congress in Barcelona, Google unveiled Gemini Live, a development set to transform smartphones into interactive personal assistants capable of seeing and understanding the world around them in real time. About a year ago, Google introduced Project Astra, a demonstration of AI capable of processing live video and answering questions in real time. Now, that technology is becoming a reality with Gemini Live, which will be available to Android users by the end of March. Until now, Gemini could process text, images, and various types of documents. With Gemini Live, it gained a new ability: live video processing and screen sharing. Simply put, users will be able to open their phone's camera, show something to Gemini Live, and receive real-time answers and relevant information. Additionally, users can share their screens and ask for assistance with various tasks, such as online shopping or troubleshooting technical issues. In a demonstration video, a user scanned their phone around a room. Gemini Live provided information about different objects, such as a computer code snippet, speakers, and a network diagram. The AI even remembered where the user had placed their glasses earlier in the video. Google promises that Gemini Live will serve as a versatile personal assistant, helping with tasks like getting information about new places, choosing outfits, or learning new skills while watching instructional videos. Gemini Live's video-processing capabilities will roll out in the Gemini app for Android at the end of March, but access will be restricted to subscribers of the AI Premium plan, which costs $20 per month. This subscription grants access to Google's most advanced AI models. Processing live video is a computationally intensive task, requiring significantly more resources than text or image processing. Google acknowledges that running Gemini Live will demand greater processing power, which may result in substantial financial losses in the early stages. Advertisement However, the company believes that this innovation will drive greater adoption of Gemini and help it compete with OpenAI and other AI industry leaders. Stay updated with the latest news! Subscribe to The Jerusalem Post Newsletter Subscribe Now Competing with ChatGPT's "seeing eye" Despite Google's strong presence in the mobile market, Gemini's adoption remains significantly lower than OpenAI's AI tools, which have already been offering ChatGPT's "seeing eye" feature for about a year. Google hopes that Gemini Live will close the gap and establish its AI as a serious contender in the race for the future of artificial intelligence. Sign up for the Business & Innovation Newsletter >>
[13]
Next Gemini Live update lets Google's AI model see the world through your camera
Editor's take: Google's push to make Gemini a more interactive AI assistant could revolutionize how we use the technology. However, as users share their camera and screen with Gemini, what are the privacy implications? While the benefits of real-time assistance are clear, the potential for data misuse or overreach is also a concern. Google took the stage at the Mobile World Congress (MWC) in Barcelona on Monday to showcase the latest enhancements to its AI assistant, Gemini. The company revealed two new features to make Gemini more interactive and context-aware, including real-time video analysis and screen-sharing capabilities. One cool upgrade to Gemini Live lets users fire up their smartphone camera and point it at objects, surroundings, or even their computer screen for instant analysis and feedback (above). Whether identifying an item, explaining something technical, or helping troubleshoot a problem, Google wants Gemini to be more than just a chatbot - it aims to be a hands-on AI assistant that actually sees what's happening. The second is a new screen-sharing feature that allows users to show Gemini Live their screen (below). The AI can then guide them through tasks, provide app-specific help, or summarize information from displayed content. Google aims to make digital assistance feel less like a chatbot and more like an ever-present AI helper that can interpret and respond to on-screen elements in real-time. However, these features won't come free. Google is locking real-time video analysis and screen sharing behind its AI Premium plan, which costs $20 per month. This move follows the industry trend of placing advanced AI capabilities behind paywalls, like OpenAI's GPT-4.5 access through ChatGPT Plus. There's also the question: How much do you trust giving Google access to your phone's camera? Google previosly demonstrated these capabilities last year for WMC 2024 attendees (below), although it was called Project Astra back then. Through the camera, Gemini could identify landmarks and objects and remember where the demonstrator's glasses were. With screen-sharing enabled, Gemini could assist in tasks like shopping or providing technical support with a simple camera scan. While the demo was impressive, Ars Technica notes that the current AI has problems with video analysis under less ideal (read: nonscripted) conditions. However, the update is more evolved, and the early response has been positive, with beta users praising the potential of an AI assistant that can see and respond to its environment. The rest of the world can soon see for themselves. Google confirmed that the updates will roll out to the Gemini app on Android later this month, with iOS availability expected soon after.
[14]
Google's Gemini Live Assistant is About to Do Some Fun Stuff With...
We may earn a commission when you click links to retailers and purchase goods. More info. Gemini Live is supposed to be an assistant that is there for you at any moment to talk you through an upcoming interview, brainstorm ideas, or have a close-to-normal conversation with AI, if that's something you need. This month, Google is further expanding Gemini Live capabilities by letting it view your world through a camera or by sharing your screen to get help with something on-device. You all remember Project Astra that accompanied the launch of Gemini 2.0 in December, right? Because you likely don't, here's a recap! Project Astra was shown off as the next-generation assistant that could be with you throughout the day with better dialogue, a better memory, improved latency, and the ability to use your camera on phone as well as in upcoming smart glasses. Today, Google announced that Gemini Live will soon be able to use the video and screen sharing capabilities of Project Astra through your device. Once live "later this month," you would be able to fire up Gemini Live and point your camera at objects to ask questions, "collaborate visually," or get help troubleshooting something. You'll also be able to fire up Gemini Live and share your screen to get help or info from apps through an assistant that wants to talk to you. Google shares ideas for this concept that involve asking Gemini Live for help putting together an outfit based on items on screen or to help choose a color to splash onto a vase. As you'll see in each example, the conversation isn't as simple as, "What is this object?" and instead is much more complex than you are probably expecting it to be. You'll find demos of each below and both are pretty wild. You all ready for this? Once live, you'll find camera and screen-sharing shortcuts at the bottom of Gemini Live sessions or Gemini make suggest going live depending on the screen.
[15]
Live video and screen-sharing are coming to Gemini Live this month
Summary Google is showcasing AI capabilities at MWC with Gemini Live's video and screen sharing features. Gemini Live will offer live video and screen sharing, giving advice on products and artistic work. Expect Gemini Live's screen sharing and video support to drop this month for advanced users. Today is the start of MWC in Barcelona, which will run through the sixth, and announcements are already starting to pour in, including fresh news from Google. Unsurprisingly, Google is showcasing more of its AI capabilities at MWC, specifically Gemini Live's video and screen sharing capabilities, something we learned about last month, and now we know these features are set to launch before the month of March is over. Related Google Gemini: 5 ways to use Google's AI-powered assistant day-to-day It can make a lot of everyday tasks a lot easier Posts Gemini Live with screen sharing Get advice from Gemini Live about content you share with your screen Google is at MWC showcasing AI for Android and how it can supposedly help in your day-to-day life, with two new capabilities for Gemini Live: live video and screen sharing (via 9to5Google). Google has dropped a couple of trailers for each feature to further explain how they work. Specifically for Gemini Live with screen sharing, you can watch as the feature is accessed through the Gemini overlay, with a box that appears above the standard Gemini search box to share your screen live (i.e. screensharing with Gemini Live). The point of this is to hold a conversation with Gemini Live about the product/item shared with your screen, like in the video where the user is given fashion advice on a pair of pants. Of course, the question remains why anyone would want Gemini Live to give them fashion advice based on the descriptions of products already listed on the page that's being interpreted by AI. Worse, the video shows how the user has to describe the pant legs as baggy. Apparently Gemini Live wasn't able to do this on its own by looking at the image it was provided access to with screen sharing. But despite the video actually showcasing how poorly Gemini Live can interpret shared screens, there is still likely a use case for doing so. It's too bad Google couldn't find one that's more compelling to share in its example video. Gemini Live with video Get advice while shooting live video Next we have Gemini Live with video, which you will be able to access in the Gemini app through a (upcoming) video button, allowing you to show Gemini Live video in real time while asking questions. The trailer offers an example where an artist is asking advice on what glaze to use on pottery to attain a mid-century modern look. Examples of colors are provided for Gemini to choose from, which it then is asked to match to pottery that's already been glazed brown, replying with a color that indeed matches the brown glaze. But much like the screen sharing video, one has to wonder how useful it is to have Gemini Live choose the colors for your artistic work. Is this actually a helpful function, or should an artist already have a basic understanding of color and how they play off each other? At what point is Gemini removing human skill from the process, and can the end result still be called art? Sure, Gemini Live apparently has the ability to match colors live on video with a basic understanding of what they represent, but as cool as that is, what problem is this feature actually solving? Who is this for, exactly, and why wasn't a better example provided, like using live video to troubleshoot an actual task few understand, like plumbing? Once again, Google could do a much better job convincing people that its AI features are groundbreaking, but it didn't. Expect these new Gemini features to drop this month Screen sharing and live video are coming to Gemini Live soon Beyond the new trailers, Google has posted about the upcoming arrival of Gemini Live's screen sharing and video support on its Keyword blog. Google has revealed we can expect these features to arrive sometime this month for Gemini Advanced users (through a subscription to the Google One AI Premium plan), but it hasn't provided an exact date. At the very least, Google has been busy expanding Gemini's capabilities for a while now. Of course, if you happen to be in Barcelona this week, you can always swing by MWC 2025 to see Gemini Live's screen sharing and video support in action for yourself. Expect to hear more about these two features as we near their launch, whatever day that may be.
[16]
Gemini live video and screensharing arrive on Android devices later this month
Mobile World Congress 2025 has officially kicked off in Barcelona. Google is on the ground previewing two AI features that will begin rolling out to Android devices starting later this month. The first is Live Video, which Google first showed at I/O 2024 last May. The tool takes advantage of Gemini's multi-modal capabilities to allow users to show the chatbot what their phone's camera sees. As you can see from the demo Google shared, Gemini is able to answer a question about mid-century modern decor and offer a suggestion when it comes how to apply that knowledge to pottery. The other new feature, screensharing, works exactly as you would imagine. You can share your screen with Gemini to ask questions about what you see in front of you on a website or app. Google isn't exactly pushing the envelope with either feature. ChatGPT's Advanced Voice Mode has offered both live video and screensharing since the end of last year, but the nice thing here is you won't need to jump into a separate app to take advantage of Gemini's new capabilities. As mentioned, Google will begin rolling out live video and screensharing to Gemini Advanced subscribers starting later this month. Gemini Advanced is included in the company's $20 per month Google One AI Premium plan. If you're a Verizon customer, the carrier recently added the subscription as a perk for some plans.
Share
Share
Copy Link
Google introduces groundbreaking features for Gemini, including live video and screen sharing capabilities, enhancing AI-powered assistance and interaction.
Google has announced a significant upgrade to its AI assistant, Gemini, introducing real-time video and screen sharing features. This development, part of Google's Project Astra, marks a pivotal shift in AI-powered assistance, moving beyond text-based interactions to more intuitive and context-aware support 1.
The new live video feature allows users to engage with Gemini using their smartphone cameras, enabling the AI to provide insights based on the user's surroundings. This functionality mimics a video call experience, where users can ask questions about their environment and receive immediate, context-aware responses 2.
Gemini's screen sharing feature enables users to share their device screens with the AI, allowing for more interactive and personalized assistance. Users can now ask questions about content displayed on their screens, whether it's a webpage, app, or image. This feature is designed to work seamlessly with various applications, including YouTube, where users can inquire about specific details in videos they're watching 1.
The update also includes improved capabilities for analyzing files, documents, and images. Gemini can now provide detailed explanations, search its knowledge base for additional information, and even create mnemonic devices to aid in learning and retention 4.
Gemini's new features extend to smart home integration through Google Home. The AI can now assist in creating complex routines based on camera footage and user prompts, showcasing the potential for more autonomous and intelligent home automation 4.
These new features are set to roll out to Gemini Advanced subscribers on Android devices later in March 2025. The advanced capabilities will be limited to paid subscribers, highlighting Google's strategy to monetize its AI innovations 5.
Google's latest updates to Gemini represent a significant leap in AI assistant technology, potentially setting new benchmarks for what AI can achieve in everyday life. As these features become widely available, they are expected to redefine user expectations and push the boundaries of AI-human interaction 1.
The introduction of these visual and interactive capabilities positions Gemini as a formidable competitor in the AI assistant market, directly challenging other AI platforms like ChatGPT. As the technology continues to evolve, it promises to make AI assistance more accessible, intuitive, and integrated into users' daily lives 3.
Reference
[1]
[2]
[3]
[4]
Google hints at upcoming features for Gemini Advanced, including video generation tools, AI agents, and improved language models, signaling a significant leap in AI capabilities and user experience.
13 Sources
13 Sources
Google introduces Gemini Live, a premium AI-powered chatbot to rival OpenAI's ChatGPT. The new service offers advanced features but faces scrutiny over its pricing and rollout strategy.
6 Sources
6 Sources
Google unveils Gemini AI integration across its ecosystem, challenging Apple's AI efforts. The Pixel 9 and Pixel Buds Pro 2 showcase advanced AI capabilities, signaling a new era in smartphone technology.
12 Sources
12 Sources
Google has made Gemini Live, its conversational AI assistant, freely available to all Android users. This move brings advanced voice AI capabilities to a wider audience, challenging competitors in the AI assistant space.
7 Sources
7 Sources
Google is testing a new interface for Gemini Live that makes interacting with the AI assistant feel more like a phone call, potentially changing how users perceive and interact with AI.
4 Sources
4 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2025 TheOutpost.AI All rights reserved