7 Sources
7 Sources
[1]
RIP Peripherals? Next-Gen Windows to Lean Heavily on AI, Voice, and Vision
Will Windows users soon abandon mice and keyboards for voice-activated controls? In a video released on Wednesday, Microsoft corporate VP Pavan Davuluri discussed how generative AI will unleash new ways to use the OS. "I think what human interfaces look like today and what they will look like five years from now is one big area of thrust for us that Windows continues to evolve," he said. Although Windows has been expanding to new device types, such as tablets and even gaming handhelds, "Experience diversity is the next space where we will continue to see voice becoming more important," Davuluri said. "So, you'll be able to speak to your computer while writing, inking, interacting with another person, for example. You should be able to have a computer semantically understand your intent to interact with it, for instance, from when it's awake or when you put a machine to sleep," he added. The Microsoft executive also envisions future Windows devices that'll be able to view your desktop screen and visually understand the content. "We certainly look at the interface becoming more multimodal and more capable, based on new interaction technologies that come to life," he said. "And whether that's mouse and keyboard moving to pen and touch, and so on." This suggests Microsoft wants Windows interactions to become more like the computer from Star Trek or the AI in the film Her. Still, the vision might creep people out since it entails Microsoft's AI software collecting more of your data. However, Davuluri also mentioned the company using AI models that can run "on the device," suggesting they can operate without an internet connection and protect user privacy. Some of these AI capabilities are already available for Windows 11 and internet users through Microsoft's Copilot AI. But for now, a large number of PC consumers remain on Windows 10, even as the OS is about to officially lose support in October.
[2]
Microsoft's Idea for Windows Sounds Like a Copilot OS
Microsoft has published a new interview with the head of Windows, Pavan Davuluri, where he shared some interesting details about how Windows is going to change in the future. He wants Windows to more closely integrate with AI technology, unsuprisingly. Davuluri said that computing will become "more ambient, more pervasive, continue to span form factors, and certainly become more multi-modal in the arc of time". This means that the operating system is going to fundamentally change and will begin to prioritize new ways of interacting with your computer, like with your voice and by having the OS understand what's on your screen at all times. Davuluri revealed that Microsoft is making a huge investment in making the OS "agentic" and multi-modal. He also teased that Windows won't look like it does today in the future. The next version of Windows will likely put voice at the center of how you interact with your computer. Davuluri says you'll be able to "speak to your computer while you're writing, inking, or interacting with another person. You should be able to have a computer semantically understand your intent to interact with it," which sounds like an AI that happens to also be a computer. This isn't the first time Microsoft has hinted at this kind of change, another Microsoft executive teased a Windows 2030 Vision video that also uses voice as a first-class input method. It seems clear that Microsoft wants us to be able to talk to our computers in addition to using a keyboard and mouse. The OS will understand your intent based on what is on your screen and will take that into account, making your workflow a lot easier if you choose to use your voice. This will let you use natural language to tell the computer what to do, and it will be able to perform tasks for you. This sounds a lot like how Gemini watches your screen when you share it. However, a lot of people are going to be worried about privacy. It will take a lot of personal data to make these features truly useful, but Davuluri states that Microsoft is building these models with privacy in mind. He says they are being thoughtful when it comes to privacy and security requirements and making sure that "data stays local customers are in control of the choices associated with the decisions themselves." Davuluri said that AI models on the device bring "a bunch of new capabilities and agencies... to the platform and the device itself". He gave a few examples of this, like an improved version of Windows Search. Traditional search is a lexical indexer that searches for keywords. The vision of a new AI would be augmented by a semantic indexer that can understand the content being searched. This gives you a much better chance of finding what you're looking for. It feels like Microsoft has been building up to an AI-focused future, since it keeps adding Copilot to apps. So, while this isn't much of an unexpected turn, it is still surprising. Source: Windows IT Pro, Windows Central
[3]
Microsoft's Windows future is built on AI, voice, cloud, and context
Can you imagine yourself having a conversation with Windows about what your PC is doing? Microsoft's Windows chief can, and is trying to build a future where those interactions are the norm. In an interview with Microsoft AI product manager Christiaan Brinkhoff, the chief of Microsoft's Windows + Devices group, Pavan Davuluri, explained that the company is trying to work toward a future where you can access Windows pretty much anywhere via the cloud, then use AI to fine-tune what you're trying to accomplish. Microsoft described the conversation as "the next chapter of Windows," with an eye toward delivering the changes within the next few years. Davuluri described what he hoped the Windows team could accomplish from a strategic level, without targeting any future version of Windows with these goals in mind. Like many projects at Microsoft, Davuluri's vision begins with businesses in mind. Instead of (or adjacent to) the PC, Microsoft said it's backing the Windows 365 Link, a sort of dumb mini PC that connects directly to the cloud and to Windows 365. "People are not talking about multiple steps; they're just trying to get to the cloud PC as seamlessly as possible," he said. "They can't tell when they're using their local CPU and when they're using a CPU in Azure," Microsoft's cloud, Davuluri added. That's not new. Microsoft began talking about this hybrid compute model a few years ago, when Microsoft chief executive Satya Nadella opened its Build developer conference by specifically calling out how future applications would combine the cloud and local compute capabilities. At Microsoft, that's now common, with Copilot sneaking into just about everything. But Nadella used his keynote to specifically call out Adobe and apps like Photoshop. And that's exactly what happened, there, top: Photoshop now uses your PC's GPU to modify images, but also calls upon the Adobe cloud for generative AI. Davuluri also hearkened further back, stating that Microsoft would expand Windows' modalities into speech and vision. Modalities are what Microsoft uses to describe the various ways you interact with your PC: you can touch the screen, type on the keyboard, or move the mouse. More recently, pens have added inking, too. "I think we will see computing become more ambient, more pervasive, continue to span form factors and certainly become more multimodal in the arc of time," Davuluri said. "I think experience diversity is the next space where we will continue to see voice becoming more important," Davuluri added. "I think more fundamentally, the concept that your computer can actually look at your screen and is context aware is going to become an important modality for us going forward. The other thing I think is going to get more intuitive is multimodal interactions. So you'll be able to speak to your computer while, you know, writing, inking, interacting with another person. For example, you should be able to have a computer semantically understand your intent to interact with it from when it's awake, or to put a machine to sleep." That's a pretty direct reference to Copilot Vision, which can now "see" your screen and talk to you about it. Microsoft built that from two AI models, Phi and Mu, which are "small" language models that can run on your PC. Those models not only enable services like Copilot Vision but can also be used to power Microsoft's new semantic search options in File Explorer and Settings, which can be changed at a user's natural-language request. Microsoft plans to augment those capabilities as assistive technologies, such as evolving Narrator from simply "reading" words on your screen to actively describing what's going on. With voice and vision being added to ChatGPT and other models, people are finding themselves chatting with AI. Will they feel as comfortable doing so with a PC? Finally, Davuluri gave a hint (or not) about the future of Microsoft's Surface PCs. After the departure of former Windows + Devices chief Panos Panay, some wondered if Microsoft would ever return to the diversity of Surface devices that prompted everything from the Surface Duo to the Surface Laptop Studio -- both seemingly on hiatus or discontinued. Davuluri didn't offer any clues. "I think today one of the things that we celebrate in Windows is the diversity on form factors in which computing is available, for sure," he said. "Between the work that the ecosystem partners do and Surface does, we have a pretty broad range of devices in themselves.
[4]
Microsoft exec's vision of Windows 12 and beyond could terrify some of you
The overall vision is for more AI, and an OS that watches what you're doing on-screen, tapping into the cloud, which may worry the privacy conscious Another Microsoft executive has provided their vision of the future of Windows, specifically framed around AI and the cloud, and how this - and voice input - is going to be a big part of the operating system down the line. Windows Central discovered a YouTube interview with Pavan Davuluri, VP of Windows and Devices at Microsoft. See the video clip below, and be warned, the technobabble is strong with this one. Davuluri says at one point: "Computing [will] become more ambient, more pervasive, continue to span form factors, and certainly become more multi-modal in the arc of time." Okay, so let's boil this - and the rest of the interview - down a bit. Computing becoming more "multi-modal" refers to using inputs beyond the traditional mouse and keyboard, and the exec touches on voice commands as an important part of the equation. This echoes what Microsoft's VP for OS Security, David Weston, said earlier this month when explaining his vision of Windows in 2030. Davuluri also says, "Fundamentally, the concept that your computer can actually look at your screen and is context aware is going to become an important modality for us going forward." Again, that follows up on what Weston observed about the next-gen Windows PC being able "to see what we see, hear what we hear, and we can talk to it and ask it to do much more sophisticated things." The key idea appears to be Windows watching what you're doing, using AI to determine context, and then applying that to your actions in the OS, and specifically making voice commands more useful due to that context. Davuluri notes: "You'll be able to speak to your computer while you're writing, inking, or interacting with another person. You should be able to have a computer semantically understand your intent to interact with it." The exec also talks about Windows becoming "increasingly agentic" (with the first AI agent recently having debuted in the Settings app in Windows 11, of course), and how the cloud will be needed to power these AI abilities. (Although some of the work will be on-device, he indicates, as we see with Copilot+ PCs already - hence the need for NPUs with these laptops.) Davuluri observes: "Compute will become pervasive, as in Windows experiences are going to use a combination of capabilities that are local [processed on the device] and that are in the cloud. I think it's our responsibility to make sure they're seamless to our customers." Which is a roundabout way of saying that the level of processing needed for some of these AI powers in next-gen Windows will need to tap the cloud to ensure performance remains responsive enough to feel 'seamless' rather than sluggish. Microsoft has clearly got a hymn sheet somewhere, as its top-level executives appear to be singing the same tune regarding how Windows will evolve as we head into the next decade. It's interesting to pick up on the mirrored points between these two interviews Microsoft has recently presented: more AI (surprise, surprise) that determines context by watching what you're doing on-screen, and also allows voice commands to be more effectively used based on that context - with the cloud at least partly powering all this. Depending on what kind of person you are, this may sound like an exciting new way forward in terms of making it easier to do what you need to do within Windows, or a privacy nightmare. The more paranoid-leaning Windows users out there are likely to be horrified at the suggestions made about the future of the OS here. An operating system that's watching what you do? The way they'll read this angle from Microsoft is that it's turning Windows into an AI-powered surveillance platform - you can guarantee that. And it's obvious where such concerns come from when we're told the "computer can actually look at your screen" and take context from there, and leverage the cloud (read: Microsoft's private servers) to crunch the data on what you're doing with your PC. If this makes using next-gen Windows a breeze, and AI is constantly firing up the apps you need, or searches you want to make, before you get to them, or proactively suggesting files you might want next - or Windows options that could be changed for your benefit in given scenarios - will people even care about what's happening in the cloud? Frankly, the truth is, they probably won't if it makes their computing lives a lot easier.
[5]
Microsoft's vision for the next version of Windows is an all-seeing, voice-controlled chatbot that gives you productivity 'superpowers'
The messaging coming out of Microsoft is becoming unmistakable. The next version of Windows is going to be all about chat. Last week, Microsoft posted a video that claimed, "the world of mousing around and keyboarding around and typing will feel as alien as it does to Gen Z to use DOS." Now there's a new interview on the official Windows IT Pro YouTube channel with Microsoft's Windows and devices boss Pavan Davuluri that doubles down on that idea and predicts, "you'll be able to speak to your computer, while writing, inking, or interacting with another person." When asked how the way we interact with computers is going to change in the next few years, Davuluri says, "I think we will see computing become more ambient, more pervasive, continue to span form factors, and certainly become multimodal in the arc of time. " Multimodal? That means, "voice and vision and pen and touch." Davuluri explains that, "today, one of the things we celebrate in Windows is the diversity on form factors." But in future, "experience diversity is the next space where we will continue to see voice become more important." So, how will that actually work? The good news is that it doesn't seem like Davuluri thinks voice will replace keyboard and mouse, as implied by David Weston, corporate vice president of enterprise and OS security at Microsoft, last week. Instead, it'll be all about that quote from earlier, about a more natural interaction where your PC "can actually look at your screen and is context aware" and will therefore, "semantically understand your intent to interact with it." Without a demo of this new approach, it's hard to be totally sure how it will work. But one can imagine typing away on a word document while making conversational requests, maybe along the lines of, "oh, can you open that webpage on jelly fish clogging up nuclear reactors I was looking at earlier," or, "make that bullet pointed list above numbered" The idea is that it will all be smooth and seamless and frictionless. To put it succinctly, the new AI-powered modalities will allow you to get into a "flow" state when using your PC. Davuluri thinks net impact will give Microsoft, "the ability to for us to find users in the flow of their tasks and have the AI models that help them with the intent and flow of their tasks" and users, "the ability to superpower their productivity on a day to day basis." Davuluri details a few specifics, such as improvements to Windows Search, which will be, "much more meaningful, much more impactful," and also discusses the broader implications of combining local AI models with cloud computing. As for time frames, Microsoft is already bringing some of these features to Windows 11 through its Copilot AI platform. But the gist from these videos seems to be that the big change and perhaps a new OS altogether, call it Windows 12 or even Windows Chat, will arrive within five years and thus probably before 2030. As as ever, watch this space for more.
[6]
Windows 12 may feature built-in AI at its foundation
Microsoft's Pavan Davuluri discussed the future of Windows in a new video interview, stating the next version will be "more ambient, pervasive, and multi-modal" as AI redefines user interaction with computers. Pavan Davuluri, Microsoft CVP and Windows boss, recently outlined a vision for the future of the Windows operating system in a newly released video interview. Davuluri detailed how artificial intelligence (AI) will fundamentally alter the desktop interface, leading to an ambient and multi-modal computing experience. This discussion provided insights into Microsoft's strategic direction for its flagship platform. During the interview, Davuluri addressed the transformative impact of AI on human-computer interaction. He stated, "I think we will see computing become more ambient, more pervasive, continue to span form factors, and certainly become more multi-modal in the arc of time." Davuluri further elaborated on the evolving nature of input methods, noting, "I think experience diversity is the next space where we will continue to see voice becoming more important. Fundamentally, the concept that your computer can actually look at your screen and is context aware is going to become an important modality for us going forward." This indicates a shift towards systems that understand environmental and user context through advanced AI capabilities. Microsoft has previously indicated a strategic emphasis on voice as a primary input method for future Windows iterations. A "Windows 2030 Vision" video, released a week prior to Davuluri's interview, featured Microsoft's CVP of Enterprise & Security discussing similar concepts regarding the future of the operating system. This earlier communication aligns with Davuluri's recent statements, reinforcing the company's commitment to integrating voice as a core interaction modality. The forthcoming version of Windows is expected to elevate voice to a first-class input method, complementing traditional mouse and keyboard interfaces. Users will reportedly be able to interact with Windows using natural language, with the operating system designed to comprehend user intent based on the content displayed on the screen. This integration aims to create a more intuitive and seamless user experience, allowing for ambient communication with the OS. Davuluri also suggested that the visual appearance and interaction paradigm of Windows are subject to significant change due to the integration of agentic AI. He observed, "I think what human interfaces look like today and what they look like in five years from now is one big area of thrust for us that Windows continues to evolve. The operating system is increasingly agentic and multi-modal." This statement highlights a substantial investment and ongoing development effort in evolving the operating system's core design and user interaction models. The cloud infrastructure is identified as a critical enabler for these advanced Windows experiences. Davuluri explained, "Compute will become pervasive, as in Windows experiences are going to use a combination of capabilities that are local and that are in the cloud. I think it's our responsibility to make sure they're seamless to our customers." This indicates a hybrid computing model where local device capabilities are seamlessly integrated with cloud-based processing to deliver enhanced functionality. Microsoft's strategic direction suggests a fundamental shift in how AI assistants are integrated into operating systems. Current AI assistants, such as Copilot on Windows, Gemini on Android, or Siri on macOS, typically function as applications or overlays operating on top of existing OS platforms. Microsoft appears to be preparing to introduce an operating system where AI is intrinsically woven into its foundational architecture, rather than existing as a separate layer. This transformation is anticipated within the next five years, potentially coinciding with the release of Windows 12. Multiple high-level Microsoft executives have alluded to this significant evolution, framing it as a major shift in computing driven by AI advancements. While the concept of voice becoming a primary input method for PCs may present an adjustment for some users, the integration of agentic AI and the OS's ability to comprehend user intent and natural language are expected to make this interaction feel more intuitive. This evolution extends beyond Microsoft, with Apple reportedly developing a similar voice-centric feature for iOS 26. This upcoming feature for iPhone users is rumored to enable application navigation solely through verbal commands, articulating user intent directly to the device. On the Windows platform, voice input is likely to augment, rather than replace, established input methods. The system will incorporate three primary modes of interaction: typing, touch/mouse, and voice. While voice input may not be mandatory for task completion, its availability is expected to streamline user workflows. However, the reliance on extensive personal user data to optimize these AI-driven experiences raises privacy considerations. Davuluri's acknowledgment of a necessary balance between local and cloud compute for these experiences suggests that these privacy concerns will be a significant factor in their implementation and public reception.
[7]
AI is taking over your PC: Microsoft teases radical Windows redesign with smart features - here's what users can expect
Microsoft is planning a big update for Windows with AI features. The new Windows will understand what you do on your PC and let you use voice to control it. It will work with keyboard, mouse, and cloud computing. The redesign may come in Windows 12 and will make computers smarter, easier, and more helpful. Microsoft released a new video interview with Pavan Davuluri, the head of Windows, to talk about the future of Windows. Davuluri was asked how AI will change computers, and he said that computing will become more ambient, everywhere, and multi-modal, meaning it will understand different types of input. He explained that voice will become more important for interacting with Windows in the future. The OS will be context-aware, meaning it can look at your screen and understand what you are doing to help you better. Users will be able to speak to their computer while writing, drawing, or interacting with others, and Windows will understand their intent, as reported by Windows Central. Microsoft has already hinted at voice-first features in Windows, including a "Windows 2030 Vision" video by another executive, showing a focus on natural language input. The future of Windows will promote voice as a primary input, along with mouse and keyboard, allowing users to talk naturally to their PC while working. Davuluri teased that Windows will look different from today, because agentic AI will be deeply built into the system. The OS will be able to act and respond intelligently, not just react to commands. The OS will use a mix of local computing and cloud computing to make AI features seamless and smooth, as per the report by Windows Central. ALSO READ: Google and Meta no longer cool? 75% of grads are walking away from Big Tech dreams for new career paths Today, AI assistants like Copilot on Windows, Gemini on Android, or Siri on Mac exist mostly as apps or floating windows, but Microsoft plans to build AI directly into the OS. This AI-powered Windows may appear in the next five years, possibly as Windows 12, bringing a big shift in PC interaction. Users may find it strange at first to use voice as a main input method, but with agentic AI understanding natural language and intent, it will feel natural and helpful. Apple is also planning voice-centered features in iOS 26, showing that tech companies are moving toward voice-first experiences, according to the report by Windows Central. On Windows, there will likely be three main input methods: typing, touch/mouse, and voice. You won't have to use voice, but it can make tasks easier. Privacy concerns are expected because AI features need a lot of personal data. Balancing cloud and local computing will be important, and some users may push back, as per the report. Microsoft emphasizes that the goal is to make AI seamless and helpful, changing how people interact with computers over the next few years. Overall, Windows is preparing a major AI-driven redesign that will make computers smarter, voice-friendly, and context-aware, changing the way we work and interact with PCs, reported by Windows Central. Q1: How will AI change Windows in the future? AI will make Windows smarter, voice-friendly, and context-aware, letting your PC understand what you are doing and respond naturally. Q2: Will Windows 12 use voice as a main input? Yes, voice will become a primary input method alongside keyboard and mouse, making tasks easier with AI understanding your intent.
Share
Share
Copy Link
Microsoft executives reveal plans for a future Windows OS that heavily integrates AI, voice control, and context awareness, promising enhanced user interaction and productivity.
Microsoft is charting an ambitious course for the future of Windows, with artificial intelligence (AI) at its core. Recent interviews with key executives, including Pavan Davuluri, Corporate VP of Windows and Devices, have shed light on the company's plans to revolutionize how users interact with their computers
1
.Source: Economic Times
The next generation of Windows is set to embrace what Microsoft calls "multimodal" interactions. This approach will expand beyond traditional mouse and keyboard inputs to include voice commands, vision-based interactions, and enhanced touch and pen capabilities
2
. Davuluri envisions a future where users can seamlessly "speak to your computer while you're writing, inking, or interacting with another person"3
.A key feature of this AI-driven Windows will be its ability to understand context. Microsoft aims to develop an operating system that can "look at your screen and is context aware," allowing it to semantically understand user intent and respond accordingly
4
. This contextual awareness is expected to enhance productivity by anticipating user needs and streamlining tasks.To power these advanced features, Microsoft plans to leverage a combination of cloud-based and on-device AI models. This hybrid approach aims to provide seamless experiences while addressing privacy concerns. Davuluri emphasized that some AI capabilities will run locally on devices, potentially operating without an internet connection to protect user privacy
1
.The integration of AI is expected to significantly improve Windows Search, making it "much more meaningful, much more impactful"
5
. Microsoft is developing semantic indexing capabilities that will better understand the content being searched, providing more accurate and relevant results.Source: pcgamer
Related Stories
While some of these features are already being introduced through Microsoft's Copilot AI in Windows 11, the full realization of this vision is expected within the next five years. This could potentially manifest as Windows 12 or even a completely new OS, unofficially dubbed "Windows Chat" by some observers
5
.Source: TechRadar
The ambitious plans have raised privacy concerns among some users. The idea of an operating system that constantly watches screen activity and processes data in the cloud has sparked debates about data security and user privacy. Microsoft has responded by emphasizing user control and the development of on-device AI models to mitigate these concerns
4
.As Microsoft continues to develop these technologies, the balance between enhanced functionality and user privacy will likely remain a central topic of discussion in the evolution of Windows.
Summarized by
Navi
[2]
1
Business and Economy
2
Technology
3
Technology