2 Sources
2 Sources
[1]
The Star Trek Holodeck just got closer - Apple's new AI tool generates 3D scenes from your photos in under a second for VR memories
Remember when Apple introduced spatial Lock Screen photos in iOS 26? This feature added a stereoscopic effect to flat images on your Lock Screen, and it's neat, if a little gimmicky. Now, though, Apple has revealed a new trick that takes the effect to an entirely new level - and it could be a brilliant addition to your Apple device if it gets a wider roll-out. That new tool is called SHARP, and it's just been unveiled in a research paper published by Apple. Titled "Sharp Monocular View Synthesis in Less Than a Second," the paper outlines a new tool that can turn 2D images into 3D spatial scenes in under one second. SHARP uses a neural network - artificial intelligence (AI), in other words - to quickly generate a 3D map of your image. That's the part that is performed in less than a second. Once that's complete, the image can be rendered in real time. Apple says it trained the model on around eight million synthetic images created in-house and 2.65 million licensed photographs, with the result that SHARP could learn to discern depth and scale and apply that knowledge to input images. It does this while maintaining consistency when it comes to aspects like scale and distance, meaning you shouldn't see the sorts of stretching and warping that can come about when performing a 2D-to-3D conversion. That's key to maintaining immersion and producing a 3D image that users actually want to keep. Right now, SHARP is more a proof-of-concept than a prime-time feature, and there's no indication of when - or if - it will come to Apple's devices. While it's available to download on GitHub, it's not yet baked into the likes of iOS 26 or macOS Tahoe. That said, it seems like a natural evolution of the spatial photos feature that Apple has already released. If the Photos app on your iPhone will let you explore your images in this way, it could be an attractive selling point for a lot of people. Add it to the Vision Pro headset and it'll be even more immersive. All that said, SHARP does have some drawbacks. For one thing, it's focused on rendering nearby scenes, meaning you can't stray too far from the original viewpoint before fidelity starts to suffer. But as a starting point, it's certainly promising, especially when you consider how quickly it can operate. Don't be surprised to see it roll out in Apple's operating systems at some point in the future.
[2]
You Can Now Try Apple's New AI Model That Creates A 3D Scene From A Single Image In Under A Second
Apple is continuing to increase its AI capabilities at a fairly rapid pace. As a case in point, consider the Cupertino giant's latest AI model that is able to create an entire 3D scene from a single 2D image, and that too in under a second. Apple has now published a study, titled "Sharp Monocular View Synthesis in Less Than a Second." The study details how Apple's engineers were able to train an AI model, called SHARP, to generate a "photorealistic" 3D view from a single 2D image. Critically, Apple claims that the view generation takes "less than a second on a standard GPU via a single feedforward pass through a neural network." Basically, SHARP predicts what a 3D scene, distilled from a 2D image, would look like by taking into account the image's "nearby viewpoints." The study notes: "The 3D Gaussian representation produced by SHARP can then be rendered in real time, yielding high-resolution photorealistic images for nearby views. The representation is metric, with absolute scale, supporting metric camera movements." For the benefit of those who might not be aware, 3D Gaussian Splatting is a technique that is used to create photorealistic 3D scenes by representing such scenes as millions of "splats," which are basically tiny colored blobs. To create a full scene, however, often requires numerous 2D images from various angles. Apple's SHARP differs in that it is able to recreate an entire photorealistic scene from a single 2D image by predicting depth and colors, and too in under a second. What's more, you can now try Apple's SHARP AI model for free by heading over to the dedicated GitHub page.
Share
Share
Copy Link
Apple has published research on SHARP, an AI model that generates 3D scenes from photos in under a second using neural networks. Trained on 8 million synthetic images and 2.65 million licensed photographs, SHARP converts 2D photographs into 3D spatial scenes while maintaining scale and distance consistency. The model is now available on GitHub.
Apple has unveiled SHARP, a new AI model that creates a 3D scene from a single image in less than a second, marking a significant advancement in the company's spatial computing capabilities
1
. The technology, detailed in a research paper titled "Sharp Monocular View Synthesis in Less Than a Second," demonstrates how Apple AI can convert 2D photographs into 3D spatial scenes with photorealistic quality2
. The model represents a natural evolution of Apple's spatial photo features, building on the stereoscopic effects introduced in iOS 26's Lock Screen photos.SHARP leverages neural networks to generate a 3D map of input images in under one second on a standard GPU via a single feedforward pass
2
. Apple trained the model on approximately 8 million synthetic images created in-house and 2.65 million licensed photographs, enabling it to discern depth perception and scale accurately1
. Once the initial 3D map is complete, the image can be rendered in real time, opening possibilities for VR memories and immersive experiences.
Source: TechRadar
The model predicts what a 3D scene would look like by taking into account the image's nearby viewpoints, maintaining consistency in scale and distance to avoid the stretching and warping common in 2D-to-3D conversions
1
.The technology employs a 3D Gaussian representation that generates photorealistic images for nearby views
2
. Unlike traditional 3D Gaussian Splatting techniques that require numerous 2D images from various angles, SHARP accomplishes monocular view synthesis from just one photograph by predicting depth and colors2
. The representation is metric, with absolute scale, supporting metric camera movements that maintain immersion2
. This approach addresses a key challenge in neural rendering: producing 3D images that users actually want to keep and explore.Related Stories
While SHARP currently exists as a proof-of-concept available on GitHub, it signals Apple's direction in spatial computing
1
. The model isn't yet integrated into iOS 26 or macOS Tahoe, but it appears positioned as a natural extension of existing spatial photo features1
. If implemented in the Photos app on iPhone, SHARP could become an attractive selling point for consumers interested in exploring their images in three dimensions. Integration with the Vision Pro headset would deliver even more immersive experiences, potentially transforming how users interact with their photo libraries1
.SHARP does face constraints in its current form, primarily its focus on rendering nearby scenes
1
. Users cannot stray too far from the original viewpoint before fidelity begins to degrade, limiting the extent of exploration within generated scenes. However, as a starting point, the technology demonstrates promising capabilities, especially given its rapid processing speed. The Cupertino giant continues to expand its AI capabilities at a rapid pace, and SHARP represents another step in that trajectory2
. Observers should watch for potential announcements regarding SHARP's integration into Apple's operating systems, as the technology matures from research to consumer-ready features.Summarized by
Navi
14 May 2025•Technology

07 Oct 2024•Technology

03 Dec 2024•Technology
