8 Sources
8 Sources
[1]
Adobe launches AI assistants for Express and Photoshop | TechCrunch
Adobe today released new AI assistants Creative Cloud products, Express and Photoshop, that can help users with image creation and editing. Most companies like to put AI assistants into a sidebar in their products to grab on-screen context. But for Express, Adobe has created a new mode that lets you use text prompts to create new images and designs. You can switch on the assistant mode to use AI prompts and then switch back to use editing tools and controls that are available in Express in the current version. Meanwhile, the new Photoshop assistant, which is in closed beta testing, lives in the sidebar. Adobe claims the assistant can understand different layers and help you automatically select objects and create masks. Adobe added that users can ask the assistant to complete repetitive tasks such as removing backgrounds or changing colors. Adobe's VP of generative AI, Alexandru Costin, told TechCrunch that the company decided to build a different mode for its AI assistant in Express to target students and professionals who use the app. The company wants to see if users can achieve what they want to do without having to switch back to the traditional interface, he explained. "We think this approach of switching between two modes, where you get the best of both worlds, is gonna make the technology both accessible and controllable," Costin said. The company said it is also experimenting with a new type of assistant that can coordinate with different assistants from other Adobe tools, and connect to a creator's social channels to better understand their style. Adobe said the product is in the early stages of development and is in private beta. The company said it is also exploring a way to connect Adobe Express with ChatGPT using OpenAI's app integrations API to let users create designs directly in ChatGPT. Adobe also announced a new set of AI features for its Creative Cloud apps. Photoshop users can now choose third-party models, like Google's Gemini 2.5 flash and Black Forest Labs' FLUX.1 Kontext, for the generative fill feature, which can remove objects or extend images. The company is also adding an AI-powered object mask in its video editing app, Premiere Pro, to let users easily identify and select objects or people to add effects or adjust colors.
[2]
I Got an Early Look at Adobe's Research Projects: How AI Makes It Quicker Than Ever to Edit Photos
Every year at Adobe Max, the software giant's annual creative conference, it shows off a number of research projects it has in development. Sneaks, as the event is known, is a great way to peek behind the vast Adobe curtain and understand where the tech giant might be headed in the future. At this year's event in Los Angeles, I got an early, exclusive first look at three of the photography projects. This year's crop of photography-focused presentations each tackle a task that could be handled with manual editing. But instead of spending hours relighting photos, removing distractors or animating 2D objects, these new tools make edits with a single click or two. They're impressive feats, and they're possible thanks to Adobe's Firefly AI models running in the background. Trace Erase is one of these projects. The new tool builds on Adobe's generative remove technology that acts like an AI-powered eraser. In Photoshop, you can use generative remove to clean up images with photobombers, distracting wires, and cables and other errant objects impeding your shot. But it can take multiple steps to completely remove an object from an image, as Lingzhi Zhang, Adobe research scientist, showed me in a demo. There are often other signs that something has been removed, like a reflection in a window. Trace Erase takes care of all those follow-up edits and, like the name implies, removes every trace of an object's existence. It's more than just removing reflections and smoothing out what remains. The AI has a wide knowledge base, so if you ask it to remove a jet ski, it will remove the entire object and the waves it created in its wake. If generative remove is an AI eraser, Trace Erase is an AI obliterator. Another photography Sneak is called Light Touch. With the tool, you can completely transform how a photo is lit. You can literally use it to change the lighting -- Adobe research scientist Zhixin Shu showed me how it can be used on an image of a lamp to turn on its light and adjust how bright it is, with the shadows aligning with the environment automatically. You can also diffuse light to adjust the prominence of shadows. Spatial lighting, the third method in the project, lets you place the source of light within an image. You can make the light appear to be coming from any point in the photo, including from within existing objects, like a lit carved pumpkin. It's an incredibly powerful tool that can transform any image simply by changing the lighting condition -- something that's very hard to do manually without having to reshoot. Turn Style is the third Sneaks photography project. Its name is very similar to the turnstiles you see at events, which makes sense for an AI-powered tool that takes 2D objects and brings them into 3D. Adobe research scientist Zhiqin Chen showed me how the 3D Render button can transform a 2D image into a fully rendered 3D object, which you can turn fully around, all 360 degrees. It maintains character consistency, so you don't have to worry about having many AI hallucinations. While AI helps to turn it into a 3D object, you still have control over what degree it's placed and what angle is visible. Some previous Sneaks research projects have eventually ended up in the hands of users, like last year's Photoshop harmonize AI tool, which is now generally available. But there's never a guarantee that will be the case for all of them. What this year's projects show is that Adobe is focusing its AI efforts for photography and photo editing on more practical use cases. "In Photoshop, it's not about just clicking a button for art. It's [about] giving you all the tools to make it easy to achieve whatever you want to achieve," said Stephen Nielsen, senior director for product management for Photoshop.
[3]
Photoshop's biggest AI update yet just dropped - how to try all the new tools
Adobe is continuing its drive into AI with the launch of several new AI-powered features for Photoshop -- and one is potentially a huge time-saver. At this week's Adobe Max conference in Los Angeles, the company unveiled multiple new initiatives, including AI assistants in Express, Firefly, and Photoshop. Also: The most popular AI image and video generator might surprise you - here's why people use it But beyond a new assistant, Photoshop is getting a serious injection of AI with the addition of tools that make editing easier and faster. Here's a look at what's new. Powered by agentic AI, Photoshop's new AI assistant is designed to free up time by automating repetitive tasks, such as resizing a batch of photos or applying the same edits, while providing personalized recommendations and tips that maintain creative control. Also: Why open source may not survive the rise of generative AI You can chat with Photoshop to make edits, get tips on how to make a photo look better, and even get help on what to do next and how to do it. When it debuted in 2023, Generative Fill quickly became one of the most popular Photoshop features. Now, Adobe says, you can use partner AI models, such as Google Gemini 2.5 Flash and Black Forest Labs FLUX.1 Kontext, in addition to Firefly's image models, to edit your content with a prompt. Want to give older photos a glow-up? Using Topaz Labs' technology, you can now upscale low-resolution images to 4K. Matching the color and lighting of an object to a background layer can be a tricky task. However, Photoshop's AI-powered Harmonize feature automatically integrates objects or people into environments in realistic-looking ways by matching the light, color, and tone of both layers. Also: Did Google just give us the ultimate AI photo-editing tool? I tested it on the Pixel, and hard agree ZDNET's David Gewirtz tested this feature earlier this month, calling it "very, very good" and even "addictive." Isolating people and objects in video frames is significantly easier with Object Mask, an AI feature that assists with color grading, blurring, and adding special effects without manual rotoscoping. Finding the best images from a large batch can be one of the most time-consuming elements of photographing an event. Assisted Culling not only identifies the best shots, but you can filter results for levels of focus, different angles, and various degrees of sharpness. Also: Adobe might've just solved one of generative AI's biggest legal risks These features are available in public beta, except for AI Assistant, which Adobe says is "rolling through a private beta waitlist on Photoshop on the web." Get the morning's top stories in your inbox each day with our Tech Today newsletter.
[4]
More Creators Are Using AI, So Adobe's Adding AI Assistants to Photoshop and Express
Adobe is introducing chatbot-like conversational editing to its content editing programs. The new AI assistants in Photoshop and Express will have an emphasis on "conversational, agentic" experiences -- meaning you can ask the chatbot to make edits, and it can independently handle them. The tools are just a few of the updates the company announced on Tuesday at its annual Max creative conference. The AI assistants in Photoshop and Express are in beta now. For example, in Express, you can ask the AI to change the color of an object or remove an obtrusive element. While pro users might be comfortable making those edits manually, the AI assistant might be more appealing to its less experienced users and folks working under a time crunch. Also announced on Tuesday is Project Moonlight, a new platform in beta on Adobe's AI hub, Firefly. It's a new tool that hopes to act as a creative partner. With your permission, it uses your data from Adobe platforms and social media accounts to help you create content. For example, you can ask it to come up with 20 ideas for what to do with your newest Lightroom photos based on your most successful Instagram posts in the past. These AI efforts represent a range of what conversational editing can look like, Mike Polner, Adobe Firefly's vice president of product marketing for creators said in an interview. "One end of the spectrum is [to] type in a prompt and say, 'Make my hat blue.' That's very simplistic," said Polner. "With Project Moonlight, it can understand your context, explore and help you come up with new ideas and then help you analyze the content that you already have," Polner said. While these AI assistants won't be the "everything machines" that ChatGPT or Gemini claim to be, it's a stark sign that Adobe's AI revolution is marching onward. Its focus on agentic AI, like much of the AI industry, aims to persuade people to delegate tasks to AI. In a recent Adobe survey of 16,000 global creators, 86% said they use creative generative AI. Over three-fourths (80%) said Gen AI helped them create content they otherwise couldn't have made. The new stats align with a growing popularity of generative media tools, like AI image and video generators, with newer models like OpenAI's Sora and Google's nano banana going viral this year. Adobe's been all in on AI for a while. This year, Adobe introduced AI-first mobile apps for Photoshop, Firefly and a new video editor called Premiere. The professional photographers, designers and illustrators who are Adobe's primary customers haven't all been sold on Adobe's AI ambitions, raising concerns about AI's legality, energy use and ethics. Agentic AI assistants are just the tip of the iceberg of all the news Adobe dropped on Tuesday. For more, check out the new generative audio and music tools in Firefly.
[5]
Adobe's new Photoshop AI Assistant can automate repetitive tasks
Among the usual slew of AI enhancements to its Creative Cloud apps, Adobe has introduced a new Photoshop AI Assistant to help automate repetitive chores and provide personalized recommendations. At Adobe Max 2025, the company also introduced new tools for Photoshop, Premiere and Lightroom, while launching a new AI generative model and bringing in new third party models from Topaz and others. A key new feature in Photoshop and Express (Adobe's all-in-one design, photo, and video tool) is the AI Assistant that lets you can chat with in a conversational manner to gain "more control, power and potential time-savings," according to Adobe. With that, you can tell it to take on a series of creative tasks like color correction on resizing. You can easily switch between prompts with the agent and manual tools like sliders to adjust brightness and contrast. It can also provide personalized recommendations and offer tutorials on how to accomplish complex tasks. In a brief demo, Adobe showed that when you switch to Photoshop's "agentic" mode in those apps, it minimizes the usual complex interface and leaves you with a simple prompt-based UI. You can then type in the task you want to accomplish, and the agent will perform those steps automatically. You can then jump back into the full interface to fine tune the result by changing things like brightness or levels. Along with the AI Assistant, Adobe introduced a few other AI tools for Photoshop. Chief among those are new partner models for generative fill that lets you easily remove unwanted objects and fill in the hole left behind. Those include Google Gemini 2.5 f!ash, Black Forest Labs FLUX.1 Kontext and Adobe's latest Firefly Image Models. It also introduced Firefly Image Model 5, Adobe's most advanced image generation model yet. Photoshop also gains new Generative Upscale option that uses Topaz Lab's AI to upscale small, cropped and other low-resolution images into 4K with "realistic detail," Adobe says. Another feature, Harmonize, lets you place objects or people into different environments in a realistic manner, eliminating much of work necessary for such compositing. Harmonize also matches the light, color and tone of foreground objects and people to the background. Premiere, meanwhile, introduced a similar feature called AI Object Mask that performs automatic identification and isolation of people and objects in video, so they can be edited and tracked without any manual rotoscoping. The app also gains new rectangle, ellipse and pen masking in Premiere to make targeted adjustments, along with a fast vector mask for quicker tracking. Finally, LIghtroom is getting a new feature called Assisted Culling. It lets you quickly and easily identify the best images in a large photo collection, with the ability to filter for things like focus level, angles and degrees of sharpness. Photoshop's Generative Fill with Partner Models, Generative Upscale and Harmonize are now available to customers today. Premiere's AI Object Mask, Rectangle, Ellipse and Pen Masking and Fast Vector Mask, along with Lightroom's AI Assisted Culling, launch today in beta. Adobe's Photoshop AI Assistant, meanwhile, will be available through a private beta waitlist.
[6]
Adobe adds chatbots to Photoshop and other design apps
The big picture: Adobe has been working for years to show creators how generative AI can be a boon to their jobs rather than an existential threat. Driving the news: Adobe is using its annual Max conference for creative types to show off a host of AI projects, including the new AI assistants coming to Photoshop and Adobe Express that allow people to describe edits in their own words and automate repetitive tasks. * Adobe will preview how its tools work within chatbots, starting with Adobe Express in ChatGPT. It expects to support more of its apps as well as other chatbots over time. * The company says it's also expanding Firefly AI playground to move beyond the concept stage to include AI-driven video editing, soundtrack creation, and voice-over tools. Between the lines: Adobe's AI assistants can handle simple fixes like removing a background or broader stylistic requests such as "make this more tropical." Yes, but: The rollout is uneven. Adobe Express' assistant is in public beta; Photoshop's is in private testing. In Firefly, speech and soundtrack generation are public, while video tools remain private. The bottom line: Adobe's blending of design tools and chatbots hints at a new AI-powered creative workflow.
[7]
Adobe's new AI assistant in Express is conversational
At Adobe MAX 2025, Adobe unveiled a new AI Assistant built into Adobe Express that aims to make design as natural as talking. The tool, now in beta, lets users describe what they want -- "make it look more tropical," for instance and watch the platform adjust colors, images, and fonts in real time. The idea is to take the friction out of design, particularly for people who aren't trained designers. Instead of navigating layers of menus or templates, users can type or say what they want and let the AI handle the details. Adobe says the assistant can modify individual layers, swap out backgrounds, and suggest new design elements while keeping the rest of a project intact. The company also emphasized that the Express assistant draws on Adobe's creative and content intelligence essentially the company's long-trained sense of what makes a design cohesive or visually appealing. It uses a mix of Adobe's own Firefly models and third-party AI to recommend tools or generate elements automatically. What sets this apart from other AI-driven design tools is its conversational tone. It doesn't just execute commands but follows up with contextual prompts asking, for example, whether the user wants to change a font to match a new color scheme. For businesses, Adobe plans to extend this technology into its enterprise suite, offering brand-locked templates and collaboration tools so teams can create on-brand content more easily. Some early adopters, like dentsu and Lumen, have suggested it could help non-designers produce professional visuals without slowing down their workflows. For now, the AI Assistant is available in beta on desktop for Adobe Express Premium users, with wider access planned through Adobe's Firefly credit system.
[8]
Is Adobe's 'conversational editing' the future of design?
In news that will come as a surprise to nobody, AI was the main focus of the opening keynote at Adobe Max in Los Angeles this morning. From new generative Firefly models (including impressive custom options) to new soundtracking tools, impressive AI features have arrived to enhance pretty much every aspect of the creation process. Among the various announcements was a phrase Adobe kept coming back to: conversational editing. With the advent of new AI assistants for Photoshop, Adobe Express and more, it seems much more of the design experience is set to be driven by prompts and chatbots. But 'conversational editing' also hints at a more human side of designing with AI - which, as Adobe told Creative Bloq, is no accident. Photoshop's new AI Assistant, powered by agentic AI, lets creatives save time by instructing it to take on series of creative tasks or provide personalised recommendations. In terms of the latter functionality, Adobe described the assistant as a "second set of eyes", offering constructive feedback on a flyer design. But perhaps the biggest cheer of the morning was awarded to the assistant's ability to clean up messy or non-existent layer names. Meanwhile, Adobe Express's new AI Assistant lets users design personalised content simply by describing what they want. The non-destructive tool lets users generate edits on any layer, including fonts and images, keeping the rest of the design intact. This natural, human style of converrational editing was a "very deliberate" focus for Adobe, Govind Balakrishnan, senior vice president and general manager at Adobe Express, told Creative Bloq. "The fact that you can converse to create a design by describing what you want to do - that's where the magic is." And at the end of the keynote, Adobe revealed a sneak peek of a tool that could take the idea even further. Project Moonlight serves as an AI Assistant that works as a personal social media assistant, letting users upload images from Lightroom and generate ideas for social posts.
Share
Share
Copy Link
Adobe launches AI assistants for Photoshop and Express at Adobe Max 2025, introducing conversational editing capabilities and automated workflows. The company also debuts new AI-powered features including third-party model integration and advanced research projects.
Adobe has unveiled a comprehensive suite of AI-powered tools at Adobe Max 2025, headlined by new AI assistants for Photoshop and Express that promise to revolutionize creative workflows through conversational editing capabilities [1](https://techcrunch.com/2025/10/28/adobe-l anches-ai-assistants-for-express-and-photoshop/). The company's latest initiative represents a significant shift toward "agentic AI" that can independently handle complex creative tasks through simple text prompts.

Source: TechCrunch
The new AI assistants operate differently across Adobe's platforms. In Express, Adobe has created a dedicated mode that allows users to switch between AI-powered prompts and traditional editing tools, targeting students and professionals who need streamlined workflows
1
.
Source: Economic Times
Meanwhile, Photoshop's assistant, currently in closed beta testing, resides in the sidebar and can understand different layers, automatically select objects, create masks, and complete repetitive tasks like background removal and color changes
1
.
Source: Axios
Adobe has significantly expanded Photoshop's AI capabilities by introducing support for third-party models in its popular Generative Fill feature. Users can now choose from Google's Gemini 2.5 Flash and Black Forest Labs' FLUX.1 Kontext, in addition to Adobe's own Firefly models, for removing objects or extending images
3
. This integration marks a notable departure from Adobe's traditionally closed ecosystem approach.The company has also introduced several powerful new features, including Generative Upscale powered by Topaz Labs technology that can enhance low-resolution images to 4K quality
5
. The new Harmonize feature automatically matches lighting, color, and tone when compositing objects or people into different environments, significantly reducing manual editing time3
.At the annual Sneaks presentation, Adobe showcased three cutting-edge research projects that demonstrate the company's vision for AI-powered photography editing
2
. Trace Erase builds upon existing generative remove technology by completely eliminating all traces of an object, including reflections and environmental effects like waves from a removed jet ski.Light Touch allows users to completely transform photo lighting conditions, including adjusting brightness, diffusing shadows, and repositioning light sources within images
2
. The third project, Turn Style, converts 2D images into fully rendered 3D objects that can be rotated 360 degrees while maintaining character consistency.Related Stories
Adobe's AI push comes as the company reports that 86% of global creators now use generative AI, with 80% stating that AI helps them create content they otherwise couldn't produce
4
. This widespread adoption has driven Adobe's strategy to focus on practical, time-saving applications rather than purely artistic generation.The company is also developing Project Moonlight, a creative partner platform that analyzes user data from Adobe platforms and social media to provide personalized content suggestions
4
. Additionally, Adobe is exploring integration with OpenAI's ChatGPT through app integration APIs, potentially allowing design creation directly within ChatGPT1
.Summarized by
Navi
[2]
10 Apr 2025•Technology

26 Oct 2024•Technology

Yesterday•Technology
