2 Sources
2 Sources
[1]
EXCLUSIVE: Luma launches creative AI agents powered by its new 'Unified Intelligence' models | TechCrunch
AI video generation startup Luma on Thursday launched Luma Agents, designed to handle end-to-end creative work across text, image, video, and audio. Luma Agents are powered by the startup's Unified Intelligence family of models, with architecture trained on a single multimodal reasoning system. Luma Agents are being pitched as a new way of doing work for ad agencies, marketing teams, design studios, and enterprises. Luma says its agents are capable of planning and generating text, image, video and audio while coordinating with other AI models, including Luma's Ray 3.14, Google's Veo 3 and Nano Banana Pro, ByteDance's Seedream, and ElevenLabs's voice models. Luma's agents are built on the startup's Uni-1 model, the first of its Unified Intelligence family of AI models. It has been trained on audio, video, image, language, and spatial reasoning, according to Amit Jain, chief executive officer and co-founder of Luma. Jain told TechCrunch that the Uni-1 model can "think in language and imagine and render in pixels or images...we call it 'intelligence in pixels.'" Other output capabilities like audio and video will come in subsequent model releases, he added. "Our customers aren't buying the tool, they're redoing how business is done," Jain said. Luma has already started rolling out its new agentic platform with existing customers including global ad agencies Publicis Groupe and Serviceplan, as well as for brands like Adidas, Mazda, and Saudi AI company Humain. Jain said the Luma Agents are a gamechanger because they are able to maintain persistent context across assets, collaborators, and creative iterations. They can also evaluate and refine outputs, improving their own results through an iterative self-critique, according to Jain. This sort of check-your-work capability is what has made coding agents so useful, Jain said. "You need that ability to evaluate your work, fix it, and do that loop until the solution is good and accurate." Jain said the current workflow for using AI tools in creative environments doesn't have the same acceleration of benefits people in the creative industry expect from AI. Instead, it's more like: "Here are 100 models. Learn how to prompt them," he said. He said what makes Luma Agents different is that you don't need to prompt back and forth for each iteration on an image or idea -- the system instead generates large sets of variations and lets users steer the direction through conversation. "With Unified Intelligence, because these models understand in addition to being able to generate, we are able to build a system that is able to do this sort of end-to-end work," Jain said. Take, for instance, a human architect designing a building. As they draw the lines, they are creating an internal mental representation of the structure, light, spatial dynamics, and lived experience. This, Jain says, is the same principle upon which Unified Intelligence is built. Jain said the system could significantly speed up creative workflows. In a demonstration, Jain showed how a 200-word brief and an image of a product (a piece of lipstick) led the system to generate various ideas for locations, models, and color schemes for an ad campaign. In another example, Luma Agents turned a brand's $15 million, year-long ad campaign into multiple localized ads for different countries in 40 hours for under $20,000, passing the brand's internal quality controls and accuracy checks, Jain said. While Luma Agents is now publicly available via API, Jain said the startup plans to roll out access gradually to ensure users maintain reliable access and avoid workflow disruptions.
[2]
Luma's Amit Jain says fragmented AI tools impacting creative workflows, launches unified intelligence - The Economic Times
In an interview with ET, CEO Amit Jain explained that the industry's reliance on 'chaining' separate models for text, image, and video has created a memory problem where AI loses the creative thread, forcing teams to rebuild context at every step.Luma AI, a Palo Alto-based startup that turns prompts into videos, says it's time to fix the fragmented nature of today's AI tools which forget details and hinder creative work. Luma's Unified Intelligence architecture, launched on Thursday, claims to fix this with one multimodal AI brain that blends reasoning with content generation and can handle text, images, video, and more in a single system, without losing track between steps. Its first model, Uni-1, mixes words and visuals together so the AI can think through ideas and create them at the same time. The startup launched Ray3 in 2025 as the world's first video reasoning model, followed by Ray3.14, which creates videos, animations, and visuals. Last November, Luma raised $900 million in a round led by Humain, a Saudi-based public investment fund (PIF), with participation from existing investors Andreessen Horowitz, Amplify Partners, and Matrix Partners. Most AI models process only a fixed amount of prior conversation or data at a time, causing them to "forget" earlier details in long projects like coding marathons or storyboarding, requiring constant re-prompting. Big tech companies have repeatedly flagged AI's "memory problem" which leads to incoherent output, higher cost, and failed projects. Luma also launched Luma Agents, a new class of AI tools built on its new architecture, to enable creative work for agencies, marketing teams, studios, and enterprises. Jain said these agents have the full picture across text, images, video, and audio. They look for different creative ideas, check and improve their own work (like coding agents), and connect easily to other tools through APIs. The agents work on external models such as Veo 3, Sora 2, Kling 2.6, Nano Banana Pro, and ElevenLabs and help cut down production timelines. Jain narrated that using Luma's agents, it took two artists 40 hours to work on a campaign that would otherwise have taken 12-13 months and $12 million. Jain explained that this represents a fundamental shift for advertising, marketing, and media workflows since the agents enable end-to-end production. "Agents collaborate end to end... everything, from the cast, locations, the lighting, colours, shots, the scenes, the motion, and how the thing flows, rather than isolated tasks". Creative work has never lacked ambition but execution capacity, with teams wasting time orchestrating tools instead of focussing on taste, direction, and strategy, he said. The startup will provide full IP ownership for customers, automated content review, and traceability, positioning its platform as an operational infrastructure that helps creative teams scale. Luma is targeting a global workforce of 100 million creative professionals, including designers, architects, and filmmakers. With global ad spend poised to cross the $1 trillion mark this year, Jain feels the shift from isolated tools to unified agents is inevitable.
Share
Share
Copy Link
AI video generation startup Luma unveiled Luma Agents, powered by its new Unified Intelligence architecture, to handle end-to-end creative production across text, image, video, and audio. The system addresses AI's memory problem by maintaining persistent context, enabling one campaign to shrink from $15 million and 12 months to under $20,000 in 40 hours.

AI video generation startup Luma on Thursday launched Luma Agents, a new class of creative AI agents designed to handle end-to-end creative production across text, image, video, and audio
1
. The announcement marks a significant shift in how creative professionals approach AI-assisted work, targeting ad agencies, marketing teams, design studios, and enterprises struggling with fragmented AI tools that lose context and hinder productivity2
.Amit Jain, CEO and co-founder of Luma, explained that the industry's reliance on chaining separate models for different tasks has created a memory problem where AI loses the creative thread, forcing teams to rebuild context at every step
2
. The current approach, he noted, amounts to telling creative professionals: "Here are 100 models. Learn how to prompt them"1
.Luma Agents are powered by the startup's new Unified Intelligence family of models, built on a single multimodal reasoning system that maintains persistent context across assets, collaborators, and creative iterations
1
. The first model in this family, Uni-1, has been trained on audio, video, image, language, and spatial reasoning, creating what Jain calls a multimodal AI brain that can "think in language and imagine and render in pixels or images"1
.This Unified Intelligence architecture addresses a fundamental flaw in existing AI systems. Most AI models process only a fixed amount of prior conversation or data at a time, causing them to "forget" earlier details in long projects like coding marathons or storyboarding, requiring constant re-prompting
2
. By blending reasoning with content generation in a single system, Luma's approach ensures the AI maintains the full picture across text, images, video, and audio without losing track between steps2
.What sets Luma Agents apart is their ability to evaluate and refine outputs through iterative self-critique, improving their own results without constant human intervention
1
. Jain compared this to coding agents, noting that "you need that ability to evaluate your work, fix it, and do that loop until the solution is good and accurate"1
.The agents can coordinate with other AI models, including Luma's Ray 3.14, Google's Veo 3, Nano Banana Pro, ByteDance's Seedream, and ElevenLabs's voice models
1
. Rather than requiring users to prompt back and forth for each iteration, the system generates large sets of variations and lets users steer the direction through conversation1
.Related Stories
The impact on creative workflows appears substantial. In one example, Luma Agents turned a brand's $15 million, year-long ad campaign into multiple localized ads for different countries in 40 hours for under $20,000, passing the brand's internal quality controls and accuracy checks
1
. Jain also shared that using Luma's agents, it took two artists 40 hours to work on a campaign that would otherwise have taken 12-13 months and $12 million2
.In a demonstration, Jain showed how a 200-word brief and an image of a product—a piece of lipstick—led the system to generate various ideas for locations, models, and color schemes for an ad campaign
1
. The agents handle everything from the cast, locations, lighting, colors, shots, scenes, motion, and how content flows, rather than isolated tasks2
.Luma has already started rolling out its new agentic platform with existing customers including global ad agencies Publicis Groupe and Serviceplan, as well as brands like Adidas, Mazda, and Saudi AI company Humain
1
. The startup will provide full IP ownership for customers, automated content review, and traceability, positioning its platform as operational infrastructure that helps creative teams scale2
.Luma is targeting a global workforce of 100 million creative professionals, including designers, architects, and filmmakers
2
. With global ad spend poised to cross the $1 trillion mark this year, Jain believes the shift from isolated tools to unified agents is inevitable2
. "Our customers aren't buying the tool, they're redoing how business is done," Jain said1
.While Luma Agents is now publicly available via API, Jain said the startup plans to roll out access gradually to ensure users maintain reliable access and avoid workflow disruptions
1
. Last November, Luma raised $900 million in a round led by Humain, a Saudi-based public investment fund, with participation from existing investors Andreessen Horowitz, Amplify Partners, and Matrix Partners2
.Summarized by
Navi
[1]
26 Nov 2024•Technology

18 Dec 2025•Technology

18 Sept 2025•Technology

1
Technology

2
Policy and Regulation

3
Policy and Regulation
