15 Sources
[1]
Dell wants to be your one-stop shop for AI infrastructure
Michael Dell is pitching a "decentralized" future for artificial intelligence that his company's devices will make possible. "The future of AI will be decentralized, low-latency, and hyper-efficient," predicted the Dell Technologies founder, chairman, and CEO in his Dell World keynote, which you can watch on YouTube. "AI will follow the data, not the other way around," Dell said at Monday's kickoff of the company's four-day customer conference in Las Vegas. Also: The rise of AI PCs: How businesses are reshaping their tech to keep up Dell is betting that the complexity of deploying generative AI on-premise is driving companies to embrace a vendor with all of the parts, plus 24-hour-a-day service and support, including monitoring. On day two of the show, Dell chief operating officer Jeffrey Clarke noted that Dell's survey of enterprise customers shows 37% want an infrastructure vendor to "build their entire AI stack for them," adding, "We think Dell is becoming an enterprise's 'one-stop shop' for all AI infrastructure." Dell's new offerings include products meant for so-called edge computing, that is, inside customers' premises rather than in the cloud. For example, the Dell AI Factory is a managed service for AI on-premise, which Dell claims can be "up to 62% more cost-effective for inferencing LLMs on-premises than the public cloud." Dell brands one offering of its AI Factory with Nvidia to showcase the chip giant's offerings. That includes, most prominently, revamped PowerEdge servers, running as many as 256 Nvidia Blackwell Ultra GPU chips, and some configurations that run the Grace-Blackwell combination of CPU and GPU. Future versions of the PowerEdge servers will support the next versions of Nvidia CPU and GPU, Vera and Rubin, said Dell, without adding more detail. Dell also unveiled new networking switches running on either Nvidia's Spectrum-X networking silicon or Nvidia's InfiniBand technology. All of these parts, the PowerEdge servers and the network switches, conform to the standardized design that Nvidia has laid out as the Nvidia Enterprise AI factory. A second batch of updated PowerEdge machines will support AMD's competing GPU family, the Instinct MI350. Both PowerEdge flavors come in configurations with either air cooling or liquid cooling. Complementing the Factory servers and switches are data storage enhancements, including updates to the company's network-attached storage appliance, the PowerScale family, and the object-based storage system, ObjectScale. Dell introduced what it calls PowerScale Cybersecurity Suite, software designed to detect ransomware, and what Dell calls an "airgap vault" that keeps immutable backups separate from production data, to "ensure your critical data is isolated and safe." Also: Everything Google unveiled at I/O 2025: Gemini, AI Search, smart glasses, more The ObjectScale products gain support for remote data access (RDMA), for use with Amazon's S3 object storage service. The technology more than triples the throughput of data transfers, said Dell, lowers the latency of transfers by 80%, and can reduce the load on CPUs by 98%. "This is a game changer for faster AI deployments," the company claimed. "We'll leverage direct memory transfers to streamline data movement with minimal CPU involvement, making it ideal for scalable AI training and inference." Dell AI Factory also emphasizes the so-called AI PC, workstations tuned for running inference. That includes a new laptop running a Qualcomm circuit board, the AI 100 PC inference card. It is meant to make local predictions with Gen AI without having to go to a central server. The Dell Pro Max Plus laptop is "the world's first mobile workstation with an enterprise-grade discrete NPU," meaning a standalone chip for neural network processing, according to Dell's analysis of workstation makers. The Pro Max Plus is expected to be available later this year. A number of Dell software offerings were put forward to aid the idea of the decentralized, "disaggregated" AI infrastructure. For example, the company made an extensive pitch for its file management software, Project Lightning, which it calls "the world's fastest parallel file system per new testing," and which it said can achieve "up to two times greater throughput than competing parallel file systems." That's important for inference operations that must rapidly intake large amounts of data, the company noted. Also: The best laptops: Expert tested and reviewed Also in the software bucket is what Dell calls its Dell Private Cloud software, which is meant to move customers between different software offerings for running servers and storage, including Broadcom's VMware hypervisors, Nutanix's hyper-converged offering, and IBM Red Hat's competing offerings. The company claimed Dell Private Cloud's automation capabilities can allow customers to "provision a private cloud stack in 90% fewer steps than manual processes, delivering a cluster in just two and a half hours with no manual effort."
[2]
Dell wants to be your one-stop shop for enterprise AI infrastructure
Michael Dell is pitching a "decentralized" future for artificial intelligence that his company's devices will make possible. "The future of AI will be decentralized, low-latency, and hyper-efficient," predicted the Dell Technologies founder, chairman, and CEO in his Dell World keynote, which you can watch on YouTube. "AI will follow the data, not the other way around," Dell said at Monday's kickoff of the company's four-day customer conference in Las Vegas. Also: The rise of AI PCs: How businesses are reshaping their tech to keep up Dell is betting that the complexity of deploying generative AI on-premise is driving companies to embrace a vendor with all of the parts, plus 24-hour-a-day service and support, including monitoring. On day two of the show, Dell chief operating officer Jeffrey Clarke noted that Dell's survey of enterprise customers shows 37% want an infrastructure vendor to "build their entire AI stack for them," adding, "We think Dell is becoming an enterprise's 'one-stop shop' for all AI infrastructure." Dell's new offerings include products meant for so-called edge computing, that is, inside customers' premises rather than in the cloud. For example, the Dell AI Factory is a managed service for AI on-premise, which Dell claims can be "up to 62% more cost-effective for inferencing LLMs on-premises than the public cloud." Dell brands one offering of its AI Factory with Nvidia to showcase the chip giant's offerings. That includes, most prominently, revamped PowerEdge servers, running as many as 256 Nvidia Blackwell Ultra GPU chips, and some configurations that run the Grace-Blackwell combination of CPU and GPU. Future versions of the PowerEdge servers will support the next versions of Nvidia CPU and GPU, Vera and Rubin, said Dell, without adding more detail. Dell also unveiled new networking switches running on either Nvidia's Spectrum-X networking silicon or Nvidia's InfiniBand technology. All of these parts, the PowerEdge servers and the network switches, conform to the standardized design that Nvidia has laid out as the Nvidia Enterprise AI factory. A second batch of updated PowerEdge machines will support AMD's competing GPU family, the Instinct MI350. Both PowerEdge flavors come in configurations with either air cooling or liquid cooling. Complementing the Factory servers and switches are data storage enhancements, including updates to the company's network-attached storage appliance, the PowerScale family, and the object-based storage system, ObjectScale. Dell introduced what it calls PowerScale Cybersecurity Suite, software designed to detect ransomware, and what Dell calls an "airgap vault" that keeps immutable backups separate from production data, to "ensure your critical data is isolated and safe." Also: Everything Google unveiled at I/O 2025: Gemini, AI Search, smart glasses, more The ObjectScale products gain support for remote data access (RDMA), for use with Amazon's S3 object storage service. The technology more than triples the throughput of data transfers, said Dell, lowers the latency of transfers by 80%, and can reduce the load on CPUs by 98%. "This is a game changer for faster AI deployments," the company claimed. "We'll leverage direct memory transfers to streamline data movement with minimal CPU involvement, making it ideal for scalable AI training and inference." Dell AI Factory also emphasizes the so-called AI PC, workstations tuned for running inference. That includes a new laptop running a Qualcomm circuit board, the AI 100 PC inference card. It is meant to make local predictions with Gen AI without having to go to a central server. The Dell Pro Max Plus laptop is "the world's first mobile workstation with an enterprise-grade discrete NPU," meaning a standalone chip for neural network processing, according to Dell's analysis of workstation makers. The Pro Max Plus is expected to be available later this year. A number of Dell software offerings were put forward to aid the idea of the decentralized, "disaggregated" AI infrastructure. For example, the company made an extensive pitch for its file management software, Project Lightning, which it calls "the world's fastest parallel file system per new testing," and which it said can achieve "up to two times greater throughput than competing parallel file systems." That's important for inference operations that must rapidly intake large amounts of data, the company noted. Also: The best laptops: Expert tested and reviewed Also in the software bucket is what Dell calls its Dell Private Cloud software, which is meant to move customers between different software offerings for running servers and storage, including Broadcom's VMware hypervisors, Nutanix's hyper-converged offering, and IBM Red Hat's competing offerings. The company claimed Dell Private Cloud's automation capabilities can allow customers to "provision a private cloud stack in 90% fewer steps than manual processes, delivering a cluster in just two and a half hours with no manual effort."
[3]
Dell unveils new AI servers powered by Nvidia chips to boost enterprise adoption
May 19 (Reuters) - Dell Technologies (DELL.N), opens new tab on Monday unveiled new servers powered by Nvidia's (NVDA.O), opens new tab Blackwell Ultra chips, aiming to capitalize on the booming demand for artificial intelligence systems. The servers, available in both air-cooled and liquid-cooled variations, support up to 192 Nvidia Blackwell Ultra chips but can be customized to include as many as 256 chips. These servers can train AI models up to four times faster than previous models, Dell said. The pricing of these products will be "competitive," Arthur Lewis, President of Dell's Infrastructure Solutions Group, told Reuters, adding that "there's a lot of interest on what's next." Dell and Super Micro Computer (SMCI.O), opens new tab have benefited from the growing demand for servers designed to handle the computer-heavy AI tasks, but the high cost of producing the systems and tough competition have pressured their margins. Dell in February forecast a decline in adjusted gross margin rate for fiscal 2026, while Super Micro projected fourth-quarter revenue below estimates earlier this month as tariff-driven economic uncertainty pressures its performance. Dell will focus on increasing sales of networking and storage products to ensure the "right level of profitability," Lewis said. The company's new servers will also support Nvidia's upcoming Vera central processing units, which will succeed the chip designer's Grace server processor. The AI server maker plans to support Nvidia's Vera Rubin chips, set to follow the Blackwell series. Dell also introduced a 'Pro Max Plus' laptop designed for AI development on Monday, featuring a neural processing unit that allows engineers to process large AI models directly on the device without relying on cloud services. Reporting by Jaspreet Singh in Bengaluru; Editing by Shreya Biswas Our Standards: The Thomson Reuters Trust Principles., opens new tab Suggested Topics:Disrupted
[4]
Dell builds out its AI offerings with an emphasis on on-premises deployment - SiliconANGLE
Dell builds out its AI offerings with an emphasis on on-premises deployment Kicking off its Dell Technologies World conference this week, Dell is expanding its AI Factory artificial intelligence product portfolio with infrastructure, software, and services updates with the stated goal of helping enterprises move from experimentation to scaled deployment of AI without using commercial cloud services. Dell said on-premises infrastructure is more efficient for AI inference at scale, especially as enterprise customers become more sensitive to cloud costs, data sovereignty and operational flexibility. The company cited Enterprise Strategy Group Inc. research that found AI Factory can save up to 62% of the cost of inferencing large models compared to public cloud options. "We're seeing up to 75% lower total cost of ownership running [large language models] on-prem with Dell infrastructure compared to the public cloud," said Sam Grocott, senior vice president of product marketing at Dell. Today's announcements span the spectrum from portable inference at the edge to high-density training clusters in data centers. At the edge, the new Pro Max Plus laptop (pictured) includes 32 AI cores, 64 gigabytes of memory and a Qualcomm Inc. AI 100 PC inference card, making it the first mobile workstation with an enterprise-grade discrete neural processing unit, according to Dell. It's aimed at running large models that typically require cloud resources locally, with support for 100-billion-parameter-plus LLMs. In the data center, Dell introduced a new approach to AI-scale thermal management with a system designed to absorb 100% of server-generated heat using a self-contained airflow design. The PowerCool Enclosed Rear Door Heat Exchanger also operates with higher water temperatures of between 32°C and 36°C to reduce reliance on traditional chillers. Dell said the heat exchanger can cut cooling-related energy costs by up to 60% and allow up to 16% greater rack density without increased power consumption. Other features include leak detection, real-time thermal monitoring and unified rack-level management. New servers in the PowerEdge lineup - the PowerEdge XE9785 and XE9785 - support Advanced Micro Devices Inc.'s new Instinct MI350 graphic processing units with 288 gigabytes of High Bandwidth Memory 3E, a new generation of memory technology designed for high-performance computing applications. Dell said the new platforms deliver up to 35 times better inferencing performance compared to previous systems while also reducing cooling demands through liquid and air-cooled options. The announcements are part of a continuing Dell push to promote what it calls "disaggregated architecture," in which computing, storage, memory and networking are managed separately. "Disaggregated infrastructure is where the puck is going for our customers because it combines the flexibility of three-tier with the simplicity of hyperconverged to enable dynamic resource allocation from a shared resource," said Varun Chhabra, senior vice president of infrastructure and telecom marketing at Dell. As part of its partnership with Nvidia Corp., Dell announced a new generation of PowerEdge servers designed around Nvidia Blackwell Ultra GPUs. Designed for LLM training and inference, the servers scale up to 256 GPUs per rack in liquid-cooled configurations and can deliver up to four times faster model training compared to the previous generation, Dell said. A new high-density PowerEdge XE9712 server features the Nvidia GB300 NVL72 rack-scale server optimized for training and offering a 50-fold gain in inference output and five-fold throughput improvements. Another model, the XE7745, will support the Nvidia RTX Pro 6000 Server Edition starting in July 2025. It's aimed at robotics, digital twins and multi-modal workloads. New networking options include the Dell PowerSwitch SN5600, SN2201 Ethernet, part of the Nvidia Spectrum-X Ethernet networking platform and Nvidia Quantum-X800 InfiniBand switches. The switches deliver up to 800 gigabits per second of throughput and are backed by Dell ProSupport and Deployment Services. On the software front, Dell AI Factory is now validated to use Nvidia NeMo microservices and tools for retrieval-augmented generation and agentic workflows. The AMD partnership includes upgraded support for the Radeon Open Compute and Day 0 container models like Llama 4. Dell also extends Red Hat OpenShift support to the AI Factory for managing containerized AI workloads. Noting that models require reliable access to quality data, Dell announced updates to its ObjectScale object storage platform to support more compact configurations and integrate with Nvidia BlueField-3 and Spectrum-4 networking. New S3 over Remote Direct Memory Access support promises throughput performance gains of up to 230%, 80% lower latency and 98% less CPU overhead, all of which help improve GPU utilization. A new Dell reference architecture combines PowerScale storage, Project Lightning, PowerEdge servers, and Nvidia's NIXL libraries for large-scale LLM inferencing. Dell will also integrate support for Nvidia's AI Data Platform for agentic application development. "Internal testing shows that, when it rolls out later this year, the AI Data Platform will be the fastest parallel file system in the world," Chhabra said. "It will outpace traditional file systems by up to 2X and offer 67% faster data access compared to its nearest competitor."
[5]
Welcome to the AI factory era: A preview of Dell Technologies World 2025 - SiliconANGLE
Welcome to the AI factory era: A preview of Dell Technologies World 2025 In just a few years, computing has undergone a massive shift. What was once a marketplace dominated by general‑purpose servers and monolithic data centers has fractured into a complex ecosystem of specialized accelerators, hyper‑scaled clusters, edge‑enabled devices, large‑scale cloud providers and sovereign‑cloud platforms. At the center of this transformation stand two sides of the market growth, 1) Nvidia Corp., the incumbent kingpin of merchant graphics processing units, and 2) everyone else, the established semiconductor and infrastructure players. I'm often asked who will win? Will it be Nvidia or everyone else? The answer is both. The demand for AI applications is fueling an infrastructure renaissance not seen since the 1990s. Now it's a 100x value enablement that weaves together competitive narratives -- and sets the agenda for Dell Technologies World this week in Las Vegas. Let's explore the trends powering the "AI Factory Era," referring to the large‑scale systems that support training, reasoning and inference at unprecedented scale. I'll break down the following areas, each critical in this supercycle of transformation: AI workloads -- whether foundation‑scale training or real‑time reasoning -- are hungry beasts, devouring FLOPS, memory bandwidth, network interconnects and rack‑scale power. Nvidia's GPUs have fed that hunger for years, carving out over 90% share of AI system shipments. Their innovative GPUs, the CUDA ecosystem and smart software libraries have made them the default choice for Amazon Web Services Inc., Google LLC, Microsoft Corp. and Meta Platforms Inc. as they built ever‑more ambitious AI Factories. But this dominance carries risk. Hyperscalers routinely face "GPU rationing" -- quarterly quotas, long lead times and price volatility. They're at the mercy of shifting U.S. export controls and potential Chinese counter‑measures. For companies spending tens of billions on AI capex, the ability to "own their destiny" is a strategic imperative. Many, like AWS, are building -- or contemplating -- custom chips and system software. Others are rallying around XPUs: custom AI accelerators co‑designed with hyperscaler partners to speed critical kernels. These ASICs sit alongside Nvidia GPUs in standard PCIe slots, enabling mixed clusters that deliver comparable throughput at 20% to 30% lower capex. More importantly, they provide a credible second source of compute, giving buyers the bargaining power to secure GPU allocations more timely. As one industry executive said, "It's not just about saving money -- it's about controlling the future of your AI infrastructure." The rise of XPUs amplifies a deeper debate. Should AI factories be built on open, composable architectures or closed, tightly integrated stacks? Today's AI landscape swings between these extremes. Hyperscalers value open composability despite the integration burden, while enterprises crave turnkey AI solutions but worry about vendor lock‑in. POCs are bleeding into production, yet startups -- starved for clarity on where to add value -- still struggle to articulate services in a half‑open, half‑closed world. The vision of an AI factory extends beyond GPUs or XPUs. It spans training, reasoning, and inference -- each with unique workload profiles. It demands networking and storage architectures tailored to massive clusters and hardware‑aware orchestration software that places workloads optimally. The rise of multistep reasoning -- on‑the‑fly decision branches and chain‑of‑thought processing -- changes everything. A 10 million‑token Q&A can balloon to 100 million tokens once reasoning is enabled, multiplying network traffic and GPU cycles tenfold. Clusters must rethink interconnect design, buffer sizing and bursting performance to support these peaks. Inference workloads range from high‑throughput, low‑latency cloud requests to resource‑constrained real‑time edge deployments. Use cases -- agentic chatbots, generative recommendation engines, digital twins -- often demand sub‑10-millisecond response times. Network operators envision AI‑powered network slices in cell towers, while embedded devices -- from autonomous drones to industrial robots -- require specialized ASICs and chip‑plus‑FPGA hybrids. The optimal AI factory spans central cloud clusters for training, regional on‑premises clusters for inference, and edge micro‑clusters for low‑latency tasks. Hardware configurations -- GPU/XPU ratios, NIC speeds, NVMe topologies -- must adapt fluidly to workload demands. A broad mix of systems and technologies is essential, covering general‑purpose AI deployments and highly specialized vertical workloads. Although enterprise AI budgets have stalled, they're beginning to rebound. A breakout is expected in 2026 as organizations seek efficiency and ROI. Meanwhile, industrial and automotive sectors highlight supply‑chain risks and uncertainties that diversified portfolios mitigate. Vendors with broad portfolios of chips, software and services are better insulated against downturns. Not all vendors will win with a single SKU; diversification is key. Enterprises who are slower and sometimes hesitant to invest in AI at scale are quietly ramping up budgets. By 2026, corporate AI capex is expected to accelerate to high levels driven by: Early examples include financial services piloting real‑time credit scoring clusters, manufacturers embedding AI‑driven defect detection, and retailers personalizing shopping experiences with on‑prem inference. At Mobile World Congress 2025 in Barcelona, conversations on theCUBE with more than 30 enterprises and operators revealed a consensus: Sovereign‑cloud AI is real and urgent. National regulations mandate that sensitive data and model execution stay within domestic borders. Operators want turnkey, on‑country AI clouds that rival AWS, Azure and Google Cloud but remain under local control. They can deploy inference nodes within regulated borders, integrate them into national 5G cores, and offer AI‑as‑a‑service without offloading data to foreign clouds. Edge AI embeds intelligence in the physical world -- smart cameras that detect safety hazards, autonomous logistics robots and digital twins monitoring factory floors. New AI‑enabled chipsets will power these devices, blurring operational technology and information technology boundaries. By 2026, tens of millions of edge nodes will form a distributed AI environment that complements centralized cloud systems. Physical AI and robotics will become major vectors in AI infrastructure. Every month we are seeing major leaps in algorithmic efficiency, fueling demand via new reasoning workloads. Traditional Q&A might consume 10 million tokens; multistep chain‑of‑thought reasoning blasts that to 100 million tokens. Real‑time agents consulting knowledge graphs add more compute layers, driving ever‑larger clusters. This self‑reinforcing cycle cements AI factories' centrality in computing and rewards vendors with unified, end‑to‑end platforms. Vendors that master AI Factory solutions stand to reap windfalls. Hyperscalers committed over $300 billion in capex for 2025, two‑thirds of which will support AI compute and networking. Enterprises, now guided by chief AI officers, are poised to unlock new capex waves for AI‑native infrastructure. No discussion of global AI infrastructure is complete without U.S.‑China tensions. In April 2025, proposed semiconductor tariffs threatened 10% to 25% duties on key components. Manufacturers are shifting export/import duties onto customers and relocating footprints from China and Vietnam to Mexico and Canada. Agile logistics -- turning geopolitical risk into a competitive moat -- will cement customer trust in an uncertain world. As we enter the heart of the AI Factory era, several innovation curves demand attention: We stand at the dawn of a new computing era. AI factories will drive the next wave of productivity, automation, and innovation across every industry. Established and emerging players are racing to fill niches in the AI factory supply chain. For enterprises and hyperscalers alike, the message is clear: The accelerator, network, cloud and supply‑chain choices you make today will determine your competitive posture for the decade to come. The AI factory era starts now.
[6]
How AI factories are reshaping enterprise infrastructure - SiliconANGLE
Michael Dell on AI factories, edge innovation and the next phase of digital transformation AI factories are accelerating a new phase of enterprise transformation, turning data into intelligence at scale -- and few leaders see this shift more clearly than Michael Dell (pictured), chief executive officer of Dell Technologies Inc. As organizations race to operationalize AI, the notion of an "AI factory" has emerged -- where data enters and intelligence is produced at scale, reshaping everything from supply chains to customer engagement. The next phase will be defined by the convergence of large-scale reasoning systems and intelligent edge devices, designed to deliver real-time insights with unprecedented precision. The future is not just cloud or edge -- it's both, working in tandem to power the knowledge economy, according to Dell. "I think we're just at this enormous speed-up moment for our species, and it's the beginning," Dell said. "When I look at how companies are able to unlock their data and accelerate their progress in everything that they're doing, it's just kind of amazing how fast this is happening." Dell spoke with theCUBE's Dave Vellante and John Furrier at Dell Technologies World, during an exclusive broadcast on theCUBE, SiliconANGLE Media's livestreaming studio. They discussed AI factories, edge computing, enterprise adoption and the infrastructure needed to scale intelligent systems. (* Disclosure below.) Michael Dell sees AI factories not just as a metaphor, but as an operational reality for modern infrastructure. These are systems purpose-built to process tens of trillions of tokens per month, requiring massive compute, data and software coordination. As demand shifts from training to inference and reasoning, deployment precision becomes even more critical. "There was also this dirty little secret in the industry for a long time. What are we doing with all this data?" Dell said. "Well, now we have tools. We can do things with it. That's why all this is happening." At the same time, the edge is becoming smarter, more efficient and a critical extension of these centralized capabilities. Dell described a future where computing power is everywhere -- from laptops and cars to industrial sites and mobile devices -- bringing AI closer to where data is generated and decisions are made. "If we dial back the clock of computing 30 years, there's always these debates on, where is the power going to be? Where is the intelligence going to be? Is it going to be at the center? Is it going to be out there in the edge? What's the answer? It's both. It's always both." The exponential increase in data volumes makes real-time processing essential. Dell noted that AI tools are finally catching up to the data deluge, enabling insights and automation that were previously out of reach. With global infrastructure ramping, Dell Technologies is deploying multiple hyperscale systems to meet enterprise needs. "That example that I showed [during the keynote] of one of the large systems that we're building, we're actually building another one that is very similar in size. We're deploying these right now," Dell said. "What's really interesting is a single system like that ... would've been more tokens than the entire world would've produced not that long ago." As the AI economy matures, so does its ecosystem. Dell Technologies is positioning itself as a foundational partner to a wide range of innovators -- from hyperscalers to open-weight model providers to enterprise software firms. In just the past year, the company has launched around 100 new solutions, including reference architectures and turnkey blueprints to jump-start AI adoption. "We need a lot more of those in every vertical in order for this to realize its potential," Dell said. "But it's definitely happening, and we're seeing it." This diversity of use cases, partners and compute needs points toward an increasingly open and hybrid AI future. Eighty-five percent of enterprises plan to move AI workloads on-premises within two years, Dell pointed out. That demand is being driven by control, cost, performance and regulatory needs, especially as organizations look to bring AI into the core of business operations. "Some of them are new partners, and some are old friends," Dell said. "The global economy, roughly half of it is a knowledge economy. That's roughly 30 or 40 trillion -- nobody's really quite sure. So, the idea of investing a trillion dollars to make the 30 trillion more competitive, more productive, that's not actually very much." Ultimately, the impact of AI must be measured not just in compute cycles, but in the value it unlocks for businesses and communities, Dell emphasized. The goal is to use technology to push human progress forward. "Technology enables human progress. That's kind of what we started with. We're still on that journey, and I passionately believe that that's what technology does," Dell added. "The progress that's occurring in our world is pretty staggering, and it's all enabled by technology. What more fun could there be that being in the center of all that? Stay tuned for the complete video interview, part of SiliconANGLE's and theCUBE's coverage of Dell Technologies World.
[7]
Dell refactors AI operations for the industrial revolution - SiliconANGLE
Dell builds out the AI infrastructure engine to power the incoming industrial revolution Artificial intelligence is not just transforming workflows -- it's igniting the next industrial revolution. As enterprises race to adopt AI at scale, the pressure is on to rethink how infrastructure is built, optimized and deployed for a new era of intelligent systems. For that revolution to happen, however, AI approaches need to be reworked to scale new heights in performance and efficiency. By implementing capabilities such as global supply chain agility, cutting-edge custom clusters and real-world enterprise AI applications, Dell Technologies Inc. is staying ahead of the development curve, according to Jeff Clarke (pictured), chief operating officer and vice chairman of Dell Technologies. "To be honest, we retooled our engineering capability," Clarke said. "With AI and this race for technology and deployment, building out clusters that are training these foundational models, we had to go back and reinvent ourselves. We had to go back and rethink engineering workflows because the rate of technology advancement no longer fit into the way that we traditionally built data center products." Clarke spoke with theCUBE's John Furrier and Dave Vellante at Dell Technologies World, during an exclusive broadcast on theCUBE, SiliconANGLE Media's livestreaming studio. They discussed how Dell Technologies is betting big on AI and backing it with real engineering muscle. (* Disclosure below.) Dell has fundamentally overhauled its engineering processes to meet the demands of AI workloads. Gone are the days of conventional data center products -- Dell now builds AI-specific architectures that scale rapidly, according to Clarke. The company's new approach enables them to go from customer collaboration to large-scale GPU cluster deployment in under six months. This agility has led to breakthroughs such as being first to market with Nvidia's GB200 NVL72 and Hopper platforms. "We've committed to engineering to solve these massive solutions problems, because we're not building a single computer; we're hooking up lots of computers to act as a single computer and to interact with one another," Clarke said. "It's remarkable, these systems, and what they're capable of." Being a custom design house, Dell is, in essence, building some of the largest AI training clusters in the world -- and then taking those learnings and scaling them for enterprises of all sizes. The scale of AI's computational needs for the future is staggering. Estimates project a leap from 25 trillion tokens in 2024 to 35 trillion by 2028, according to Clarke. The sharp uptick is driven by the rise of agentic systems and reasoning engines that multiply demand for compute and data infrastructure. "The truth is most data is created out in the wild, out in a smart factory, out in a smart hospital, out in a smart city, in your sneakers," Clarke said. "That's where data is actually generated. What we see and continue to believe is that the AI migrates to where the data is created to be dealt with." This unprecedented surge in data generation is changing the very fabric of enterprise computing. To absorb the data demands from the unfolding industrial revolution, Dell is building what it calls "token factories," which are hyper-efficient, high-performance compute clusters capable of processing this deluge of data, according to Clarke. Here's the complete video interview, part of SiliconANGLE's and theCUBE's coverage of Dell Technologies World:
[8]
Dell AI Factory 2.0 with Nvidia supports agentic AI, semantic queries - SiliconANGLE
Dell and Nvidia level up Dell AI Factory for agentic use cases Dell Technologies Inc. and Nvidia Corp. aim to stay ahead of the curve on artificial intelligence infrastructure by releasing the latest version of Dell AI Factory with Nvidia, described as "AI in a box." Since the product was announced at Nvidia's 2024 GTC conference, Dell and Nvidia have collaborated on numerous updates, leading to the 2.0 version Dell announced this week. The upgrades are focused on supporting agentic AI and improving networking compatibility between the two companies' technologies, according to Varun Chhabra (pictured, right), senior vice president of product marketing, Infrastructure Solutions Group and Telecom, at Dell. "With 2.0, there is a big focus on obviously refreshing the latest infrastructure based on all the innovation that Nvidia is bringing in," Chhabra said. "We also have a greater and expanded relationship with Nvidia on the networking side, where Nvidia networking is now available to Dell for customers. Compute, storage and networking ... need support for their AI Factory with Nvidia. They are able to get that through us." Chhabra and Kari Briski (left), vice president of generative AI software for enterprise at Nvidia, spoke with theCUBE's Dave Vellante and John Furrier at Dell Technologies World, during an exclusive broadcast on theCUBE, SiliconANGLE Media's livestreaming studio. They discussed what's new about Dell AI Factory 2.0 and what is next for machine learning. (* Disclosure below.) Since generative artificial intelligence burst onto the scene, the technology has evolved rapidly. Now, AI agents and enterprise-scale large language models are at the center of the conversation, alongside concerns over how to manage the data centers that sustain them. Dell AI Factory manages the full AI lifecycle -- from the data center to the user-specific application -- to applying the best tools to the task, according to Briski. "Enterprises are moving from [central processing unit]-powered enterprise operations to [graphics processing unit]-enterprise operations ... because of LLMs and agentic AI," she said. "When you're working with these enterprise systems, you need agents that can perceive and understand the tools that they have access to. These storage servers are going to have to be semantic storage servers." Semantic storage servers provide knowledge instead of data, managing the relationship between different data points to answer queries faster. This is how the Dell data lakehouse offers more flexibility to customers: By doing all the semantic work and serving it up in a ready-to-go package, according to Chhabra. "Whether it's structured data or unstructured data, whether it's sitting on Dell storage or out on the edge or in the cloud, what apps and developers want is access to all of that data in a seamless way," he said. "They don't want to have to build a lot of capability around semantic search. They want that inbuilt into the platform. They want federated access." Governing agentic AI continues to be front-of-mind for infrastructure providers, ensuring that agents are constantly improving and providing accurate results. In this age of AI, the IT department has become the managers of a digital workforce, according to Briski. "You evaluate these digital workers," she explained. "And then, you're going to put them back out. The data flywheel is meant to do that. So as an agent is out working from those logs, from implicit and explicit feedback, from humans and logs, you're able to curate that data, customize, evaluate, guardrail and put them back out." Here's the complete video interview, part of SiliconANGLE's and theCUBE's coverage of Dell Technologies World:
[9]
Dell unveils new AI servers powered by Nvidia chips to boost enterprise adoption
Dell and Super Micro Computer have benefited from the growing demand for servers designed to handle the computer-heavy AI tasks, but the high cost of producing the systems and tough competition have pressured their margins. Dell in February forecast a decline in adjusted gross margin rate for fiscal 2026, while Super Micro projected fourth-quarter revenue below estimates earlier this month as tariff-driven economic uncertainty pressures its performance.Dell Technologies on Monday unveiled new servers powered by Nvidia's Blackwell Ultra chips, aiming to capitalize on the booming demand for artificial intelligence systems. The servers, available in both air-cooled and liquid-cooled variations, support up to 192 Nvidia Blackwell Ultra chips but can be customized to include as many as 256 chips. These servers can train AI models up to four times faster than previous models, Dell said. The pricing of these products will be "competitive," Arthur Lewis, President of Dell's Infrastructure Solutions Group, told Reuters, adding that "there's a lot of interest on what's next." Dell and Super Micro Computer have benefited from the growing demand for servers designed to handle the computer-heavy AI tasks, but the high cost of producing the systems and tough competition have pressured their margins. Dell in February forecast a decline in adjusted gross margin rate for fiscal 2026, while Super Micro projected fourth-quarter revenue below estimates earlier this month as tariff-driven economic uncertainty pressures its performance. Dell will focus on increasing sales of networking and storage products to ensure the "right level of profitability," Lewis said. The company's new servers will also support Nvidia's upcoming Vera central processing units, which will succeed the chip designer's Grace server processor. The AI server maker plans to support Nvidia's Vera Rubin chips, set to follow the Blackwell series. Dell also introduced a 'Pro Max Plus' laptop designed for AI development on Monday, featuring a neural processing unit that allows engineers to process large AI models directly on the device without relying on cloud services.
[10]
New PCs, Servers To Drive 'Breakthrough' Year Unveiled At Dell Technologies World 2025
At its annual Dell Technologies World conference, Dell is unveiling new laptops, servers and storage to help partners meet a market in which 75 percent of organizations say AI is key to their strategy and 65 percent are successfully moving AI concepts into production, according to one executive. Dell Technologies said AI projects that have been in the works for years are now moving out of proof of concept and into production, setting the company and its partners up to have a "breakthrough year," said Sam Grocott, senior vice president of product. Dell Technologies is unveiling new laptops, servers and storage at its Dell Technologies World conference this week to help partners meet a market in which 75 percent of organizations say AI is key to their strategy and 65 percent are successfully moving AI concepts into production, Grocott said during a media pre-briefing ahead of the show. "We are super excited about the maturing market and the way this is going and where I think this is setting us up to have a really breakthrough year," Grocott said. "Last year we introduced the Dell AI Factory and we've been busy over the last year with over 200 new releases and innovations across the Dell AI Factory. We've also deployed more than 3,000 of these across many enterprises." [RELATED: 'Fast and Furious' Dell Brings AI To Speed Partner Sales: Channel Chief] Grocott said Dell's AI Factory approach, which drives users to place AI workloads on prem, can be up to 62 percent more cost-effective for inferencing large language models (LLMs) than the public cloud and helps organizations securely and easily deploy enterprise. Grocott said the market is noticing and Dell is picking up customers among traditional technology clients such as GPU as a service, hyperscale and cloud services providers, but also seeing AI deployments in retail, finance, health care, and public sector. To meet the demand for new devices and capabilities, Dell has new AI PCs, leading with its Dell Pro Max lineup of desktops and notebooks that incorporate Qualcomm chips to power the CPU as well as Nvidia GB300s to tackle the AI or graphics workloads. Additionally, Dell's newest top-end PowerEdge servers, XE9780 and XE9780L, come in both air-cooled and liquid-cooled varieties, and can incorporate either AMD or Nvidia GPU architectures. Michael Dell introduced the audience to the devices on stage at Dell Technologies World 2025, calling the products the "infrastructure backbone" of enterprises as they create intelligence with their data. "Dell is the infrastructure, the backbone enabling enterprises to think faster, to act smarter and to dream bigger," he told the crowd. "Today we have now more than 3,000 customers running Dell AI Factories with a lot of success. The Dell AI Factory is up to 60 percent more cost-effective than the public cloud. Now with agents, and test time compute and deep reasoning, the models are helping us think and they're thinking and acting on their own, with autonomy." Dell Pro Max Plus laptop with Qualcomm During a pre-briefing, the company said the Dell Pro Max Plus laptop comes with a Qualcomm AI-100 NPU that provides 32 AI cores. It is the industry's first-ever enterprise-grade discrete NPU in a mobile form factor. Dell said the device is designed to enable AI inferencing at the edge with an eye towards AI engineers and data scientists who can now run large models at scale at the edge with the ability to power a 109 billion parameter model on the device, such as Meta Llama 4. Dell PowerCool Enclosed Rear Door Heat Exchanger (eRDHx) Dell said this is another industry first, with self-contained airflow systems that captures 100 percent of IT heat. It can do this with warm water systems. The devices can save customers 60 percent on their cooling costs compared to the standard rear door heat exchanger units available today, Dell said. The upside for compute is that organizations can deploy 16 percent more dense AI and HPC racks without increasing their power consumption, Dell said. It can cool up to 80-kilowatt racks and includes features such as hot swapable fans. This pairs with the new Dell Integrated Rack Controller, which can be placed inside OpenManage Enterprise software. That combination offers IT administrators leak detection and thermal management, through rack scale visibility. Dell AI Data Platform The platform is built to ingest, discover and process data from the edge to the data center or in the cloud, said Varun Chhabra, senior vice president of infrastructure and telecom marketing at Dell. One of the enhancements to the platform being unveiled is Project Lightning, which when it is released later this year will be the fastest parallel file system in the world, double the speed of traditional file systems. The AI data platform also integrates vector search for direct data set creation and querying within the Dell Data Lake House, which helps eliminate external systems for those capabilities. Dell has also built query functions around SQL-accessible LLM functions, text summarization, sentiment analysis, with all of it baked into a customer's instance of Dell's AI Data Platform. PowerEdge XE9785 and XE9785L servers with AMD Instinct MI350 Series These are the newest versions of Dell's flagship servers that come equipped with 288 gigabytes of HBM3E memory per GPU and deliver up to 35 times greater inferencing performance and are designed to house the AMD MI350 series chips. The new server comes in both air-cooled and liquid-cooled versions. Eventually they will be used to power Dell's AI Platform With AMD, which currently runs on XE9680s and has also been upgraded in time for Dell Tech World. Chhabra said while the platform was just introduced last year, this is the fourth update that combines Dell's PowerEdge servers, PowerScale storage, and PowerSwitch networking with AMD's ROCm AI accelerator framework, and MI300X GPUs for an 86 percent increase in customer's time to value, Chhabra said. Dell PowerEdge servers with Nvidia HGX B300 The newest versions of Dell's PowerEdge servers, the XE9780 and XE9785, run Nvidia's most advanced GB300 GPUs inside both liquid-cooled and air-cooled systems. The air-cooled versions are 10U systems that can support Intel Xeon 6 as well as AMD Epyc 5 processors, and can fit inside standard 19-inch racks. It provides 1.8 gigabytes per second of GPU to GPU throughput for faster model training as well as faster inferencing. It also features eight ConnectX 800-gigabit network interfaces per server, which lets customers cluster GPU servers together to create large scale systems for training and inferencing. The newest version of PowerEdge also has liquid-cooled versions that come in 3U form factors and can create the highest density GPU systems that Dell can deploy. It comes standard with the ability to install 192 GPUs per rack, and it is customizable up to 256 HGX B300 GPUs in a single rack. The machines are built to handle highly demanding AI workloads as well as HPC clusters.
[11]
Dell Unveils The Integration of NVIDIA's GB300 "Blackwell Ultra" GPUs With Its AI Factories, Taking Performance & Scalability to New Levels
Following NVIDIA's Computex keynote, Dell showcased its latest AI solutions featuring NVIDIA's latest AI architecture, designed to take the "AI factory" ecosystem to new levels. Dell has expanded its AI infrastructure massively over the past few years, especially after NVIDIA introduced its Hopper generation solutions, which managed to gain massive adoption in the AI markets. It seems like Dell isn't stopping just yet, as the firm has announced new and advanced PowerEdge AI racks, which feature NVIDIA's latest Blackwell GPUs, the HGX B300, which can support up to a 192-GPU configuration, making them one of the fastest options out there. Dell has also become one of the first firms to offer NVIDIA's RTX PRO 6000 "Blackwell" GPUs as well. Starting with Blackwell Ultra, Dell has unveiled the PowerEdge XE9780 and XE9785 "air-cooled" servers, which are said to come with up to 192 Blackwell Ultra HGX B300 GPUs. Interestingly, there are also liquid-cooled models, the PowerEdge XE9780L and XE9785L, which are said to support up to 256 Blackwell Ultra GPUs per rack. Through this, the firm has managed to bring in a whopping 4x faster LLM training performance compared to the previous generation offerings, which shows that the generational increment in performance has been massive over the past few years, which is why Dell is a leading integrator. Interestingly, Dell has also announced plans to release an AI solution based on NVIDIA's latest RTX PRO Blackwell GPUs, making them the first of their kind. The PowerEdge XE7745 is expected to be released by July 2025, and it is set to target the industrial segment of the market, particularly applications like multi-modal AI. The firm has also announced plans to develop PowerEdge designs featuring NVIDIA's Vera CPUs and Rubin GPUs as well, so the firm has extensive plans for the future as well. We're on a mission to bring AI to millions of customers worldwide. Our job is to make AI more accessible. Our job is to make AI more accessible. With the Dell AI Factory with NVIDIA, enterprises can manage the entire AI lifecycle across use cases, from deployment training, at any scale. - Dell's CEO Michael Dell Dell is probably one of the first to offer the most extensive NVIDIA-focused AI portfolio, and the firm has managed to take its "AI factory" ambitions to a new level with the recent Blackwell Ultra-based products.
[12]
Dell Technologies Launches Next-Gen Enterprise AI Solutions Powered by NVIDIA
Dell Technologies announces innovations across the Dell AI Factory with NVIDIA - all designed to help enterprises accelerate AI adoption and achieve faster time to value. As enterprises make AI central to their strategy and progress from experimentation to implementation, their demand for accessible AI skills and technologies grows exponentially. Dell and NVIDIA continue the rapid pace of innovation with updates to the Dell AI Factory with NVIDIA, including robust AI infrastructure, solutions and services that streamline the path to full-scale implementation. Dell infrastructure advances enterprise AI innovation with enhanced power, efficiency and scalability Dell Technologies introduces the next generation of advanced compute, data storage, data management and networking solutions:
[13]
Dell Technologies Unveils Next Generation Enterprise AI Solutions with NVIDIA
The Dell AI Factory with NVIDIA delivers new AI solutions that drive innovation for accelerated compute and data processing, streamline operations and achieve faster results at every stage of AI deployment Upgraded infrastructure, full-stack enterprise AI solutions and managed services are the latest additions to the industry's first and only end-to-end enterprise AI solution[1] The world's top provider of AI-centric infrastructure,[2] Dell Technologies (NYSE: DELL), announces innovations across the Dell AI Factory with NVIDIA - all designed to help enterprises accelerate AI adoption and achieve faster time to value. Why it matters As enterprises make AI central to their strategy and progress from experimentation to implementation, their demand for accessible AI skills and technologies grows exponentially. Dell and NVIDIA continue the rapid pace of innovation with updates to the Dell AI Factory with NVIDIA, including robust AI infrastructure, solutions and services that streamline the path to full-scale implementation. Dell infrastructure advances enterprise AI innovation with enhanced power, efficiency and scalability Dell Technologies introduces the next generation of advanced compute, data storage, data management and networking solutions: Air-cooled Dell PowerEdge XE9780 and XE9785 servers simplify integration into existing enterprise data centers, while liquid-cooled Dell PowerEdge XE9780L and XE9785L servers accelerate rack-scale deployment. The new PowerEdge servers support up to 192 NVIDIA Blackwell Ultra GPUs with direct to chip liquid cooling and can be customized with up to 256 NVIDIA Blackwell Ultra GPUs per Dell IR7000 rack. As the successors to Dell's fastest ramping solution ever,[3] the Dell PowerEdge XE9680, these platforms can deliver up to four times faster large language model (LLM) training with the 8-way NVIDIA HGX B300.[4] The Dell PowerEdge XE9712 featuring NVIDIA GB300 NVL72 offers efficiency at rack scale for training and 50 times more AI reasoning inference output and 5x improvement in throughput.[5] With new Dell PowerCool technology, this platform helps businesses achieve greater power efficiency. The Dell PowerEdge XE7745 server will be available with NVIDIA RTX Pro™ 6000 Blackwell Server Edition GPUs in July 2025. This platform - supported in the NVIDIA Enterprise AI Factory validated design - provides a universal platform to help meet the needs of physical and agentic AI use cases like robotics, digital twins and multi-modal AI applications with support for up to 8 GPUs in a 4U chassis. Dell plans to support the NVIDIA Vera CPU, offering speed, efficiency and performance. Dell plans to support the NVIDIA Vera Rubin platform with a new Dell PowerEdge XE server designed for Dell Integrated Rack Scalable Systems. Connecting it all, Dell extends its networking portfolio to include the Dell PowerSwitch SN5600, SN2201 Ethernet, part of the NVIDIA Spectrum-X Ethernet networking platform, and NVIDIA Quantum-X800 InfiniBand switches. These high-density, low-latency switches deliver up to 800 gigabits per second of throughput and are now backed by Dell ProSupport and Deployment Services to provide expert guidance at every stage of AI deployment. Dell AI Factory with NVIDIA solutions support the NVIDIA Enterprise AI Factory validated design, featuring Dell and NVIDIA compute, networking, storage and NVIDIA AI Enterprise software providing an end-to-end fully integrated AI solution for enterprises. Dell AI Data Platform advancements improve AI data management Because AI is only as powerful as the data that fuels it, organizations need a platform designed for performance and scalability. Dell AI Data Platform advancements provide AI applications with always-on access to high quality data. Dell ObjectScale supports large-scale AI deployments while helping reduce cost and data center footprint with the introduction of a denser, software-defined system. NVIDIA BlueField-3 and Spectrum-4 networking integrations boost performance and scalability. Dell introduces a high-performance solution built with Dell PowerScale, Dell Project Lightning and PowerEdge XE Using KV cache and integrating NVIDIA's NIXL Libraries, this solution is ideal for large-scale distributed inference workloads. Dell ObjectScale will support S3 over RDMA, achieving up to 230% higher throughput, up to 80% lower latency and 98% reduced CPU load compared to traditional S3 for better GPU utilization.[6] Dell announces an integrated solution that incorporates the NVIDIA AI Data Platform to accelerate curated insights from data and accelerate agentic AI applications and tools. Software updates help organizations seamlessly deploy agentic AI The NVIDIA AI Enterprise software platform, available directly from Dell, offers businesses the option to innovate on the Dell AI Factory with NVIDIA with NVIDIA NIM, NVIDIA NeMo microservices, and NVIDIA Blueprints, NVIDIA NeMo Retriever for RAG and NVIDIA Llama Nemotron reasoning models, and seamlessly develop agentic workflows while accelerating time-to-value for AI outcomes. Simplify business-critical AI deployments while providing flexibility and security with Red Hat OpenShift available on the Dell AI Factory with NVIDIA. New managed services streamline operations and drive faster outcomes The new Dell Managed Services for the Dell AI Factory with NVIDIA simplify AI operations with management of the full NVIDIA AI solutions stack -- including AI platforms, infrastructure and NVIDIA AI Enterprise software. Dell managed services experts handle 24×7 monitoring, reporting, version upgrades and patching, helping teams overcome resource and expertise constraints by providing cost-effective, scalable and proactive IT support. Perspectives "We're on a mission to bring AI to millions of customers around the world," said Michael Dell, chairman and chief executive officer, Dell Technologies. "Our job is to make AI more accessible. With the Dell AI Factory with NVIDIA, enterprises can manage the entire AI lifecycle across use cases, from training to deployment, at any scale." "AI factories are the infrastructure of modern industry, generating intelligence to power work across healthcare, finance and manufacturing," said Jensen Huang, founder and chief executive officer, NVIDIA. "With Dell Technologies, we're offering the broadest line of Blackwell AI systems to serve AI factories in clouds, enterprises and at the edge." Availability Check availability here. Additional Resources Connect with Dell on X and LinkedIn Learn more about the Dell AI Factory with NVIDIA. Blog: Continuing to Power the Future of AI with Dell's Direct Liquid Cooling and Computing Innovations Blog: Accelerate Your AI Workflows with the New Dell Data Lakehouse Features Blog: Pioneering Enterprise AI with Dell AI Factory and NVIDIA Blog: Operationalize AI Across Your Enterprise with Dell and Red Hat Blog: Managed Services: The Ultimate Enterprise AI Easy Button About Dell Technologies Dell Technologies (NYSE: DELL) helps organizations and individuals build their digital future and transform how they work, live and play. The company provides customers with the industry's broadest and most innovative technology and services portfolio for the AI era.
[14]
Dell Technologies Fuels Enterprise AI Innovation with Infrastructure, Solutions and Services
Advancements to the Dell AI Factory -- from industry-first AI PCs to edge and data center enhancements -- simplify and speed AI deployments for organizations of any size Powerful AI infrastructure and solutions, backed by a broad partner ecosystem and global services, empower organizations to embrace applications from building foundational models to running agentic AI Dell Technologies (NYSE: DELL), the world's No. 1 provider of AI infrastructure,1 announces Dell AI Factory advancements, including powerful and energy-efficient AI infrastructure, integrated partner ecosystem solutions and professional services to drive simpler and faster AI deployments. Why it matters AI is now essential for businesses, with 75% of organizations saying AI is key to their strategy2 and 65% successfully moving AI projects into production.3 However, challenges like data quality, security concerns and high costs can slow progress. The Dell AI Factory approach can be up to 62% more cost effective for inferencing LLMs on-premises than the public cloud4 and helps organizations securely and easily deploy enterprise AI workloads at any scale. Dell offers the industry's most comprehensive AI portfolio designed for deployments across client devices, data centers, edge locations and clouds.5 More than 3,000 global customers across industries are accelerating their AI initiatives with the Dell AI Factory.6 Dell infrastructure advancements help organizations deploy and manage AI at any scale Dell introduces end-to-end AI infrastructure to support everything from edge inferencing on an AI PC to managing massive enterprise AI workloads in the data center. Dell Pro Max AI PC delivers industry's first enterprise-grade discrete NPU in a mobile form factor7 The Dell Pro Max Plus laptop with Qualcomm® AI 100 PC Inference Card is the world's first mobile workstation with an enterprise-grade discrete NPU.8 It offers fast and secure on-device inferencing at the edge for large AI models typically run in the cloud, such as today's 109-billion-parameter model. The Qualcomm AI 100 PC Inference Card features 32 AI-cores and 64GB memory, providing power to meet the needs of AI engineers and data scientists deploying large models for edge inferencing. Dell redefines AI cooling with innovations that reduce cooling energy costs by up to 60%9 The industry-first Dell PowerCool Enclosed Rear Door Heat Exchanger (eRDHx) is a Dell-engineered alternative to standard rear door heat exchangers. Designed to capture 100% of IT heat generated with its self-contained airflow system, the eRDHx can reduce cooling energy costs by up to 60%10 compared to currently available solutions. With Dell's factory integrated IR7000 racks equipped with future-ready eRDHx technology, organizations can: Significantly cut costs and eliminate reliance on expensive chillers given the eRDHx operates with water temperatures warmer than traditional solutions (between 32 and 36 degrees Celsius). Maximize data center capacity by deploying up to 16% more racks11 of dense compute, without increasing power consumption. Enable air cooling capacity up to 80 kW per rack for dense AI and HPC deployments.12 Minimize risk with advanced leak detection, real-time thermal monitoring and unified management of all rack-level components with the Dell Integrated Rack Controller. Dell PowerEdge servers with AMD GPUs maximize performance and efficiency Dell PowerEdge XE9785 and XE9785L servers will support AMD Instinct™ MI350 series GPUs with 288 GB of HBM3e memory per GPU and up to 35 times greater13 inferencing performance.14 Available in liquid-cooled and air-cooled configurations, the servers will reduce facility cooling energy costs. Dell advancements power efficient and secure AI deployments and workflows Because AI is only as powerful as the data that fuels it, organizations need a platform designed for performance and scalability. The Dell AI Data Platform updates improve access to high quality structured, semi-structured and unstructured data across the AI lifecycle. Dell Project Lightning is the world's fastest parallel file system per new testing, delivering up to two times greater throughput than competing parallel file systems.15 Project Lightning will accelerate training time for large-scale and complex AI workflows. Dell Data Lakehouse enhancements simplify AI workflows and accelerate use cases -- such as recommendation engines, semantic search and customer intent detection -- by creating and querying AI-ready datasets. "We're excited to work with Dell to support our cutting-edge AI initiatives, and we expect Project Lightning to be a critical storage technology for our AI innovations," said Dr. Paul Calleja, director, Cambridge Open Zettascale Lab and Research Computing Services, University of Cambridge. With additional advancements, organizations can: Lower power consumption, reduce latency and boost cost savings for high performance computing (HPC) and AI fabrics with Dell Linear Pluggable Optics. Increase trust in the security of their AI infrastructure and solutions with Dell AI Security and Resilience Services, which provide full stack protection across AI infrastructure, data, applications and models. Dell expands AI partner ecosystem with customizable AI solutions and applications Dell is collaborating with AI ecosystem players to deliver tailored solutions that simply and quickly integrate into organizations' existing IT environments. Organizations can: Enable intelligent, autonomous workflows with a first-of-its-kind on-premises deployment of Cohere North, which integrates various data sources while ensuring control over operations. Innovate where the data is with Google Gemini and Google Distributed Cloud on-premises available on Dell PowerEdge XE9680 and XE9780 servers. Prototype and build agent-based enterprise AI applications with Dell AI Solutions with Llama, using Meta's latest Llama Stack distribution and Llama 4 models. Securely run scalable AI agents and enterprise search on-premises with Glean. Dell and Glean's collaboration will deliver the first on-premises deployment architecture for Glean's Work AI platform.16 Build and deploy secure, customizable AI applications and knowledge management workflows with solutions jointly engineered by Dell and Mistral AI. The Dell AI Factory also expands to include: Advancements to the Dell AI Platform with AMD add 200G of storage networking and an upgraded AMD ROCm open software stack for organizations to simplify workflows, support LLMs and efficiently manage complex workloads. Dell and AMD are collaborating to provide Day 0 support and performance optimized containers for AI models such as Llama 4. The new Dell AI Platform with Intel helps enterprises deploy a full stack of high performance, scalable AI infrastructure with Intel® Gaudi® 3 AI accelerators. Dell also announced advancements to the Dell AI Factory with NVIDIA and updates to Dell NativeEdge to support AI deployments and inferencing at the edge. Perspectives "It has been a non-stop year of innovating for enterprises, and we're not slowing down. We have introduced more than 200 updates to the Dell AI Factory since last year," said Jeff Clarke, chief operating officer, Dell Technologies. "Our latest AI advancements -- from groundbreaking AI PCs to cutting-edge data center solutions -- are designed to help organizations of every size to seamlessly adopt AI, drive faster insights, improve efficiency and accelerate their results." "We leverage the Dell AI Factory for our oceanic research at Oregon State University to revolutionize and address some of the planet's most critical challenges," said Christopher M. Sullivan, director of Research and Academic Computing for the College of Earth, Ocean and Atmospheric Sciences, Oregon State University. "Through advanced AI solutions, we're accelerating insights that empower global decision-makers to tackle climate change, safeguard marine ecosystems and drive meaningful progress for humanity." Additional Resources: Blog Blog Blog Connect with Dell on X and LinkedIn About Dell Technologies Dell Technologies (NYSE: DELL) helps organizations and individuals build their digital future and transform how they work, live and play. The company provides customers with the industry's broadest and most innovative technology and services portfolio for the AI era.
[15]
Dell unveils new AI servers powered by Nvidia chips to boost enterprise adoption
(Reuters) -Dell Technologies on Monday unveiled new servers powered by Nvidia's Blackwell Ultra chips, aiming to capitalize on the booming demand for artificial intelligence systems. The servers, available in both air-cooled and liquid-cooled variations, support up to 192 Nvidia Blackwell Ultra chips but can be customized to include as many as 256 chips. These servers can train AI models up to four times faster than previous models, Dell said. The pricing of these products will be "competitive," Arthur Lewis, President of Dell's Infrastructure Solutions Group, told Reuters, adding that "there's a lot of interest on what's next." Dell and Super Micro Computer have benefited from the growing demand for servers designed to handle the computer-heavy AI tasks, but the high cost of producing the systems and tough competition have pressured their margins. Dell in February forecast a decline in adjusted gross margin rate for fiscal 2026, while Super Micro projected fourth-quarter revenue below estimates earlier this month as tariff-driven economic uncertainty pressures its performance. Dell will focus on increasing sales of networking and storage products to ensure the "right level of profitability," Lewis said. The company's new servers will also support Nvidia's upcoming Vera central processing units, which will succeed the chip designer's Grace server processor. The AI server maker plans to support Nvidia's Vera Rubin chips, set to follow the Blackwell series. Dell also introduced a 'Pro Max Plus' laptop designed for AI development on Monday, featuring a neural processing unit that allows engineers to process large AI models directly on the device without relying on cloud services. (Reporting by Jaspreet Singh in Bengaluru; Editing by Shreya Biswas)
Share
Copy Link
Dell Technologies announces a range of new AI-focused products and services at its annual conference, positioning itself as a one-stop shop for enterprise AI infrastructure with an emphasis on on-premises deployment.
At Dell Technologies World 2025, CEO Michael Dell outlined a vision for a "decentralized, low-latency, and hyper-efficient" future of AI 1. This strategy aligns with Dell's new offerings, which focus on on-premises AI solutions and edge computing, positioning the company as a comprehensive provider of AI infrastructure 12.
Source: ZDNet
The Dell AI Factory, a managed service for on-premise AI, is at the heart of Dell's new offerings. The company claims it can be "up to 62% more cost-effective for inferencing LLMs on-premises than the public cloud" 1. This aligns with Dell's survey findings that 37% of enterprise customers want an infrastructure vendor to build their entire AI stack 1.
Key components of the AI Factory include:
Dell is also emphasizing edge computing solutions:
To support its hardware offerings, Dell introduced several software solutions:
Source: SiliconANGLE
Dell introduced the PowerCool Enclosed Rear Door Heat Exchanger, a new approach to AI-scale thermal management. This system is designed to absorb 100% of server-generated heat and can potentially reduce cooling-related energy costs by up to 60% 4.
Dell's strategy appears to be a response to the growing demand for AI infrastructure and the challenges faced by enterprises in deploying AI solutions. By offering a comprehensive suite of products and services, Dell aims to simplify the AI deployment process for its customers 123.
However, the company faces stiff competition in the AI infrastructure market, particularly from Nvidia, which dominates the GPU market for AI applications 5. Dell's emphasis on on-premises solutions and partnerships with multiple chip manufacturers (including Nvidia and AMD) may help differentiate its offerings in this competitive landscape 123.
Source: SiliconANGLE
As the AI infrastructure market continues to evolve rapidly, Dell's success will likely depend on its ability to deliver on its promises of cost-effectiveness, performance, and comprehensive support for enterprise AI deployments.
Summarized by
Navi
[4]
Google launches its new Pixel 10 smartphone series, showcasing advanced AI capabilities powered by Gemini, aiming to challenge competitors in the premium handset market.
20 Sources
Technology
2 hrs ago
20 Sources
Technology
2 hrs ago
Google's Pixel 10 series introduces groundbreaking AI features, including Magic Cue, Camera Coach, and Voice Translate, powered by the new Tensor G5 chip and Gemini Nano model.
12 Sources
Technology
3 hrs ago
12 Sources
Technology
3 hrs ago
NASA and IBM have developed Surya, an open-source AI model that can predict solar flares and space weather with improved accuracy, potentially helping to protect Earth's infrastructure from solar storm damage.
6 Sources
Technology
10 hrs ago
6 Sources
Technology
10 hrs ago
Google's latest smartwatch, the Pixel Watch 4, introduces significant upgrades including a curved display, enhanced AI features, and improved health tracking capabilities.
17 Sources
Technology
2 hrs ago
17 Sources
Technology
2 hrs ago
FieldAI, a robotics startup, has raised $405 million to develop "foundational embodied AI models" for various robot types. The company's innovative approach integrates physics principles into AI, enabling safer and more adaptable robot operations across diverse environments.
7 Sources
Technology
2 hrs ago
7 Sources
Technology
2 hrs ago