Curated by THEOUTPOST
On Tue, 8 Oct, 4:12 PM UTC
3 Sources
[1]
AMD-based AI cloud platform TensorWave raises $43M to increase data center capacity - SiliconANGLE
AMD-based AI cloud platform TensorWave raises $43M to increase data center capacity TensorWave Inc., an artificial intelligence GPU cloud delivery platform aimed at delivering an alternative to Nvidia Corp.'s market dominance using Advanced Micro Devices Inc. chips, said today it has raised $43 million in a so-called "simple agreement for future equity" investment round. The round was led by Nexus VP with participation from Maverick Capital, Translink Capital, Javelin Venture Partners, StartupNV and AMD Ventures. SAFE agreements, conceived in 2013 by the startup incubator Y Combinator, provide potential future equity in a company for the investor in exchange for immediate cash to the company. Founded in 2023, TensorWave sought to help solve key challenges in the AI and machine learning market by providing greater availability and optionality to GPUs for AI training and inference. Currently, it's estimated that Nvidia controls between 70% and 95% of the chip market for training and deploying AI models. Speaking to SiliconANGLE in an interview, TensorWave Chief Executive Darrick Horton said that before launching the company, even if people could get equivalent or better compute and capability delivered compared with Nvidia cards, if it wasn't simpler or faster to implement, then they wouldn't want to switch. That's when his team took notice of AMD's Instinct MI300x series data center AI accelerators, released last year, which fit the bill. As for the funding, Horton said it will be used to scale up the company's primary data center with AMD Instinct accelerators to tackle AI workloads and also prepare its infrastructure for the next generation of AMD chips. The chipmaker intends to roll out a new generation of AI accelerators, the MI325X, in the fourth quarter of this year that appear to put Nvidia in the company's sights. "The MI300 for context is, you know, around the same as the H100, sometimes it's better, sometimes it's worse. It depends on the workload," said Horton. "The MI325 is going to be significantly better than the H200. So, it will have the most memory of any chip on the market and the most memory bandwidth of any chip on the market. It will dominate on inference workloads." The H100 and H200 are Nvidia's most powerful Tensor Core GPUs based on Nvidia Hopper architecture, making them capable of providing powerful accelerated AI compute. Horton explained that AMD's chip offerings have been helping compete with Nvidia on price-performance ratios for the customers of its cloud compute solution. TensorWave said the company intends to add MI325X instances to its offering as early as the end of the year after AMD launches its new silicon. The global AI market is projected to grow from $196.6 billion in 2023 to $1.81 trillion by 2030, according to Grand View Research. Much of this will be driven by the accelerating rate of AI research and innovation and the continuous expansion of AI product offerings and services across every industry, including large language models from companies such as OpenAI and Google LLC. All of these models require massive amounts of compute for both training and deployment. Horton added that the funding will also help the company launch Manifest, a specialized AI deployment platform. "Manifest is a purpose-built, highly optimized inference stack for AMD that is accessible only at TensorWave," said Horton. "We're building in a lot of unique features that take advantage of AMD and offer solutions for actual enterprise use cases that are missing in the marketplace right now." The platform will support larger context windows, which is the amount of data that an AI system can ingest at a time to work on, and offer lower response times. Manifest will provide efficient analysis of complex documents for enterprise customers, provide accelerated reasoning and a secure platform for private data storage. It will launch during the fourth quarter of 2024.
[2]
TensorWave Rides AI Wave with Record-Breaking $43M Funding Round
In a move that encourages Nevada's growing tech scene, AI cloud platform TensorWave announced today it has secured a statement inducing $43 million in SAFE funding. The largest of its kind for a Nevada-based startup, the funding highlights the increasing attractiveness of Nevada as a hub for tech innovation, with the state's favorable business environment and growing talent pool drawing significant attention from investors and entrepreneurs alike, supporting strong growth in sectors like cybersecurity, aerospace, and logistics technology TensorWave aims to address a typical bottleneck: access to powerful and affordable GPU resources. The company offers a compelling alternative to the existing options currently available, leveraging AMD Instinct Series GPUs rather than siding with NVIDIA, to deliver unparalleled scalability and performance - a strategy it recently discussed with Dataconomy. By eliminating the wait times and complexities associated with on-premise server management, TensorWave aims to democratize AI compute for startups, enterprises, and researchers alike. "We are thrilled to have the support of such esteemed investors, partners, and the State of Nevada as we embark on this next phase of growth," said Darrick Horton, CEO at TensorWave. "This funding allows us to significantly scale our team and deploy thousands of AMD Instinct AI accelerators to empower the startups and enterprises shaping our technological future." The funding, led by Nexus VP, with significant contributions from Maverick Capital, Translink Capital, Javelin Venture Partners, StartupNV, and AMD Ventures, will fuel TensorWave's expansion plans. The company plans to bolster its core team, significantly increase capacity at its primary data center with the addition of thousands of AMD Instinct MI300X GPUs, and prepare for the integration of the next-generation MI325X GPUs. Furthermore, TensorWave is gearing up to launch Manifest, an inference platform designed for handling large context windows with minimal latency, enabling advanced document analysis and accelerated reasoning. "AMD Ventures shares TensorWave's vision to transform AI compute infrastructure," said Mathew Hein, Senior Vice President, Chief Strategy Officer and Corporate Development at AMD. "Their deployment of the AMD Instinct MI300X and ability to offer public instances to AI customers and developers positions them as an early competitor in the AI space, and we are excited to support their growth through this latest round of funding." TensorWave's strategic deployment of AMD GPUs positions the company at the forefront of the rapidly expanding AI market, which is projected to reach a staggering $1.81 trillion by 2030, according to Grand View Research. Investors have recognized the company's potential to become a key player in this burgeoning sector. "TensorWave has impressed with their mission, vision, and rapid execution. We're thrilled to support this dynamic startup, which is poised to lead AI cloud computing with AMD GPUs," says Brendan Walsh, Venture Partner at Translink Capital. "Their deep expertise in both the physical plant and software elements of cloud compute and artificial intelligence, along with their commitment to fostering a competitive, open-source future for AI, makes them an exciting company to watch as they continue to scale." With its focus on accessibility, performance, and innovation, TensorWave is set to make waves in the AI industry. The company's commitment to providing a robust and competitive cloud platform powered by AMD's cutting-edge GPU technology has the potential to reshape the AI landscape and accelerate the development of groundbreaking applications across various sectors.
[3]
TensorWave thinks it can break Nvidia's grip on AI compute with an AMD-powered cloud
Chipmaker Nvidia notched $30 billion in revenue last fiscal quarter, driven in large part by the AI industry's insatiable demand for GPUs. GPUs are essential for training and running AI models; they contain thousands of cores that work in parallel to quickly perform the linear algebra equations scaffolding the models. The appetite for AI remains high, and Nvidia's GPUs have become the chip of choice among AI players of all sizes. But TensorWave, a company founded late last year, is going against the grain by launching a cloud that only offers access to hardware from Nvidia rival AMD for AI workloads. "We recognized an unhealthy monopoly at work -- one that was starving end-users of compute access and stifling innovation in the AI space," Darrick Horton, TensorWave's CEO and one of its co-founders, told TechCrunch. "Motivated by our desire to democratize AI, we set out to provide a viable alternative and restore competition and choice." Winding paths Pickleball initially brought Horton together with TensorWave's other two co-founders, Jeff Tatarchuk and Piotr Tomasik -- or at least, it's what got the ball rolling (excuse the pun). After a match one day, Tomasik and Tatarchuk -- close friends and longtime pickleball doubles partners -- invited Horton, a former colleague of Tatarchuk's, to join them at their favorite Las Vegas watering hole. "As the conversation unfolded, we discussed the monopolistic grip on GPU compute capacity, which was leading to supply constraints," Horton said. "This realization led to the formation of TensorWave." The three partners didn't just know each other from pickleball. Tatarchuk co-founded cloud vendor VMAccel with Horton before selling another startup, CRM developer Lets Rolo, to digital identity firm LifeKey. Horton, who has bachelor's degrees in mechanical engineering and physics, once worked at Lockheed Martin's Skunk Works R&D division, then co-founded VaultMiner Technologies, a crypto mining company and VMAccel's corporate parent. As for Tomasik, he co-launched Lets Rolo alongside Tatarchuk. (Tomasik is also a co-founder of influencer marketer site Influential, which French PR firm Publicis acquired for $500 million in July.) So what made three entrepreneurs with little knowledge of the hyperscaler landscape think they could compete with titans of AI industry? Tenacity, basically. "We believed we could solve the GPU supply problem," Horton said. Vegas, inc. TensorWave is headquartered in Las Vegas, an unusual choice of city for a cloud infrastructure startup. But Horton said that the team liked the odds. "We thought that Vegas had the potential to become a thriving technology and startup ecosystem," he said. That prediction's not totally off base. According to Dealroom.co data, Las Vegas is home to just over 600 startups employing more than 11,000 people, which attracted over $4 billion in investments in 2022. Energy costs and overhead are lower in Vegas than in many major U.S. cities, too. And both Tomasik and Tatarchuk have close ties to the city's VC community. Tomasik was previously a GP at Vegas-based seed fund 1864 Fund, and now works with the nonprofit accelerators StartUp Vegas and Vegas Tech Ventures. (Strangely, Vegas Tech Ventures' site threw a 404 for the pages listing its partners and portfolio companies; a spokesperson said that it was a technical error and would be corrected.) Tatarchuk is an angel investor at Fruition Lab, a Vegas incubator that rather unusually began as a Christian religious organization. These connections -- along with Horton's -- helped bootstrap TensorWave into becoming one of the first clouds to market with AMD Instinct MI300X instances for AI workloads. Delivering setups with dedicated storage and high-speed interconnects upon request, TensorWave rents GPU capacity by the hour, and requires a minimum six-month contract. "In the cloud space as whole, we are in good company," Horton said. "We see ourselves as complementary, offering additional AI specific compute at competitive price-to-performance." CoreWeave, the GPU infrastructure provider that began life as a crypto mining operation, recently raised $1.1 billion in new funds (and $7.5 billion in debt) and signed a multi-billion-dollar capacity deal with Microsoft. Lambda Labs in early April secured a special purpose financing vehicle of up to $500 million, and is reportedly seeking an additional $800 million. The nonprofit Voltage Park, backed by crypto billionaire Jed McCaleb, last October announced that it's investing $500 million in GPU-backed data centers. And Together AI, a cloud GPU host that also conducts generative AI research, in March landed $106 million in a Salesforce-led round. So how does TensorWave hope to compete? First, on price. Horton notes that the MI300X is significantly cheaper than Nvidia's most popular GPU for AI workloads at present, the H100, and that this allows TensorWave to pass savings onto customers. He wouldn't reveal TensorWave's exact instance pricing. But to beat the more competitive H100 plans, it would have to come under ~$2.50 per hour -- a challenging but not inconceivable feat. "Pricing ranges from approximately $1 per hour to $10 per hour, depending on the bespoke requirements of the workload and the GPU configurations chosen," Horton said. "As for the cost per instance that TensorWave incurs, we are unable to share those details due to confidentiality agreements." Second, on performance. Horton points to benchmarks showing the MI300X outgunning the H100 when it comes to running (but not training) AI models, specifically text-generating models like Meta's Llama 2. (Other evaluations suggest that the advantage may be workload-dependent.) There seems to be some credence to Horton's claims, given interest from tech industry movers and shakers in the MI300X. Meta said in December that it'll use MI300X chips for use cases like running its Meta AI assistant, while OpenAI, the maker of ChatGPT, plans to support the MI300X in its developer tooling. Working in all of these vendors' favor right now is the continued Nvidia GPU shortage and the delay of Nvidia's upcoming Blackwell chip. But the shortage could ease soon with a ramp-up in the manufacturing of critical chip components, in particular memory. And that could allow Nvidia to scale up shipments of the H200, the H100's successor, which boasts dramatically improved performance. Another existential dilemma for upstart clouds betting on AMD hardware is bridging the competitive moats Nvidia has built around AI chips. Nvidia's development software is perceived as more mature and easier to use than AMD's -- and it's widely deployed. AMD CEO Lisa Su herself has admitted that it "takes work" to adopt AMD. On the far horizon, competing on pricing might become challenging down the line as hyperscalers increase their investments in custom hardware to run and train models. Google offers its TPUs; Microsoft recently unveiled two custom chips, Azure Maia and Azure Cobalt; and AWS has Trainium, Inferentia and Graviton. "As developers seek alternatives that can effectively handle their AI workloads, especially with increased memory and performance demands, along with ongoing production issues causing delays, AMD will maintain superiority for even longer, playing a key role in the democratization of compute in the AI era," Horton said. Early demand TensorWave began onboarding customers late this spring in preview. But it's already generating $3 million in annual recurring revenue, Horton says. He expects that figure will reach $25 million by the end of the year -- an 8x leap -- once TensorWave ratchets up capacity to 20,000 MI300Xs. Assuming $15,000 per GPU, 20,000 MI300Xs would amount to a $300 million investment -- yet Horton claims TensorWave's burn rate is "well within sustainable levels." TensorWave previously told The Register that it would use its GPUs as collateral for a large round of debt financing, an approach employed by other data center operators including CoreWeave; Horton says that's still the plan. "This reflects our strong financial health," he continued. "We're strategically positioned to weather potential headwinds by delivering value where it's most needed." I asked Horton how many customers TensorWave has today. He declined to answer due to "confidentiality," but highlighted TensorWave's publicly announced partnerships with networking backbone provider Edgecore Networks and MK1, an AI inferencing startup founded by ex-Neuralink engineers. "We are rapidly expanding our capacity, with multiple nodes available, and we are continually increasing capacity to meet the growing demands of our pipeline," Horton said, adding that TensorWave plans to bring AMD's next-gen MI325X GPUs, which are scheduled to be released in Q4 2024, online as early as November/December. Investors seem pleased with TensorWave's growth trajectory so far. Nexus VP revealed on Wednesday that it led a $43 million round in the company, which also had participation from Maverick Capital, StartupNV, Translink Capital, and AMD Ventures. The tranche -- TensorWave's first -- values the startup at $100 million post-money. "AMD Ventures shares TensorWave's vision to transform AI compute infrastructure," AMD Ventures SVP Matthew Hein said in a statement. "Their deployment of the AMD Instinct MI300X and ability to offer public instances to AI customers and developers positions them as an early competitor in the AI space, and we are excited to support their growth through this latest round of funding."
Share
Share
Copy Link
TensorWave, an AI cloud platform using AMD GPUs, raises $43 million to expand its data center capacity and launch a new inference platform, aiming to provide an alternative to Nvidia's dominance in the AI chip market.
TensorWave, an AI cloud platform startup, has successfully raised $43 million in a SAFE (Simple Agreement for Future Equity) funding round, marking the largest of its kind for a Nevada-based startup [1][2]. The funding round was led by Nexus VP, with participation from Maverick Capital, Translink Capital, Javelin Venture Partners, StartupNV, and AMD Ventures [1][2].
Founded in 2023, TensorWave aims to address a critical challenge in the AI and machine learning market by providing greater availability and optionality for GPUs used in AI training and inference [1]. The company is positioning itself as an alternative to Nvidia's market dominance, which is estimated to control between 70% and 95% of the chip market for AI applications [1].
TensorWave's unique selling point is its use of AMD Instinct Series GPUs, particularly the MI300X, instead of Nvidia chips [1][2]. The company's CEO, Darrick Horton, highlighted the competitive edge of AMD's offerings:
"The MI300 for context is, you know, around the same as the H100, sometimes it's better, sometimes it's worse. It depends on the workload," said Horton. "The MI325 is going to be significantly better than the H200. So, it will have the most memory of any chip on the market and the most memory bandwidth of any chip on the market. It will dominate on inference workloads." [1]
The funding will be used to scale up TensorWave's primary data center with AMD Instinct accelerators and prepare for the integration of the next-generation MI325X GPUs [1][2]. Additionally, the company plans to launch Manifest, a specialized AI deployment platform designed for handling large context windows with minimal latency [2].
The global AI market is projected to grow from $196.6 billion in 2023 to $1.81 trillion by 2030, according to Grand View Research [1]. This growth potential has attracted significant investor interest in TensorWave's approach.
Mathew Hein, Senior Vice President at AMD, expressed support for TensorWave's vision: "AMD Ventures shares TensorWave's vision to transform AI compute infrastructure. Their deployment of the AMD Instinct MI300X and ability to offer public instances to AI customers and developers positions them as an early competitor in the AI space." [2]
TensorWave aims to compete on both price and performance. Horton claims that the MI300X is significantly cheaper than Nvidia's H100, allowing TensorWave to offer more competitive pricing to customers [3]. While exact pricing details were not disclosed, Horton suggested a range of approximately $1 to $10 per hour, depending on workload requirements and GPU configurations [3].
TensorWave's founding team brings diverse experience to the table. The three co-founders – Darrick Horton, Jeff Tatarchuk, and Piotr Tomasik – have backgrounds in cloud computing, CRM development, and influencer marketing [3]. Their decision to base the company in Las Vegas was driven by the city's potential as a growing tech hub and its favorable business environment [3].
As TensorWave moves forward with its ambitious plans, the company's success could potentially reshape the AI infrastructure landscape, offering a compelling alternative to Nvidia's established dominance in the market.
CoreWeave, an AI-optimized cloud platform operator, has closed a $650 million secondary sale led by major investors. The deal values the company at $23 billion, reflecting growing interest in AI cloud infrastructure.
3 Sources
GMI Cloud, a GPU cloud infrastructure provider, has raised $82 million in a Series A funding round to expand its AI-focused services and data center capacity.
2 Sources
Intel launches Tiber AI Cloud, powered by Gaudi 3 chips, partnering with Inflection AI to offer enterprise AI solutions, competing with major cloud providers and NVIDIA in the AI accelerator market.
4 Sources
AMD's AI GPU business, led by the Instinct MI300, has grown rapidly to match the company's entire CPU operations in revenue. CEO Lisa Su predicts significant market growth, positioning AMD as a strong competitor to Nvidia in the AI hardware sector.
4 Sources
AMD unveils its next-generation AI accelerator, the Instinct MI325X, along with new networking solutions, aiming to compete with Nvidia in the rapidly growing AI infrastructure market.
16 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2024 TheOutpost.AI All rights reserved