2 Sources
2 Sources
[1]
AWS custom AI silicon helped Metagenomi cut AI bill 56%
Gene editing startup Metagenomi has tapped AWS's Inferentia 2 accelerators to speed the discovery of potentially life-saving therapies, and said its efforts cost 56 percent less than it would have incurred using Nvidia GPUs. Founded in 2018, Metagenomi is using a Nobel prize-winning approach developed by Jennifer Doudna and Emmanuelle Charpentier called CRISPR, which allows for the targeted editing of gene sequences. "Gene editing is a new therapeutic modality aimed at treating disease by addressing the cause of disease at the genetic level. So rather than treating the symptoms, actually going after a cure," Chris Brown, VP of discovery at Metagenomi, told El Reg. These therapies rely on identifying enzymes - essentially biological catalysts that facilitate chemical reactions - that can bind to the RNA sequences that guide them to their destination, cut the target DNA in the right spot, and - critically - fit in the delivery mechanism of choice. To find these enzymes, the startup is using a class of generative AI known as protein language models (PLMs), like Progen2, to rapidly generate millions of potential candidates. "It's about finding that one thing in a million. So if you've got access to twice as many, you're doubling your chances of potentially getting a product at the end," Brown said. Developed by researchers at Salesforce, Johns Hopkins, and Columbia Universities in 2022, Progen2 is an auto-regressive transformer model not unlike GPT-2. But rather than spitting out strings of text, it synthesizes novel protein sequences. Weighing in at about 800 million parameters for the base model, Progen2 is tiny compared to modern large language models like GPT-4 or DeepSeek R1, which means running it doesn't require massive quantities of high-bandwidth memory. For the trial, Metagenomi compared AWS's Inferentia 2 accelerator with Nvidia's L40S, which the biotech startup had previously been using to run Progen2. Launched in 2023, Inferentia 2 is (as its name suggests) an inference-optimized accelerator, with 32GB of HBM, 820 GB/s of memory bandwidth, and 190 teraFLOPS of 16-bit performance. By comparison, the L40S, based on Nvidia's previous-gen Ada Lovelace GPU architecture, features 48GB of GDDR6 good for 864 GB/s of memory bandwidth and 362 teraFLOPS at 16-bit precision. But while the L40S outperforms Inferentia 2 on paper, Amazon claims its chip can do the job cheaper by taking advantage of its batch processing pipeline, AWS Batch, and spot instances. "Spot Instances are generally 70-ish percent lower cost than on demand. Because the workflows that they were optimizing for could be scheduled around spot Instances utilizing AWS Batch, it really simplified these deployments ... and allowed them to schedule different types of experimentation to run around the clock," Kamran Khan, head of business development for the machine learning wing of AWS's Annapurna Labs team, told The Register. A chunk of the savings from using Inferentia came from greater availability. The cloud giant says the interruption rate for its homegrown chip is roughly five percent compared to 20 percent for Nvidia's L40S-based spot instances. In theory, this means that only one in 20 of Metagenomi's protein generation batches should be interrupted, versus one in five for Nvidia's accelerator. For Brown, Inferentia's lower operating cost translates directly into more science, increasing the likelihood of discovering enzymes capable of targeting different ailments. "We took a problem where it would have been one project for the year, and instead we turned it into something that my team can do multiple times a day or a week," Brown said. The collab also highlights that for AI workloads that aren't interactive, faster hardware isn't always better - older, heavily discounted accelerators may offer better value. ®
[2]
Metagenomi taps Amazon's custom chips to develop gene-editing tools
SAN FRANCISCO, Oct 22 (Reuters) - Biotech firm Metagenomi (MGX.O), opens new tab said it is using artificial intelligence chips from Amazon.com's (AMZN.O), opens new tab cloud computing unit to power some of its work in developing gene-editing technologies. The work represents one of the first major uses of Amazon Web Services chips beyond chatbots and other products powered by large language models, with Metagenomi saying AWS's Inferentia chips had proven far more cost-effective than products from rivals like Nvidia (NVDA.O), opens new tab. Emeryville, California-based firm Metagenomi is one of a number of companies working to develop tools to inject genetic material into the human body where it can edit genes to treat diseases. To do that, Metagenomi sifts through the natural world looking for proteins that might help with the challenge of precisely delivering a treatment to any gene in any cell of the human body. When it finds one that looks close, it uses AI to generate many similar examples to find just the right ones. "So we did this impressive thing where we generated over a million different proteins from this rare class of enzyme that we use for doing gene editing. And in that case, it was really a clear cost advantage for using the Inferentia platform," Chris Brown, head of discovery for Metagenomi. "Unless you cast a broad enough net in the beginning, you actually just miss what can be the most important discoveries and systems entirely," he added. The Inferentia chips - which were introduced in 2019 to power AI features for Amazon's Alexa virtual assistant - were able to do the same job as rival chips at roughly half the price, Brown said. Reporting by Stephen Nellis in San Francisco; Editing by Edwina Gibbs Our Standards: The Thomson Reuters Trust Principles., opens new tab
Share
Share
Copy Link
Biotech startup Metagenomi has leveraged AWS's Inferentia 2 accelerators to significantly reduce costs and accelerate gene-editing research. The collaboration highlights the potential of custom AI chips in biotechnology applications.

Gene editing startup Metagenomi has made significant strides in its research by leveraging Amazon Web Services' (AWS) custom AI chips, the Inferentia 2 accelerators. This collaboration has resulted in a remarkable 56% reduction in costs compared to using traditional Nvidia GPUs for their AI-driven gene therapy discovery process
1
.Metagenomi, founded in 2018, is utilizing the Nobel prize-winning CRISPR technology to develop targeted gene-editing therapies. Their approach aims to treat diseases at the genetic level rather than merely addressing symptoms. The key to their research lies in identifying specific enzymes that can bind to RNA sequences, cut DNA at precise locations, and fit within the chosen delivery mechanism
1
.To accelerate their research, Metagenomi employs a class of generative AI known as protein language models (PLMs), specifically Progen2. This model, developed by researchers at Salesforce, Johns Hopkins, and Columbia Universities in 2022, is capable of synthesizing novel protein sequences. With approximately 800 million parameters, Progen2 is relatively small compared to modern large language models, making it suitable for running on specialized hardware like AWS's Inferentia 2
1
.The trial conducted by Metagenomi compared AWS's Inferentia 2 accelerator with Nvidia's L40S GPU. While the L40S boasts higher raw performance metrics, Amazon's Inferentia 2 proved more cost-effective for Metagenomi's specific workload. The Inferentia 2, launched in 2023, features 32GB of HBM, 820 GB/s of memory bandwidth, and 190 teraFLOPS of 16-bit performance
1
.AWS's solution leverages its batch processing pipeline, AWS Batch, and spot instances to achieve significant cost reductions. The lower interruption rate of Inferentia 2 spot instances (5% compared to 20% for L40S) contributes to greater availability and efficiency. Chris Brown, VP of discovery at Metagenomi, emphasized that this cost-effectiveness translates directly into more science, enabling his team to perform multiple experiments daily or weekly instead of just one project per year
1
2
.Related Stories
This collaboration between Metagenomi and AWS represents one of the first major applications of Amazon's custom AI chips beyond chatbots and large language models. It demonstrates the potential for specialized AI accelerators in biotechnology research, particularly in tasks that don't require interactive performance but benefit from cost-effective batch processing
2
.The success of this partnership opens up new possibilities for the use of custom AI chips in various scientific fields. As biotechnology companies continue to harness the power of AI for drug discovery and gene therapy development, the demand for cost-effective and efficient computing solutions is likely to grow. This trend could reshape the landscape of AI hardware in scientific research, potentially challenging the dominance of traditional GPU manufacturers in certain niche applications.
Summarized by
Navi
[1]
12 Nov 2024•Technology

04 Dec 2024•Technology

03 Dec 2024•Technology
