

AMD AI chips are nearly as fast as Nvidia's, MosaicML says | TechSpot
source link: https://www.techspot.com/news/99271-amd-ai-chips-punch-above-their-weight-mosaicml.html
Go to the source link to view the article. You can view the picture content, updated content and better typesetting reading experience. If the link is broken, please click the button below to view the snapshot at that time.

AMD AI chips are nearly as fast as Nvidia's, MosaicML says
Software is key in LLM training
By Alfonso Maruccia July 3, 2023, 11:16 AM 7 comments
Why it matters: MosaicML is an AI startup that was recently acquired by Databricks for $1.3 billion. Both companies advocate for a DIY approach to AI systems and LLM training platforms, enabling companies to maintain control over their AI applications. Regarding hardware, MosaicML claims AMD chips can deliver nearly equivalent performance to Nvidia chips.
As Nvidia's recent surge in market capitalization clearly demonstrates, the AI industry is in desperate need of new hardware to train large language models (LLMs) and other AI-based algorithms. While server and HPC GPUs may be worthless for gaming, they serve as the foundation for data centers and supercomputers that perform highly parallelized computations necessary for these systems.
When it comes to AI training, Nvidia's GPUs have been the most desirable to date. In recent weeks, the company briefly achieved an unprecedented $1 trillion market capitalization due to this very reason. However, MosaicML now emphasizes that Nvidia is just one choice in a multifaceted hardware market, suggesting companies investing in AI should not blindly spend a fortune on Team Green's highly sought-after chips.
The AI startup tested AMD MI250 and Nvidia A100 cards, both of which are one generation behind each company's current flagship HPC GPUs. They used their own software tools, along with the Meta-backed open-source software PyTorch and AMD's proprietary software, for testing.
MosaicML trained an LLM algorithm without making any changes to the underlying software code, and found that AMD chips performed nearly as well as those from Nvidia.

In real workload-based tests, MosaicML reports that the LLM training stack remained stable and performed well without any additional configuration. AMD MI250 GPUs were "competitive," the company stated, delivering 80 percent of the per-GPU data throughput offered by Nvidia's A100 40 GB model and within 73 percent compared to the A100 800 GB model.
Hanlin Tang, chief technology officer at MosaicML, states that the major weakness for most companies manufacturing chips for ML algorithm acceleration lies in their software. AMD excelled in this area, and the company is expecting even better performance on new HPC GPUs as software tools continue to improve. It should be mentioned however that CUDA – Nvidia's low-level chip programming framework, has become a sort of standard in the industry, at least for now. CUDA is not perfect, or elegant, or especially easy, but it is familiar and it is Nvidia-only.
AMD is understandably pleased with MosaicML's findings, which seemingly validate the company's strategy of supporting an "open and easy to implement software ecosystem" for AI training and inference on its chips. Nvidia, meanwhile, declined to comment.
Recommend
-
7
AI ‘a lens through which we can move data,...
-
6
MosaicML debuts inference service to make generative AI deployment affordable
-
6
Cloudflare R2 and MosaicML enable training LLMs on any compute, anywhere in the world, with zero switching costs 05/16/2023
-
6
MosaicML challenges OpenAI with its new open-source language model
-
2
This Week in AI: Databricks’ Acquisition of MosaicML
-
5
AIGC领域最大收购:Databricks 13亿美元买下MosaicML,成立仅2年员工60人 硅星人 发表于 2023年06月27日...
-
4
Databricks is acquiring MosaicML for a jaw-dropping $1.3 billion
-
9
Databricks and MosaicML CEOs say $1.3 billion deal is about helping enterprises create AI IP
-
5
2023-07-25 08:05 AI 公司 MosaicML 推出 70 亿参数模型 MPT-7B-8K,号称“一次处理 8000 字长文本、可商用” 据 IT 之家 7 月 25 日消息,AI 创业公司 MosaicML 近日发布了其 70 亿参数模型 MPT-7B-8K。据悉,该模型一次可以处理 80...
-
7
US Looking Into Nvidia’s AI Chips for Chin...
About Joyk
Aggregate valuable and interesting links.
Joyk means Joy of geeK