IBM and AMD have announced a significant collaboration to provide advanced artificial intelligence infrastructure to Zyphra, a San Francisco-based AI research company. Under a new multi-year agreement, IBM will deliver a large computing cluster featuring AMD's latest technology on IBM Cloud, enabling Zyphra to train sophisticated AI models.
This partnership marks one of the largest generative AI training systems to date built on an AMD hardware stack and hosted on IBM's cloud platform. The infrastructure is designed to support Zyphra's work in developing next-generation open-source AI.
Key Takeaways
- IBM and AMD have partnered to supply AI infrastructure to the open-source AI company Zyphra.
- Zyphra will use a large cluster of AMD Instinct MI300X GPUs hosted on IBM Cloud.
- The collaboration will support the training of advanced multimodal foundation models.
- Zyphra, recently valued at $1 billion, aims to develop a general-purpose AI assistant named Maia.
A New AI Training Infrastructure Partnership
The collaboration between the three companies centers on providing substantial computational power for AI development. IBM is set to supply Zyphra with a dedicated, large-scale cluster of AMD Instinct™ MI300X GPUs through its cloud services. This arrangement provides Zyphra with the necessary resources to train what are known as frontier multimodal foundation models.
These models are designed to understand and process information from multiple sources, such as text, images, and audio, simultaneously. The scale of this deployment is notable, representing a major commitment by all three companies to advance AI research. According to the announcement, the initial hardware was made available to Zyphra in early September, with further expansion planned for 2026.
What is a Multimodal Foundation Model?
A multimodal foundation model is a type of large-scale artificial intelligence that can process and integrate information from different data types, like text, images, and sound. Unlike models that only understand language, these can make connections between a picture and its description or a sound and its source, leading to more comprehensive and human-like understanding.
Zyphra's Vision for Open-Source Superintelligence
Zyphra is an AI research and product company focused on building open-source superintelligence. The company recently secured a Series A funding round that brought its valuation to $1 billion. Its research is concentrated on creating novel neural network architectures, improving long-term memory in AI systems, and enabling continual learning capabilities.
The company plans to use the new IBM and AMD infrastructure to develop Maia, a general-purpose AI assistant. The goal for Maia is to enhance productivity for knowledge workers in enterprise environments by handling complex tasks across different data formats.
"This collaboration marks the first time AMD's full-stack training platform—spanning compute through networking—has been successfully integrated and scaled on IBM Cloud, and Zyphra is honored to lead the way in developing frontier models with AMD silicon on IBM Cloud," stated Krithik Puthalath, CEO and Chairman of Zyphra.
Puthalath added, "We're excited to partner with IBM and AMD to power the next era of open-source, enterprise superintelligence."
The Technology Powering the Initiative
The agreement represents the first large, dedicated training cluster on IBM Cloud that leverages a complete AMD hardware solution. This includes not only the powerful AMD Instinct MI300X GPUs but also advanced networking components.
Hardware Stack Components
- Compute: AMD Instinct™ MI300X GPUs, designed specifically for AI and high-performance computing workloads.
- Networking: AMD Pensando™ Pollara 400 AI NICs (Network Interface Cards) and AMD Pensando Ortano DPUs (Data Processing Units) to manage data flow efficiently between the GPUs.
This integrated stack is designed to provide high performance and efficiency, which are critical for training massive AI models that can take weeks or months and consume significant energy. The ability to scale these resources is a key reason Zyphra cited for choosing the partnership.
Strategic Implications for IBM and AMD
For IBM and AMD, this collaboration serves as a major validation of their joint efforts in the competitive AI hardware and cloud services market. Last year, the two companies announced plans to offer AMD's MI300X accelerators as a service on IBM Cloud, targeting enterprise clients with demanding AI and high-performance computing (HPC) needs.
Alan Peacock, General Manager of IBM Cloud, emphasized the importance of efficient AI workload scaling. "We are delighted to support Zyphra's strategic roadmap as we collaborate with AMD to deliver scalable, economical AI infrastructure that can accelerate Zyphra's model training," he said.
This partnership showcases the ability of IBM Cloud to support cutting-edge AI research while highlighting the performance of AMD's comprehensive AI hardware platform. It positions them as a strong alternative for companies and researchers seeking powerful and scalable AI training solutions.
"By combining IBM enterprise cloud expertise with AMD leadership in high-performance computing and AI acceleration, we are supporting Zyphra's pioneering work in multimodal and inference-efficient AI," said Philip Guido, EVP and Chief Commercial Officer at AMD. He noted that the collaboration helps organizations build smarter businesses and unlock AI solutions.
Broader Collaboration Efforts
Beyond this specific project, IBM and AMD are also working on longer-term initiatives. The companies recently revealed plans to develop next-generation computing architectures that combine classical supercomputing with quantum computing. This research aims to leverage IBM's leadership in quantum computers with AMD's expertise in high-performance accelerators, pushing the boundaries of scientific computation even further.





