Delivering new AI technologies at scale also means rethinking every layer of our infrastructure – from silicon and software systems and even our data center designs.
For the second year in a row, Meta’s engineering and infrastructure teams returned for the AI Infra @ Scale conference, where they discussed the challenges of scaling up an infrastructure for AI as well as work being done on our large-scale GPU clusters, open hardware designs for next-generation data center hardware, and how Meta is building custom silicon like the Meta Training and Inference Accelerator (MTIA) to handle some of our AI training workloads.
Aparna Ramani, VP of Engineering at Meta, responsible for AI infrastructure, data infrastructure and developer infrastructure, delivered the opening keynote at AI Infra @Scale 2024 and discussed the AI landscape up to today, the technical challenges, and how solutions like open models and hardware can push AI to new frontiers.
Watch the full keynote presentation below: