TLDR: We are hiring! Specifically, we are hiring PhDs graduating in 2025, professionals with industry experience, and managers. We are NOT hiring new graduates with a bachelor's or master's degree. Our summer internship positions have been filled. Candidates please email us at "codesign AT meta DOT com".
The AI and Systems Co-Design team at Meta (formerly known as Facebook), led by Chunqiang Tang (a.k.a. CQ Tang), consists of over 100 employees, mostly PhDs, including many world-class research scientists and engineers. As reflected in our team name "co-design", we conduct interdisciplinary research and development across AI, hardware, and software, with a focus on performance, efficiency, and scalability.
- We own the company's overall strategy for exploring innovative hardware technologies for CPUs, GPUs, memory, storage, and Meta's custom AI chips, and we productionize them in Meta's hyperscale fleet of O(1,000,000) servers and O(100,000) GPUs, powering all Meta products such as Facebook, Instagram, and meta.ai.
- We apply novel software optimizations across the whole stack---from ML models and applications to the Linux kernel---to achieve optimal performance on the hardware.
- We develop innovative AI technologies for large language models (Llama), ranking systems, and more.
Overall, our work largely corresponds to the research communities of systems in general and especially systems for ML (MLSys, SOSP, OSDI, SIGCOMM, NSDI), hardware architecture (ISCA, ASPLOS), ML (NeurIPS, ICML, ICLR) and supercomputing (SC, ICS). Here are selected publications that showcase our work in diverse areas:
HPC and collective communications library (MPI, NCCL, RCCL)
Performance benchmarking and projection for both AI and non-AI workloads
Hardware and software co-design
Like research labs, our team consists primarily of PhDs, and we strongly encourage and excel in research publications. However, we differ from traditional research labs in several key ways:
- Direct ownership: Like traditional research labs, we build strong partnerships with numerous teams across diverse areas for broad influence. However, what sets us apart is our direct ownership of the hardware strategy for Meta's hyperscale fleet. This enables us to lead in many areas while fostering seamless partnerships in others.
- Production systems: Our primary goal is to develop forward-looking innovations in AI, hardware, and software, and directly implement them in production systems that serve billions of people. The billions of users of Meta products and Meta's hyperscale fleet of O(1,000,000) servers and O(100,000) GPUs are, in effect, our lab. In contrast, traditional research labs often rely on technology transfer for a less direct impact.
- Impact: Our impact is widely acknowledged within the company and throughout the industry. We drive Meta's hardware strategy to save billions of dollars, and directly develop innovative technologies in Meta's flagship products like Llama and Ads ranking models.
Open Source Projects
- DCPerf: An open source benchmark suite for hyperscale compute applications
- DLRM: An advanced, open source deep learning recommendation model
- FBGEMM: ML kernels
- Pytorch distributed Shampoo optimizer. This work won the competition of the external tuning track of the inaugural AlgoPerf training algorithms benchmark.
Selected Publications
2024
2023
2022
2021
2020
2019