Zuckerberg says Meta will need 10x more computing power to train Llama 4 than Llama 3 | TechCrunch
๐ Abstract
The article discusses Meta's plans to significantly increase the computing power needed to train its large language model, Llama, in the future. It highlights Meta's CEO Mark Zuckerberg's comments on the growing compute requirements for training Llama 4, which is expected to be 10 times more than what was needed for Llama 3. The article also covers Meta's efforts to build capacity for training future AI models, as well as the company's focus on consumer-facing AI products like its chatbot.
๐ Q&A
[01] Meta's Plans for Llama 4 and Future Models
1. What are Meta's plans for training Llama 4 and future language models?
- Meta CEO Mark Zuckerberg stated that the computing power needed to train Llama 4 will likely be 10 times more than what was used for Llama 3.
- Future models beyond Llama 4 are expected to continue growing in their compute requirements.
- Meta is planning to build capacity for training these future AI models, rather than falling behind its competitors.
2. Why does Meta want to build capacity for training future AI models?
- Zuckerberg said he would rather "risk building capacity before it is needed rather than too late, given the long lead times for spinning up new inference projects."
- Meta's CFO, Susan Li, mentioned that the company is thinking about different data center projects and building capacity to train future AI models.
- This investment in infrastructure is expected to increase Meta's capital expenditures in 2025.
[02] Meta's Investments in AI Infrastructure
1. What is the current state of Meta's investments in AI infrastructure?
- Meta's capital expenditures rose nearly 33% to $8.5 billion in Q2 2024, from $6.4 billion a year earlier, driven by investments in servers, data centers, and network infrastructure.
- According to a report, OpenAI spends $3 billion on training models and an additional $4 billion on renting servers at a discount rate from Microsoft.
- Meta's CFO, Susan Li, stated that the company will continue to build its infrastructure in a way that provides flexibility in how the capacity is used, allowing it to direct training to either generative AI inference or its core ranking and recommendation work.
[03] Meta's Consumer-facing AI Products
1. How is Meta's consumer-facing AI product, Meta AI, performing?
- During the earnings call, Meta mentioned that India is the largest market for its chatbot, which is part of the Meta AI product.
- However, Meta's CFO, Susan Li, noted that the company does not expect its generative AI products to contribute significantly to revenue in the near future.