Rumor Denied: Meta Executive Addresses Concerns Over AI Model Benchmarking
Meta Denies Training AI Models on Test Sets
A Meta executive has denied a rumor that the company trained its new AI models to present well on specific benchmarks while concealing the models’ weaknesses. Ahmad Al-Dahle, VP of generative AI at Meta, stated in a post on X that it’s "simply not true" that Meta trained its Llama 4 Maverick and Llama 4 Scout models on "test sets." This practice could misleadingly inflate a model’s benchmark scores, making it appear more capable than it actually is.
Origin of the Rumor
The rumor appears to have originated from a post on a Chinese social media site from a user claiming to have resigned from Meta in protest over the company’s benchmarking practices. This unsubstantiated claim was further fueled by reports that Maverick and Scout perform poorly on certain tasks. Additionally, Meta’s decision to use an experimental, unreleased version of Maverick to achieve better scores on the benchmark LM Arena has raised concerns among researchers.
Differences in Model Performance
Researchers on X have observed stark differences in the behavior of the publicly downloadable Maverick compared with the model hosted on LM Arena. This inconsistency has led to questions about the accuracy of the models’ benchmark scores.
Acknowledging Quality Issues
Al-Dahle acknowledged that some users are seeing "mixed quality" from Maverick and Scout across the different cloud providers hosting the models. He attributed this to the models being released as soon as they were ready, stating that it will take several days for all public implementations to get "dialled in." Meta will continue to work on bug fixes and onboarding partners to improve the models’ performance.
Conclusion
In conclusion, Meta has denied the rumor that it trained its AI models to present well on specific benchmarks while concealing their weaknesses. The company is working to address the quality issues and inconsistencies in the models’ performance.
FAQs
Q: What is the rumor about Meta’s AI models?
A: The rumor claims that Meta trained its AI models, Llama 4 Maverick and Llama 4 Scout, on "test sets" to artificially inflate their benchmark scores.
Q: Is this rumor true?
A: No, according to Meta executive Ahmad Al-Dahle, the company did not train its AI models on test sets.
Q: Why are some users seeing mixed quality from Maverick and Scout?
A: The inconsistent performance is due to the models being released as soon as they were ready, and it will take several days for all public implementations to get "dialled in."
Q: What is Meta doing to address the quality issues?
A: Meta is working on bug fixes and onboarding partners to improve the models’ performance.

