Thu. Apr 17th, 2025 8:00:00 PM
Meta's vanilla maverick ai model ranks under rivals in the

Earlier this week, Meta landed in warm water to achieve a high score on the crowdsourced benchmark LM Arena using an experimental, unpublished version of the Llama 4 Maverick model. Incident I apologized to the LM Arena maintainer.change the policy and get the unchanged vanilla maverick.

After all, it’s not very competitive.

Unchanged Maverick, “llama-4-maverick-17b-128e-instruct” Ranked below model Includes Openai’s GPT-4O, Anthropic’s Claude 3.5 Sonnet, and Google’s Gemini 1.5 Pro on Friday. Many of these models were a few months ago.

Why is the performance poor? Meta’s experimental Maverick, Lama-4-Maverick-03-26-experience, was “optimized for conversation,” the company explained. Published charts Last Saturday. These optimizations clearly worked well for LM arenas where human evaluators compare the outputs of the models and select what they like.

As I wrote before, for a variety of reasons, LM arena was not the most reliable measure of AI models’ performance. Still, tuning your model to your benchmark is not only misleading, but it also makes it difficult for developers to accurately predict how well a model will work in different contexts.

In a statement, a Meta spokesperson told TechCrunch that Meta will experiment with “all kinds of custom variants.”

“‘llama-4-maverick-03-26-Experimmal’ is a chat-optimized version that also works well in the LM arena,” the spokesman said. “We are currently releasing an open source version and see how developers can customize Llama 4 for their use cases. We look forward to seeing what they build and ongoing feedback.”

Leave a Reply

Your email address will not be published. Required fields are marked *