Ranking the best performing large language models for digital product development.
LLM Benchmarks April 2025

The highlights from the world of language models in April:
- New OpenAI models - o3-mini, o4-mini and GPT-4.1
- Qwen3 - pushing state-of-the-art for local models
- Google Is Breathing Down the Neck: Gemini Flash 2.5 Preview and Pro 2.5 v2
- Practical insights from AI Coding
Model Scores
# | Model | bi | compliance | code | reason | Score | Local | Features |
---|---|---|---|---|---|---|---|---|
1 | openai/o3-mini-2025-01-31 | 45% | 70% | 100% | 74% | 76.7% | SO, Reason | |
2 | openai/o4-mini-2025-04-16 | 45% | 70% | 100% | 74% | 76.7% | SO, Reason | |
3 | google/gemini-2.5-flash-preview:thinking | 45% | 57% | 100% | 68% | 71.2% | Reason | |
4 | google/gemini-2.5-pro-preview-03-25 | 45% | 70% | 93% | 68% | 71.1% | Reason | |
5 | qwen/qwen3-32b | 54% | 40% | 96% | 68% | 71.1% | ✓ | Reason, Open |
6 | anthropic/claude-3.7-sonnet:thinking | 54% | 32% | 100% | 67% | 70.4% | Reason | |
7 | openai/o1-2024-12-17 | 45% | 70% | 84% | 67% | 70.0% | SO, Reason | |
8 | openai/gpt-4.1-2025-04-14 | 45% | 70% | 77% | 67% | 67.2% | SO | |
9 | deepseek/deepseek-r1 | 27% | 64% | 100% | 63% | 66.1% | ✓ | SO, Reason, Open |
10 | google/gemini-2.5-pro-preview-05-06 | 45% | 70% | 80% | 65% | 65.6% | Reason | |
11 | qwen/qwen3-30b-a3b | 45% | 37% | 96% | 61% | 65.0% | ✓ | Reason, Open |
12 | qwen/qwen3-235b-a22b | 36% | 45% | 100% | 59% | 62.8% | ✓ | Reason, Open |
13 | openai/gpt-4.1-mini-2025-04-14 | 36% | 80% | 63% | 60% | 61.1% | SO | |
14 | deepseek/deepseek-r1-distill-llama-70b | 36% | 32% | 96% | 56% | 60.0% | ✓ | Open |
15 | deepseek/deepseek-chat-v3-0324 | 45% | 60% | 70% | 55% | 59.6% | ✓ | Reason, Open |
16 | google/gemini-2.5-flash-preview | 45% | 60% | 70% | 58% | 59.4% | ||
17 | anthropic/claude-3.7-sonnet | 45% | 47% | 65% | 55% | 56.5% | ||
18 | qwen/qwen3-14b | 27% | 15% | 100% | 52% | 56.1% | ✓ | Reason, Open |
19 | openai/gpt-4o-2024-11-20 | 36% | 55% | 62% | 55% | 53.6% | SO | |
20 | openai/gpt-4.5-preview-2025-02-27 | 45% | 47% | 62% | 53% | 51.9% | SO | |
21 | deepseek-v3 | 36% | 47% | 58% | 49% | 50.6% | ✓ | SO, Open |
22 | openai/gpt-4o-2024-08-06 | 18% | 62% | 63% | 52% | 50.5% | SO | |
23 | mistralai/mistral-medium-3 | 36% | 35% | 70% | 45% | 49.9% | SO, Reason | |
24 | microsoft/phi-4 | 36% | 62% | 57% | 48% | 49.7% | ✓ | Open |
25 | meta-llama/llama-4-maverick | 27% | 42% | 70% | 44% | 49.1% | ✓ | SO, Open |
26 | qwen/qwen-max | 45% | 45% | 45% | 50% | 46.3% | ||
27 | google/gemma-3-27b-it | 27% | 27% | 70% | 43% | 45.0% | ✓ | Open |
28 | anthropic/claude-3.5-sonnet | 36% | 32% | 57% | 44% | 43.6% | ||
29 | meta-llama/llama-3.1-70b-instruct | 36% | 50% | 44% | 43% | 42.6% | ✓ | SO, Open |
30 | meta-llama/llama-3.3-70b-instruct | 27% | 50% | 48% | 41% | 40.8% | ✓ | SO, Open |
31 | google/gemini-2.0-flash-001 | 27% | 24% | 57% | 38% | 40.7% | ||
32 | qwen/qwq-32b | 36% | 52% | 41% | 37% | 40.0% | ✓ | SO, Reason, Open |
33 | qwen/qwen-2.5-72b-instruct | 27% | 30% | 47% | 39% | 39.2% | ✓ | SO, Open |
34 | mistralai/mistral-small-3.1-24b-instruct | 36% | 42% | 41% | 39% | 39.2% | ✓ | SO, Open |
35 | openai/gpt-4.1-nano-2025-04-14 | 9% | 32% | 64% | 32% | 37.7% | SO | |
36 | qwen/qwen2.5-32b-instruct | 27% | 20% | 53% | 36% | 36.6% | ✓ | Open |
37 | qwen/qwen-2.5-coder-32b-instruct | 18% | 35% | 54% | 39% | 36.5% | ✓ | SO, Open |
38 | meta-llama/llama-3.1-405b-instruct | 18% | 55% | 40% | 38% | 35.5% | ✓ | SO, Open |
39 | google/gemma-3-12b-it | 9% | 17% | 61% | 30% | 33.4% | ✓ | Open |
40 | qwen/qwen-plus | 18% | 25% | 40% | 31% | 31.7% | ||
41 | google/gemma-3-12b-it-qat-q4_0-gguf | 18% | 47% | 34% | 24% | 30.6% | ✓ | SO, Open |
42 | mistralai/mixtral-8x22b-instruct | 9% | 27% | 47% | 28% | 29.2% | ✓ | SO, Open |
43 | openai/gpt-4o-mini-2024-07-18 | 9% | 32% | 41% | 30% | 28.4% | SO | |
44 | mistral/mistral-small-24b-instruct-2501 | 27% | 22% | 33% | 30% | 27.8% | ✓ | SO, Open |
45 | qwen/qwen-turbo | 0% | 15% | 41% | 20% | 21.9% | ||
46 | deepseek/deepseek-r1-distill-qwen-32b | 9% | 22% | 29% | 17% | 21.2% | ✓ | SO, Open |
47 | meta-llama/llama-4-scout | 9% | 25% | 22% | 16% | 18.0% | ✓ | SO, Open |
48 | mistral/ministral-8b | 18% | 0% | 20% | 13% | 14.8% | ✓ | SO, Open |
49 | meta-llama/llama-3.2-3b-instruct | 0% | 17% | 16% | 11% | 10.6% | ✓ | SO, Open |
50 | mistralai/mistral-large-2411 | 0% | 0% | 0% | 0% | 0.0% | ✓ | SO, Open |
51 | ByteDance-Seed/Seed-Coder-8B-Reasoning | 0% | 0% | 0% | 0% | 0.0% | ✓ | SO, Reason, Open |
Averages | 29% | 41% | 60% | 44% |
New OpenAI models: o3-mini, o4-mini, GPT-4.1
New models from OpenAI landed nicely in our reasoning leaderboard. o3-mini and o4-mini scored at the top, while less expensive 4.1 (base and mini) took 8th and 13th places. There are not that many surprises here, except for the cost of running these models.

Qwen3 - pushing state-of-the-art for local models
Qwen models are a well-known “secret” of teams that need to run LLM on their premises. We heard about successful uses of Qwen 2.5 in many projects. This was represented by relatively high scores on our reasoning benchmark.
Enterprise-Grade Accuracy, Even for Complex Tasks
However the latest release of Qwen3 models has pushed state-of-the-art even further. Qwen3 models come in a variety of plain models and Mixture-of-Expert models. They may take longer to process problems, but their accuracy is comparable to the best cloud models. This makes them suitable for enterprise jobs that require high accuracy and can run over night.
- Qwen3 32B
32.8B parameters, 32k native context and 131k extended context (download on Huggingface) - Qwen3-30B-A3B
30.5B parameters, but only 3.3B are activated per token (leading to faster inference) (download on Huggingface)

Qwen3 release pushes state-of-the-art in local language models and creates more pressure for the cloud providers. Things don’t stop there, though.
Google Is Breathing Down the Neck: Gemini Flash 2.5 Preview and Pro 2.5 v2
We all remember how Google was late in the game in the past. OpenAI was much better than first Gemini models, while praised Gemini Ultra - never came out.
Google has managed to gradually shift the tides since then. While it still doesn’t beat the top OpenAI models, it consistently delivers new models within the TOP-10 range, while beating OpenAI on price and stability.

Gemini-2.5-Flash Preview in thinking mode is currently the best Google model on our benchmark. Gemini 2.5 Pro Preview scores 4th and 10th places. Gemini 2.5 Flash in non-thinking mode is at 16th place.
However, benchmarks are only an approximation of the real world. So let us give you two additional data points:
First of all, experienced engineers (with access to Mistral / Anthropic / ChatGPT Pro and Gemini) are starting to gradually rely more on Gemini 2.5 Pro as their tool of choice in AI Coding. They praise quality of the answers and reliable work with larger context. For example, it is common to copy essential chunk of the source code into the prompt (up to 50k tokens) and then work interactively with that chat until the overall context reaches 200k-500k.
Another example is about complex document understanding. Within the last months we’ve heard from our peers in AI Research that “if you need to handle enterprise documents reliably - use Google Gemini LLM”.
We went on to verify that. In one of the evaluations, we load a compliance PDF into a Graph database for further analysis. The tricky part is that such documents can span hundreds of pages, so they have to be loaded page by page. Another tricky part is that the documents are deeply nested, and it takes extra care to properly attribute clauses across the page breaks.

Models from Anthropic, Mistral, and even OpenAI fail to handle the task of “stitching” content across pages. They get confused easily even within a small context, misrepresenting the original text.
Google Gemini 2.5 Pro, on the other hand, handles the same task (given the same inputs) reliably. This makes it a good contender to OpenAI models when implementing document-heavy AI workflows.
Practical insights from AI Coding
Embrace AI Initiative
As a part of the development program at TIMETOACT GROUP Austria we are running Embrace AI initiative - an experiment in training senior-level software engineers to better understand and use modern AI coding tools. This program is also integrated into our AI Research process, helping to strengthen it with insights from the wider community of experts.
As we’ve mentioned earlier, the first insight from this process is that practitioners are gradually shifting from OpenAI and Sonnet 3.7 models to Google Gemini 2.5 Pro in their everyday tasks. This affects even those developers that used to stand by Sonnet 3.5 for a long time.
The reason for this shift is the effective combination of context size (the amount of context the model can reliably handle), speed, quality, and cost offered by Gemini models.

This shift is not mainstream - it affects only experienced engineers that are not bound to a specific AI provider and are free to choose the best tools for the job.
Second insight involves tooling - chats vs complex multi-agent IDEs and environments. To put things into the context, let us provide you with an example of one exercise from our Embrace AI program:
Example of one exercise from the Embrace AI program
Please implement a WebUI tool that can send a request to OpenAI/Gemini/LLM of your choice, while appending contents of files into the prompt. You should be able to pick the files that you want to append.
Requirements:
Tool is passed a directory as argument on startup (e.g. `node server.js ../../projects/demo-project`)
When loading, it will list all files (recursively) in the left pane
When user clicks on the file - it is added to the right pane
If user clicks on the file in the right pane - it is removed
When user types a prompt and clicks on “Submit”, then text of these files is appended to the prompt and sent to LLM.
Response from LLM is streamed back.
NOT Requirements:
multi-turn chat conversations or follow-up questions.
any persistence. Reloading the page may lose all information
Software engineers were allowed to use any combination of tooling to accomplish this, starting from chats to the most advanced IDEs and Coding agents. We allowed them to experiment, only asking to measure the time.
The fastest solution involved using Anthropic Claude with tools to access powershell and directory. It took 30 minutes and used quite a lot of prompts and tokens under the hood.

Other solutions took 1.5h to 2h to implement the code, they involved human control and a bit of prodding the agents.
Writing the same code from the scratch without AI would probably take a few hours, so the best solution with AI provided a noticeable productivity boost. Median solution with AI took the same time it would take an experienced developer to code this app.
However, then we flipped the question and pushed the challenge further:
"Write a prompt, that, once copied to any decent chat LLM model, will produce the code implementing the tool from previous kata straight away (without any agents or complex IDEs? It has to work reliably and across multiple models range of models.
The purpose of the challenge is to build an intuition about how models work and why sometimes you don't need complex tools/agents to accomplish a task."
The purpose of this exercise was to demonstrate that modern AI tooling gets in the way more often than not, and that understanding the fundamentals of coding with AI can let engineers get better results much faster.
Our own solution to the challenge took 15 minutes and required 2 steps:
- Copy-paste kata to ChatGPT o1 pro, request to ask questions clarifying the task, then capture in succinct prompt. This pattern is called Prompt Distillation in our books.
- The resulting prompt was 432 tokens (or 1833 characters) long. It produced working solution from the first try, when copied to GPT-o1 pro, GPT 4.5 and Claude 3.7.

Another engineer used Gemini 2.5 Pro for Prompt Distillation, arriving to the similar result in 15 minutes (although his final Gemini-generated prompt was 1191 tokens or 4598 characters).
This looks pretty impressive, right? Our AI Research community took the challenge to heart and managed to push it even further.
One engineer manually wrote a prompt that is just 70 tokens (313 characters) long and worked on Gemini 2.5 Pro and Sonnet 3.7. It also worked almost out-of-the-box (required one follow-up question) on GPT-4o.
The prompt looked like this
Another engineer wrote in 7 minutes a prompt (194 tokens+298 custom instructions) that produced valid result on Claude 3.7.
Our conclusion:
- Modern AI Coding tools can bring real productivity gains, when used properly.
- Practice and experience can beat fancy tooling. In fact, complex AI tools can often get in the way of productivity.
- Claude 3.7 Sonnet and Gemini 2.5 Pro are currently preferred by the AI Coding practitioners.
Turning AI Potential into Business Impact
From first insights to enterprise-ready solutions — we help you turn AI into real business value.