A bar chart comparing the performance of deepseek ai and openai models across key benchmarks like aime 2024 codeforces gpqa diamond math 500 mmlu and swe bench verified with deepseek r1 showing strong results
A comparison of DeepSeek AI and OpenAI models across various benchmarks including AIME 2024 Codeforces GPQA Diamond MATH 500 MMLU and SWE bench Verified The results show DeepSeek R1 leading in multiple categories highlighting its competitive performance in AI reasoning mathematics and coding tasks

The image presents a bar chart comparing the benchmark results of various deepseek ai and openai models The benchmarks include

aime 2024 mathematical problem solving accuracy
codeforces competitive programming percentile ranking
gpqa diamond general knowledge and reasoning
math 500 advanced mathematical problem solving
mmlu massive multitask language understanding
swe bench verified software engineering and coding accuracy
the results indicate that deepseek r1 performs competitively surpassing openai models in several categories particularly in math 500 and mmlu reflecting its strengths in logical reasoning and technical problem solving