marketing_sales
Best LLM for Sales Outreach
Ranked models for drafting personalized outbound emails tailored to prospect personas.
#1 Recommendation
qwen-2.5-72b-instruct
Strong on Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct and Open LLM Leaderboard GPQA gpqa
external/qwen/qwen-2-5-72b-instruct
29.1%
Score
45.3%
Confidence
15
Evidence
Ranked Models
30
Evidence Quality
97%
Evidence Points
15
Top Signal
Open LLM Leaderboard MMLU-Pro: mmlu_pro_accuracy_pct
Benchmark Sources
35
Last Updated
6h ago
All Ranked Models
| Rank | Model | Score |
|---|---|---|
| 🥇 | qwen-2.5-72b-instruct Strong on Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct and Open LLM Leaderboard GPQA gpqa | 29.1% |
| 🥈 | claude-sonnet-4 Strong on Galileo Agent Leaderboard v2 Avg TSQ and EQ-Bench Leaderboard eq_bench_score | 23.7% |
| 🥉 | Grok-4-0709 Strong on Galileo Agent Leaderboard v2 Avg TSQ and EQ-Bench Leaderboard eq_bench_score | 23.6% |
| #4 | gemini-2.5-pro Strong on EQ-Bench Leaderboard eq_bench_score and Galileo Agent Leaderboard v2 Avg TSQ | 23.0% |
| #5 | Mistral-Large-Instruct-2411 Strong on Open LLM Leaderboard GPQA gpqa and Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct | 21.9% |
| #6 | gpt-5-2025-08-07 Strong on EQ-Bench Leaderboard eq_bench_score and UGI Leaderboard Writing ✍️ | 21.2% |
| #7 | Mixtral-8x22B-Instruct-v0.1 Strong on Open LLM Leaderboard GPQA gpqa and Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct | 20.6% |
| #8 | gemma-2-27b-it Strong on Open LLM Leaderboard GPQA gpqa and Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct | 20.0% |
| #9 | Steelskull/L3.3-MS-Nevoria-70b Strong on Open LLM Leaderboard GPQA gpqa and Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct | 19.8% |
| #10 | o3-20250416 Strong on EQ-Bench Leaderboard eq_bench_score and UGI Leaderboard Writing ✍️ | 19.3% |
| #11 | gpt-4o Strong on CRMArena Function Calling overall_score_pct and EQ-Bench Leaderboard eq_bench_score | 19.2% |
| #12 | Qwen2-72B-Instruct Strong on Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct and Open LLM Leaderboard GPQA gpqa | 19.1% |
| #13 | Sao10K/70B-L3.3-Cirrus-x1 Strong on Open LLM Leaderboard GPQA gpqa and Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct | 18.7% |
| #14 | RYS-XLarge Strong on Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct and Open LLM Leaderboard GPQA gpqa | 18.7% |
| #15 | RYS-XLarge-base Strong on Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct and Open LLM Leaderboard GPQA gpqa | 18.5% |
| #16 | MaziyarPanahi/calme-3.2-instruct-78b Strong on Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct and Open LLM Leaderboard GPQA gpqa | 18.4% |
| #17 | Steelskull/L3.3-Nevoria-R1-70b Strong on Open LLM Leaderboard GPQA gpqa and Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct | 18.3% |
| #18 | Qwen2.5-32B-Instruct Strong on Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct and Open LLM Leaderboard GPQA gpqa | 18.3% |
| #19 | phi-4 Strong on Open LLM Leaderboard GPQA gpqa and Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct | 18.1% |
| #20 | gpt-4.1-20250414 Strong on Galileo Agent Leaderboard v2 Avg TSQ and Galileo Agent Leaderboard v2 Avg AC | 18.0% |
| #21 | MaziyarPanahi/calme-2.4-rys-78b Strong on Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct and Open LLM Leaderboard GPQA gpqa | 17.9% |
| #22 | MaziyarPanahi/calme-3.1-instruct-78b Strong on Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct and Open LLM Leaderboard GPQA gpqa | 17.9% |
| #23 | wizardlm-2-8x22b Strong on Open LLM Leaderboard GPQA gpqa and Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct | 17.9% |
| #24 | Tarek07/Progenitor-V1.1-LLaMa-70B Strong on Open LLM Leaderboard GPQA gpqa and Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct | 17.8% |
| #25 | CalmeRys-78B-Orpo-v0.1 Strong on Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct and Open LLM Leaderboard GPQA gpqa | 17.8% |
| #26 | solar-pro-preview-instruct Strong on Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct and Open LLM Leaderboard GPQA gpqa | 17.5% |
| #27 | Apollo-70B Strong on Open LLM Leaderboard GPQA gpqa and Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct | 17.3% |
| #28 | Triangle104/Set-70b Strong on Open LLM Leaderboard GPQA gpqa and Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct | 17.3% |
| #29 | Homer-v1.0-Qwen2.5-72B Strong on Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct and Open LLM Leaderboard GPQA gpqa | 17.1% |
| #30 | Tarek07/Thalassic-Alpha-LLaMa-70B Strong on Open LLM Leaderboard GPQA gpqa and Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct | 17.1% |
Head-to-Head: #1 vs #2
#1
Top Pickqwen-2.5-72b-instruct
Strong on Open LLM Leaderboard MMLU-Pro mmlu_pro_accuracy_pct and Open LLM Leaderboard GPQA gpqa
Conf 45.3%
#2
anthropic/claude-sonnet-4
Strong on Galileo Agent Leaderboard v2 Avg TSQ and EQ-Bench Leaderboard eq_bench_score
Conf 30.6%
Related Lookups
Best LLM for Code Generation
Benchmark-backed ranking of models for generating correct, secure code from requirements.
Best LLM for Debugging
Find the top-ranked models for localizing bugs and proposing fixes with explanations.
Best LLM for Unit Test Generation
Ranked models for generating meaningful unit tests and edge cases from code.
Best LLM for Code Review
Compare models for automated PR review covering correctness, security, and maintainability.
Best LLM for Autonomous Coding
Benchmark-backed ranking of models for end-to-end autonomous software engineering and issue resolution.
Best LLM for Function Calling
Compare models for reliable tool use, function selection, and multi-step API orchestration.