Zubnet AIAprenderWiki › AI Benchmarks
Fundamentos

AI Benchmarks

MMLU, HumanEval, ARC, HellaSwag
Testes padronizados usados para medir e comparar capacidades de modelos IA. MMLU testa conhecimento através de 57 matérias acadêmicas. HumanEval testa geração de código. ARC testa raciocínio científico. HellaSwag testa raciocínio de senso comum. GSM8K testa matemática. Pontuações de benchmark provêem uma linguagem comum para comparar modelos, embora tenham limitações significativas.

Por que importa

Benchmarks são como a indústria marca o placar. Quando a Anthropic diz que o Claude pontua X% no MMLU e Y% no HumanEval, esses números só significam algo se você sabe o que os benchmarks testam, como são pontuados e quais são suas limitações. Entender benchmarks te ajuda a cortar através de afirmações de marketing e avaliar qual modelo é realmente melhor para seu caso de uso específico.

Deep Dive

Key benchmarks: MMLU (Massive Multitask Language Understanding) — 14,042 multiple-choice questions across 57 subjects from STEM to humanities. HumanEval — 164 coding problems testing function generation in Python. ARC (AI2 Reasoning Challenge) — science exam questions requiring reasoning. HellaSwag — sentence completion testing commonsense knowledge. GSM8K — 8,500 grade-school math word problems. Each tests a different capability.

Why Benchmarks Are Problematic

Several issues: contamination (test questions appear in training data, inflating scores), saturation (when all models score 95%+, the benchmark stops discriminating), gaming (training specifically to maximize benchmark scores without genuinely improving capability), and narrow coverage (benchmarks test what's testable, not necessarily what matters to users). A model that scores 90% on MMLU might be worse at actually helping a user than one that scores 80% but follows instructions better.

The Move Beyond Static Benchmarks

The field is evolving: Chatbot Arena uses real-time human preferences (hard to contaminate, always current). LiveBench uses frequently refreshed questions. SEAL and other private benchmarks keep test data secret. Task-specific evaluations (SWE-bench for real GitHub issue solving, GPQA for PhD-level science) test capabilities that general benchmarks miss. The trend is toward evaluation that looks more like real-world use and less like standardized testing.

Conceitos relacionados

← Todos os termos
← AGI AI Coding Assistants →