Model Profile
openai/gpt-4.1
Use this page to decide where this model is a strong fit. Rankings below are benchmark-backed by use case, with explicit confidence and contributor metrics.
Identity
ID: external/openai/gpt-4-1
Author: openai
Origin: external_benchmark_shadow
Arch: unknown
Benchmark Coverage
Scored use cases: 12
Avg confidence: 23.5%
Evidence points: 178
Raw rows: 87
Weighted rows: 25
Catalog Metadata
Parameters: unknown
Context window: 4096
Downloads: 0
Price / 1M tokens: $3.50 (blended 3:1)
Intelligence Profile
Dimension Breakdown
No creativity benchmarks found
No based benchmarks found
* Low confidence — limited benchmark evidence for this dimension
3/5 dimensions scored · Last updated Apr 21, 2026
Benchmark Signals
Click through to the benchmark source behind this model profile.
DuckDB NSQL Leaderboard
all_execution_accuracy
Normalized value 96.2% · confidence 100.0%
Strongest impact in Metric definition workshop
duckdb_nsql_leaderboard.all_execution_accuracy · Apr 1, 2026
LanguageBench Translation Official (Split)
translation_to:bleu
Normalized value 82.7% · confidence 100.0%
Strongest impact in Archaic and historical translation
languagebench_translation_official.translation_to_bleu · Apr 1, 2026
LanguageBench
overall:mean
Normalized value 98.3% · confidence 100.0%
Strongest impact in Archaic and historical translation
languagebench.overall_mean · Apr 1, 2026
SWE-bench Verified Leaderboard
swe_verified_resolved_pct
Normalized value 94.1% · confidence 100.0%
Strongest impact in Verilog/VHDL generation
swebench_verified_official.swe_verified_resolved_pct · Apr 1, 2026
LanguageBench Grammar/Clarity Official (Split)
grammar_clarity_score_pct
Normalized value 95.8% · confidence 100.0%
Strongest impact in Translation and localization
languagebench_grammar_clarity_official.grammar_clarity_score_pct · Apr 1, 2026
Aider Polyglot Leaderboard
percent_correct_pct
Normalized value 88.2% · confidence 100.0%
Strongest impact in Verilog/VHDL generation
aider_polyglot.percent_correct_pct · Apr 1, 2026
Some fit rows have limited benchmark evidence.
9 of 12 scored use cases have low confidence or thin contributor coverage.
Coverage Diagnostics
actively scoredUse-Case Scores
101
Total Measurements
87
Weighted Measurements
25
Weighted Sources
15
Raw Source Coverage
Weighted Source Coverage
Best Use Cases for This Model
| Use Case | Score |
|---|---|
| Archaic and historical translation use_case.history.archaic_translation | 26.2% |
| Legal translation use_case.legal.legal_translation | 24.7% |
| Brand voice localization use_case.mkt.brand_voice_localization | 20.9% |
| Historical document summarization use_case.history.historical_doc_summarization | 20.6% |
| Verilog/VHDL generation use_case.eda.verilog_generation | 19.6% |
| Metric definition workshop use_case.data.metric_definition_workshop | 17.8% |
| Integration test generation use_case.dev.integration_tests | 16.6% |
| Grammar and writing coach use_case.lang.grammar_coach | 16.6% |
| Data quality assistant use_case.data.data_quality_assistant | 16.3% |
| Translation and localization use_case.business.translation_localization | 16.0% |
| Contract term extraction use_case.legal.contract_term_extraction | 15.9% |
| Clause playbook check use_case.legal.playbook_clause_check | 15.9% |