Skip to content
#

mbpp

Here are 6 public repositories matching this topic...

Benchmark suite for evaluating LLMs and SLMs on coding and SE tasks. Features HumanEval, MBPP, SWE-bench, and BigCodeBench with an interactive Streamlit UI. Supports cloud APIs (OpenAI, Anthropic, Google) and local models via Ollama. Tracks pass rates, latency, token usage, and costs.

  • Updated Apr 23, 2026
  • Python

Stress-validation of Qwen3.6-27B inference configurations on dual RTX PRO 6000 Blackwell. 5 configs x 4 phases (gates, throughput matrix, HumanEval, MBPP) = 2,105 hard coding problems, zero crashes. Headline: FP8+MTP=3 wins HumanEval (79.3%), BF16+DFlash wins MBPP (89.5%). MTP=5 dominated on correctness despite faster raw tok/s.

  • Updated May 7, 2026
  • Python

Improve this page

Add a description, image, and links to the mbpp topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the mbpp topic, visit your repo's landing page and select "manage topics."

Learn more