Skip to main content
Back to registry

evaluating-code-models

davila7/claude-code-templates

BigCode Evaluation Harness evaluates code generation models across 15+ benchmarks including HumanEval, MBPP, and MultiPL-E (18 languages).

Installs174
Install command
npx skills add https://github.com/davila7/claude-code-templates --skill evaluating-code-models
Security audits
Gen Agent Trust HubFAIL
SocketPASS
SnykWARN
Community Reviews

Latest reviews

Sign in to review

No community reviews yet. Be the first to review.

Browse this skill in context
FAQ
What does evaluating-code-models do?

BigCode Evaluation Harness evaluates code generation models across 15+ benchmarks including HumanEval, MBPP, and MultiPL-E (18 languages).

Is evaluating-code-models good?

evaluating-code-models does not have approved reviews yet, so SkillJury cannot publish a community verdict.

What agent does evaluating-code-models work with?

evaluating-code-models currently lists compatibility with codex, gemini-cli, opencode, cursor, github-copilot, claude-code.

What are alternatives to evaluating-code-models?

Skills in the same category include telegram-bot-builder, flutter-app-size, sharp-edges, iterative-retrieval.

How do I install evaluating-code-models?

npx skills add https://github.com/davila7/claude-code-templates --skill evaluating-code-models

Related skills

More from davila7/claude-code-templates

Related skills

Alternatives in Software Engineering