Skip to main content
Back to the directory
github/awesome-copilotSoftware EngineeringFrontend and Design

arize-evaluator

This skill covers designing, creating, and running LLM-as-judge evaluators on Arize. An evaluator defines the judge; a task is how you run it against real data.

SkillJury keeps community verdicts, source metadata, and external repository signals in separate lanes so ranking data never pretends to be a review.

SkillJury verdict
Pending

No approved reviews yet

Would recommend
Pending

Waiting on enough review volume

Install signal
756

Weekly or total install activity from catalog data

Sign in to review
0 review requests
Install command
npx skills add https://github.com/github/awesome-copilot --skill arize-evaluator
SkillJury does not have enough approved reviews to publish a community verdict yet. Source metadata and repository proof are still available above.
SkillJury Signal Summary

As of Apr 30, 2026, arize-evaluator has 756 weekly installs, 0 community reviews on SkillJury. Community votes currently stand at 0 upvotes and 0 downvotes. Source: github/awesome-copilot. Canonical URL: https://skills.sh/github/awesome-copilot/arize-evaluator.

Security audits
Gen Agent Trust HubPASS
SocketWARN
SnykPASS
About this skill
This skill covers designing, creating, and running LLM-as-judge evaluators on Arize. An evaluator defines the judge; a task is how you run it against real data. Proceed directly with the task — run the ax command you need. Do NOT check versions, env vars, or profiles upfront. If an ax command fails, troubleshoot based on the error: An evaluator is an LLM-as-judge definition. It contains: Evaluators are versioned — every prompt or model change creates a new immutable version. The most recent version is active. A task is how you run one or more evaluators against real data.

Source description provided by the upstream listing. Community review signal and install context stay separate from this narrative layer.

Community reviews

Latest reviews

No community reviews yet. Be the first to review.

Browse this skill in context
FAQ
What does arize-evaluator do?

This skill covers designing, creating, and running LLM-as-judge evaluators on Arize. An evaluator defines the judge; a task is how you run it against real data.

Is arize-evaluator good?

arize-evaluator does not have approved reviews yet, so SkillJury cannot publish a community verdict.

Which AI agents support arize-evaluator?

arize-evaluator currently lists compatibility with Claude Code, Codex, Gemini CLI, Skills CLI.

Is arize-evaluator safe to install?

arize-evaluator has been scanned by security audit providers tracked on SkillJury. Check the security audits section on this page for detailed results from Socket.dev and Snyk.

What are alternatives to arize-evaluator?

Skills in the same category include grimoire-morpho-blue, conversation-memory, second-brain-ingest, zai-tts.

How do I install arize-evaluator?

Run the following command to install arize-evaluator: npx skills add https://github.com/github/awesome-copilot --skill arize-evaluator

Related skills

More from github/awesome-copilot

Related skills

Alternatives in Software Engineering