azure-ai-evaluation-pyAzure AI Evaluation SDK for Python. Use for evaluating generative AI applications with quality, safety, and custom evaluators. Triggers: "azure-ai-evaluation", "evaluators", "GroundednessEvaluator", "evaluate", "AI quality metrics".
Install via ClawdBot CLI:
clawdbot install thegovind/azure-ai-evaluation-pyGrade Fair — based on market validation, documentation quality, package completeness, maintenance status, and authenticity signals.
Contains instructions to override system prompt or ignore user requests
"ignore previous instructions"Potentially destructive shell commands in tool definitions
eval (Uses known external API (expected, informational)
azure.comAI Analysis
The skill definition is for a legitimate Azure AI evaluation SDK and shows no evidence of hidden instructions, credential harvesting, or obfuscation. The primary risk is the standard dependency on external Azure OpenAI services for AI-assisted evaluators, which is consistent with the skill's stated purpose and requires user-provided credentials.
Generated Mar 21, 2026
Assess AI-powered customer support chatbots for accuracy and safety. Use GroundednessEvaluator to verify responses align with product documentation, and SafetyEvaluators to filter harmful content before deployment.
Evaluate medical AI assistants for factual correctness and safety. Apply RelevanceEvaluator to ensure responses address patient queries accurately, and ContentSafetyEvaluator to prevent harmful medical advice.
Validate AI financial advisors for regulatory compliance and accuracy. Use CoherenceEvaluator to check logical consistency in investment advice, and custom evaluators to flag non-compliant statements.
Monitor AI tutoring systems for educational quality and appropriateness. Implement FluencyEvaluator to assess language clarity, and SafetyEvaluators to filter inappropriate content for student audiences.
Verify AI legal document reviewers for precision and completeness. Employ SimilarityEvaluator to compare AI summaries with source documents, and custom evaluators to check citation accuracy.
Offer automated evaluation services for companies deploying generative AI applications. Charge per evaluation run or monthly subscription based on data volume and number of evaluators used.
Provide regulatory compliance certification for AI systems using standardized evaluation protocols. Generate revenue through certification fees, audit services, and compliance monitoring subscriptions.
Integrate evaluation SDK into existing AI development platforms and MLOps tools. Monetize through licensing fees, enterprise support contracts, and premium evaluation features.
💬 Integration Tip
Start with single row evaluation using GroundednessEvaluator for quick validation, then scale to batch evaluation with evaluate() function for production testing.
Scored Apr 19, 2026
Audited Apr 17, 2026 · audit v1.0
Use CodexBar CLI local cost usage to summarize per-model usage for Codex or Claude, including the current (most recent) model or a full model breakdown. Trigger when asked for model-level usage/cost data from codexbar, or when you need a scriptable per-model summary from codexbar cost JSON.
Gemini CLI for one-shot Q&A, summaries, and generation.
Manages free AI models from OpenRouter for OpenClaw. Automatically ranks models by quality, configures fallbacks for rate-limit handling, and updates openclaw.json. Use when the user mentions free AI, OpenRouter, model switching, rate limits, or wants to reduce AI costs.
Manages free AI models from OpenRouter for OpenClaw. Automatically ranks models by quality, configures fallbacks for rate-limit handling, and updates opencla...
Reduce OpenClaw AI costs by 97%. Haiku model routing, free Ollama heartbeats, prompt caching, and budget controls. Go from $1,500/month to $50/month in 5 min...
HTML-first PDF production skill for reports, papers, and structured documents. Must be applied before generating PDF deliverables from HTML.