llm-modelsAccess Claude, Gemini, Kimi, GLM and 100+ LLMs via inference.sh CLI using OpenRouter. Models: Claude Opus 4.5, Claude Sonnet 4.5, Claude Haiku 4.5, Gemini 3...
Install via ClawdBot CLI:
clawdbot install okaris/llm-modelsGrade Fair — based on market validation, documentation quality, package completeness, maintenance status, and authenticity signals.
Calls external URL not in known-safe list
https://inference.shAudited Apr 17, 2026 · audit v1.0
Generated Mar 20, 2026
Deploy a chatbot using Claude Sonnet or Gemini 3 Pro to handle customer inquiries, providing balanced performance for answering questions and resolving issues. This reduces human agent workload and improves response times in industries like e-commerce or SaaS.
Integrate Claude Opus or GLM-4.6 into development workflows to generate, review, and debug code snippets, enhancing productivity for software teams. This is ideal for tech companies seeking to streamline coding tasks and ensure code quality.
Use models like Claude Haiku for fast, economical generation of marketing copy, blog posts, or social media content, helping marketing agencies produce scalable content. This supports industries such as digital marketing and media.
Combine Kimi K2 Thinking with web search skills to build agents that perform multi-step reasoning for market research or data interpretation, useful in consulting or academic settings. This automates complex analysis tasks efficiently.
Offer tiered subscription plans for access to multiple LLMs via OpenRouter, charging based on usage volume or model tiers. This generates recurring revenue from developers and businesses needing scalable AI capabilities.
Charge users per API call or token processed, with cost optimization through automatic fallback to economical models like Claude Haiku. This attracts startups and small projects with flexible, usage-based pricing.
Provide custom integration services, training, and premium support for large organizations deploying AI agents in production workflows. This leverages the skill's capabilities for high-value, long-term contracts.
💬 Integration Tip
Start by installing the inference.sh CLI and testing with the 'any-model' app for automatic cost-effective selections before customizing specific models for advanced use cases.
Scored Apr 19, 2026
Use CodexBar CLI local cost usage to summarize per-model usage for Codex or Claude, including the current (most recent) model or a full model breakdown. Trigger when asked for model-level usage/cost data from codexbar, or when you need a scriptable per-model summary from codexbar cost JSON.
Gemini CLI for one-shot Q&A, summaries, and generation.
Manages free AI models from OpenRouter for OpenClaw. Automatically ranks models by quality, configures fallbacks for rate-limit handling, and updates openclaw.json. Use when the user mentions free AI, OpenRouter, model switching, rate limits, or wants to reduce AI costs.
Manages free AI models from OpenRouter for OpenClaw. Automatically ranks models by quality, configures fallbacks for rate-limit handling, and updates opencla...
Reduce OpenClaw AI costs by 97%. Haiku model routing, free Ollama heartbeats, prompt caching, and budget controls. Go from $1,500/month to $50/month in 5 min...
HTML-first PDF production skill for reports, papers, and structured documents. Must be applied before generating PDF deliverables from HTML.