fleet-embeddingsEmbeddings with nomic-embed-text, mxbai-embed, and snowflake-arctic-embed across your device fleet. Fleet-routed via Ollama for RAG, semantic search, and vec...
Install via ClawdBot CLI:
clawdbot install twinsgeeks/fleet-embeddingsGrade Fair — based on market validation, documentation quality, package completeness, maintenance status, and authenticity signals.
Sends data to undocumented external endpoint (potential exfiltration)
POST → http://localhost:11435/dashboard/api/pullCalls external URL not in known-safe list
https://github.com/geeks-accelerator/ollama-herdAI Analysis
The skill routes embeddings through a local fleet router (localhost:11435) for distributed processing, which aligns with its stated purpose and keeps data on-premises. The external GitHub reference is for documentation/installation, not runtime data transmission. The 'UNKNOWN_DATA_SINK' signal appears to be a false positive, as the endpoint is part of the local fleet management dashboard.
Audited Apr 16, 2026 · audit v1.0
Generated May 6, 2026
Embed thousands of document chunks for a retrieval-augmented generation pipeline. The fleet distributes embedding loads across nodes, allowing large-scale indexing without blocking LLM inference.
Enable semantic search on a fleet of devices by embedding user queries and documents. Fleet routing ensures low-latency responses even with many concurrent requests.
Use embeddings to find near-duplicate documents or messages across a large corpus. Batch embedding across nodes speeds up comparison and deduplication.
Generate item embeddings for a recommendation engine, then compute similarities to find top candidates for users. Fleet parallelism accelerates embedding updates.
Index text descriptions from images or audio transcripts into embeddings for cross-modal search. The fleet handles the embedding workload while other nodes do inference.
Offer basic embedding API access for free with rate limits, then charge for higher throughput or batch processing. Revenue comes from subscription tiers or pay-per-request.
License the fleet embedding solution to companies that want to embed documents privately on their own infrastructure. Revenue from licensing and support contracts.
Provide a fully managed embedding fleet with monitoring, scaling, and maintenance. Charge monthly based on number of nodes and embedding volume.
💬 Integration Tip
Start with the curl or Python examples—just replace localhost:11435 with your fleet router address. Use batch embedding for RAG to maximize throughput.
Scored May 6, 2026
Use CodexBar CLI local cost usage to summarize per-model usage for Codex or Claude, including the current (most recent) model or a full model breakdown. Trigger when asked for model-level usage/cost data from codexbar, or when you need a scriptable per-model summary from codexbar cost JSON.
Gemini CLI for one-shot Q&A, summaries, and generation.
Manages free AI models from OpenRouter for OpenClaw. Automatically ranks models by quality, configures fallbacks for rate-limit handling, and updates openclaw.json. Use when the user mentions free AI, OpenRouter, model switching, rate limits, or wants to reduce AI costs.
Manages free AI models from OpenRouter for OpenClaw. Automatically ranks models by quality, configures fallbacks for rate-limit handling, and updates opencla...
Reduce OpenClaw AI costs by 97%. Haiku model routing, free Ollama heartbeats, prompt caching, and budget controls. Go from $1,500/month to $50/month in 5 min...
HTML-first PDF production skill for reports, papers, and structured documents. Must be applied before generating PDF deliverables from HTML.