web-pilotSearch the web and read page contents without API keys. Use when you need to search via DuckDuckGo/Brave/Google (multi-page), extract readable text from URLs...
Install via ClawdBot CLI:
clawdbot install LiranUdi/web-pilotFour scripts, zero API keys. All output is JSON by default.
Dependencies: requests, beautifulsoup4, playwright (with Chromium).
Optional: pdfplumber or PyPDF2 for PDF text extraction.
Install: pip install requests beautifulsoup4 playwright && playwright install chromium
python3 scripts/google_search.py "query" --pages N --engine ENGINE
--engine — duckduckgo (default), brave, or google[{title, url, snippet}, ...]python3 scripts/read_page.py "https://url" [--max-chars N] [--visible] [--format json|markdown|text] [--no-dismiss]
--format — json (default), markdown, or text--no-dismiss)python3 scripts/browser_session.py open "https://url" # Open + extract
python3 scripts/browser_session.py navigate "https://other" # Go to new URL
python3 scripts/browser_session.py extract [--format FMT] # Re-read page
python3 scripts/browser_session.py screenshot [path] [--full] # Save screenshot
python3 scripts/browser_session.py click "Submit" # Click by text/selector
python3 scripts/browser_session.py search "keyword" # Search text in page
python3 scripts/browser_session.py tab new "https://url" # Open new tab
python3 scripts/browser_session.py tab list # List all tabs
python3 scripts/browser_session.py tab switch 1 # Switch to tab index
python3 scripts/browser_session.py tab close [index] # Close tab
python3 scripts/browser_session.py dismiss-cookies # Manually dismiss cookies
python3 scripts/browser_session.py close # Close browser
python3 scripts/download_file.py "https://example.com/doc.pdf" [--output DIR] [--filename NAME]
{status, path, filename, size_bytes, content_type, extracted_text}Generated Feb 26, 2026
Researchers can use the search and page reading capabilities to gather data on competitors, industry trends, and customer reviews from multiple sources without API limitations. The persistent browser session allows for interactive exploration of websites to extract structured information like pricing or features.
Students and legal professionals can search for academic papers, legal documents, or news articles, download PDFs, and extract text for analysis. The ability to handle PDFs and dismiss cookie banners ensures efficient access to publicly available resources without manual intervention.
Organizations can automate the monitoring of websites for updates, such as product changes or news announcements, using the browser session to navigate and extract content periodically. Screenshot functionality provides visual records for compliance or archival purposes.
Developers and QA teams can simulate user interactions like clicking buttons and searching text to test web applications. The persistent browser with tab management allows for complex multi-step workflows without needing a full browser automation suite.
AI researchers can collect large datasets from the web by searching and extracting readable text from multiple pages, with output in JSON for easy integration into machine learning pipelines. The tool's ability to handle dynamic content and cookies ensures reliable data extraction.
Offer a free version with basic search and page reading, while charging for advanced features like high-volume scraping, priority support, or cloud-based persistent sessions. Revenue can come from subscriptions or enterprise licenses.
Provide the skill as an API or plugin for existing platforms like CRM systems or data analytics tools, enabling clients to embed web data collection directly into their workflows. Charge based on usage or through partnership agreements.
Offer tailored solutions for specific industries, such as building custom scrapers or automation scripts using the skill's capabilities. Revenue is generated from project-based fees and ongoing maintenance contracts.
💬 Integration Tip
Integrate with data pipelines by using the JSON output format for seamless parsing, and consider adding error handling for network issues to improve reliability in automated workflows.
Use CodexBar CLI local cost usage to summarize per-model usage for Codex or Claude, including the current (most recent) model or a full model breakdown. Trigger when asked for model-level usage/cost data from codexbar, or when you need a scriptable per-model summary from codexbar cost JSON.
Gemini CLI for one-shot Q&A, summaries, and generation.
Research any topic from the last 30 days on Reddit + X + Web, synthesize findings, and write copy-paste-ready prompts. Use when the user wants recent social/web research on a topic, asks "what are people saying about X", or wants to learn current best practices. Requires OPENAI_API_KEY and/or XAI_API_KEY for full Reddit+X access, falls back to web search.
Check Antigravity account quotas for Claude and Gemini models. Shows remaining quota and reset times with ban detection.
Manages free AI models from OpenRouter for OpenClaw. Automatically ranks models by quality, configures fallbacks for rate-limit handling, and updates opencla...
Manages free AI models from OpenRouter for OpenClaw. Automatically ranks models by quality, configures fallbacks for rate-limit handling, and updates openclaw.json. Use when the user mentions free AI, OpenRouter, model switching, rate limits, or wants to reduce AI costs.