clawpodRead any website or search Google, even when sites block bots or are geo-restricted. Handles CAPTCHAs, JavaScript rendering, and anti-bot protection server-s...
Install via ClawdBot CLI:
clawdbot install codeninja23/clawpodFetch and extract content from URLs using Massive's Unblocker Browser API. Handles JS rendering, CAPTCHAs, and retries automatically. Returns clean markdown with headings, links, lists, tables, and code blocks.
Check for the API token and try to set up node-html-markdown:
echo "TOKEN=${MASSIVE_UNBLOCKER_TOKEN:-MISSING}"
node -e "require('node-html-markdown')" 2>/dev/null || npm install -g node-html-markdown 2>/dev/null || echo "NHM=UNAVAILABLE (will fall back to raw HTML)"
If token is MISSING, stop and tell the user:
If node-html-markdown is unavailable, proceed anyway — raw HTML will be returned and the LLM can parse it directly.
Single endpoint. GET request. Returns rendered HTML. Pipe through node-html-markdown for clean markdown (falls back to raw HTML if unavailable).
https://unblocker.joinmassive.com/browser?url=<encoded-url>
Auth header: Authorization: Bearer $MASSIVE_UNBLOCKER_TOKEN
curl -s -G --data-urlencode "url=THE_URL" \
-H "Authorization: Bearer $MASSIVE_UNBLOCKER_TOKEN" \
"https://unblocker.joinmassive.com/browser" -o /tmp/_page.html && \
(node -e "const{NodeHtmlMarkdown}=require('node-html-markdown');console.log(NodeHtmlMarkdown.translate(require('fs').readFileSync('/tmp/_page.html','utf8')))" 2>/dev/null || cat /tmp/_page.html)
Replace THE_URL with the actual URL. curl --data-urlencode handles URL-encoding automatically.
Loop through them sequentially. Each call can take up to 2 minutes (CAPTCHA solving, retries).
URLS=(
"https://example.com/page1"
"https://example.com/page2"
)
for url in "${URLS[@]}"; do
echo "=== $url ==="
curl -s -G --data-urlencode "url=$url" \
-H "Authorization: Bearer $MASSIVE_UNBLOCKER_TOKEN" \
"https://unblocker.joinmassive.com/browser" -o /tmp/_page.html && \
(node -e "const{NodeHtmlMarkdown}=require('node-html-markdown');console.log(NodeHtmlMarkdown.translate(require('fs').readFileSync('/tmp/_page.html','utf8')))" 2>/dev/null || cat /tmp/_page.html)
done
Append to the query string as needed:
| Parameter | Values | Default | Use when |
|-----------|--------|---------|----------|
| format | rendered, raw | rendered | Use raw to skip JS rendering (faster) |
| expiration | 0 to N (days) | 1 | Set 0 to bypass cache |
| delay | 0.1 to 10 (seconds) | none | Page needs extra time to load dynamic content |
| device | device name string | desktop | Need mobile-specific content |
| ip | residential, isp | residential | ISP IPs for less detection |
Example with options:
curl -s -G --data-urlencode "url=THE_URL" \
-H "Authorization: Bearer $MASSIVE_UNBLOCKER_TOKEN" \
"https://unblocker.joinmassive.com/browser?expiration=0&delay=2" -o /tmp/_page.html && \
(node -e "const{NodeHtmlMarkdown}=require('node-html-markdown');console.log(NodeHtmlMarkdown.translate(require('fs').readFileSync('/tmp/_page.html','utf8')))" 2>/dev/null || cat /tmp/_page.html)
Generated Mar 1, 2026
Marketing teams can extract pricing, product features, and promotional content from competitor websites that use anti-scraping measures. This enables real-time market analysis without being blocked by CAPTCHAs or geo-restrictions.
Researchers can gather content from paywalled academic journals, news sites with article limits, and dynamically loaded research databases. The tool handles JavaScript rendering to access full articles behind subscription barriers.
Financial analysts can extract real-time stock data, earnings reports, and market news from financial portals that implement bot protection. The residential IP option helps avoid detection while gathering time-sensitive information.
Travel agencies can scrape flight and hotel prices from booking websites that use sophisticated bot detection. The delay parameter ensures dynamic pricing content fully loads before extraction, providing accurate rate comparisons.
Property platforms can collect listings from multiple real estate websites that implement anti-scraping technologies. The tool bypasses geo-restrictions to access location-specific property data across different markets.
Offer tiered monthly subscriptions based on request volume, with premium tiers providing residential IPs and faster response times. Enterprise clients pay for dedicated infrastructure and custom device profiles.
Aggregate and clean extracted content from high-value sources, then package and sell structured datasets to businesses. Focus on industries like finance, real estate, and e-commerce where fresh data has premium value.
License the technology to other SaaS platforms that need web data extraction capabilities. Provide custom integrations with their existing workflows and handle all the infrastructure maintenance and CAPTCHA solving.
💬 Integration Tip
Always check for the MASSIVE_UNBLOCKER_TOKEN environment variable first and guide users to the waitlist if missing. Use the raw format parameter for faster responses when JavaScript rendering isn't needed.
Set up and use 1Password CLI (op). Use when installing the CLI, enabling desktop app integration, signing in (single or multi-account), or reading/injecting/running secrets via op.
Security-first skill vetting for AI agents. Use before installing any skill from ClawdHub, GitHub, or other sources. Checks for red flags, permission scope, and suspicious patterns.
Perform a comprehensive read-only security audit of Clawdbot's own configuration. This is a knowledge-based skill that teaches Clawdbot to identify hardening opportunities across the system. Use when user asks to "run security check", "audit clawdbot", "check security hardening", or "what vulnerabilities does my Clawdbot have". This skill uses Clawdbot's internal capabilities and file system access to inspect configuration, detect misconfigurations, and recommend remediations. It is designed to be extensible - new checks can be added by updating this skill's knowledge.
Use when reviewing code for security vulnerabilities, implementing authentication flows, auditing OWASP Top 10, configuring CORS/CSP headers, handling secrets, input validation, SQL injection prevention, XSS protection, or any security-related code review.
Security check for ClawHub skills powered by Koi. Query the Clawdex API before installing any skill to verify it's safe.
Scan Clawdbot and MCP skills for malware, spyware, crypto-miners, and malicious code patterns before you install them. Security audit tool that detects data exfiltration, system modification attempts, backdoors, and obfuscation techniques.