parallel-extractURL content extraction via Parallel API. Extracts clean markdown from webpages, articles, PDFs, and JS-heavy sites. Use for reading specific URLs with LLM-ready output.
Install via ClawdBot CLI:
clawdbot install NormallyGaussian/parallel-extractExtract clean, LLM-ready content from URLs. Handles webpages, articles, PDFs, and JavaScript-heavy sites that need rendering.
Trigger this skill when the user asks for:
Use Search to discover; use Extract to read.
parallel-cli extract "https://example.com/article" --json
parallel-cli extract "<url>" [options]
| Flag | Description |
|------|-------------|
| --url " | URL to extract (repeatable, max 10) |
| --objective " | Focus extraction on specific content |
| --json | Output as JSON |
| --excerpts / --no-excerpts | Include relevant excerpts (default: on) |
| --full-content / --no-full-content | Include full page content |
| --excerpts-max-chars N | Max chars per excerpt |
| --excerpts-max-total-chars N | Max total excerpt chars |
| --full-max-chars N | Max full content chars |
| -o | Save output to file |
Basic extraction:
parallel-cli extract "https://example.com/article" --json
Focused extraction:
parallel-cli extract "https://example.com/pricing" \
--objective "pricing tiers and features" \
--json
Full content for PDFs:
parallel-cli extract "https://example.com/whitepaper.pdf" \
--full-content \
--json
Multiple URLs:
parallel-cli extract \
--url "https://example.com/page1" \
--url "https://example.com/page2" \
--json
When extracting, provide context:
Good: --objective "Find the installation steps and system requirements"
Poor: --objective "Read the page"
Returns structured JSON with:
url — source URLtitle — page titleexcerpts[] — relevant text excerpts (if enabled)full_content — complete page content (if enabled)publish_date — when availableWhen turning extracted content into a user-facing answer:
For long conversations, save results and use sessions_spawn:
parallel-cli extract "<url>" --json -o /tmp/extract-<topic>.json
Then spawn a sub-agent:
{
"tool": "sessions_spawn",
"task": "Read /tmp/extract-<topic>.json and summarize the key content.",
"label": "extract-summary"
}
| Exit Code | Meaning |
|-----------|---------|
| 0 | Success |
| 1 | Unexpected error (network, parse) |
| 2 | Invalid arguments |
| 3 | API error (non-2xx) |
curl -fsSL https://parallel.ai/install.sh | bash
export PARALLEL_API_KEY=your-key
Generated Mar 1, 2026
Business analysts use the skill to extract pricing, features, and specifications from competitor websites and PDF whitepapers, enabling rapid comparison and strategic planning. It handles JavaScript-heavy pages that might hide dynamic content, ensuring comprehensive data collection.
Researchers and legal professionals extract clean text from academic articles, legal documents, and paywalled content for analysis, summarization, or evidence gathering. The ability to process PDFs and focus on specific objectives streamlines literature reviews and case preparation.
Media companies and content creators fetch articles from multiple URLs to compile newsletters, reports, or summaries, using excerpts to highlight key points. The JSON output integrates easily with automation tools for regular updates.
Support teams extract troubleshooting steps or product documentation from internal or external webpages to answer user queries accurately. The skill helps maintain up-to-date information by reading updated guides and FAQs.
Financial analysts extract earnings reports, market analyses, and regulatory filings from websites to monitor trends and compliance. The focus on preserving numbers and dates ensures data integrity for further processing.
Parallel offers the extraction API as a paid service, charging based on usage volume or subscription tiers. Revenue comes from developers and enterprises integrating it into their applications for automated content processing.
Targeting large organizations with custom integrations, support, and enhanced features like higher rate limits or dedicated instances. Revenue is generated through annual contracts and tailored service packages.
Monetizing the CLI tool and associated developer resources, potentially through freemium models where basic features are free and advanced capabilities require payment. Revenue streams include premium support and advanced API access.
💬 Integration Tip
Use the --json flag for easy parsing in scripts, and combine with sessions_spawn for long-running tasks to manage context limits effectively.
Advanced filesystem operations - listing, searching, batch processing, and directory analysis for Clawdbot
Perform advanced filesystem tasks including listing, recursive searching by name or content, batch copying/moving/deleting files, and analyzing directory siz...
Essential SSH commands for secure remote access, key management, tunneling, and file transfers.
Extract text from PDF files for LLM processing
The directory for AI agent services. Discover tools, platforms, and infrastructure built for agents.
Advanced filesystem operations - listing, searching, batch processing, and directory analysis for Clawdbot