decodo-scraperSearch Google, scrape web pages, Amazon product pages, YouTube subtitles, or Reddit (post/subreddit) using the Decodo Scraper OpenClaw Skill.
Install via ClawdBot CLI:
clawdbot install DonatasDecodo/decodo-scraperUse this skill to search Google, scrape any URL, or fetch YouTube subtitles via the Decodo Web Scraping API. Search outputs a JSON object of result sections; Scrape URL outputs plain markdown; Amazon and Amazon search output parsed product-page or search results (JSON). Amazon search uses --query. YouTube subtitles outputs transcript/subtitles. Reddit post and Reddit subreddit output post/listing content (JSON).
Authentication: Set DECODO_AUTH_TOKEN (Basic auth token from Decodo Dashboard β Scraping APIs) in your environment or in a .env file in the repo root.
Errors: On failure the script writes a JSON error to stderr and exits with code 1.
Use this to find URLs, answers, or structured search results. The API returns a JSON object whose results key contains several sections (not all may be present for every query):
| Section | Description |
|--------|--------------|
| organic | Main search results (titles, links, snippets). |
| ai_overviews | AI-generated overviews or summaries when Google shows them. |
| paid | Paid/sponsored results (ads). |
| related_questions | βPeople also askββstyle questions and answers. |
| related_searches | Suggested related search queries. |
| discussions_and_forums | Forum or discussion results (e.g. Reddit, Stack Exchange). |
The script outputs only the inner results object (these sections); pagination info (page, last_visible_page, parse_status_code) is not included.
Command:
python3 tools/scrape.py --target google_search --query "your search query"
Examples:
python3 tools/scrape.py --target google_search --query "best laptops 2025"
python3 tools/scrape.py --target google_search --query "python requests tutorial"
Optional: --geo us or --locale en for location/language.
Use this to get the content of a specific web page. By default the API returns content as Markdown (cleaner for LLMs and lower token usage).
Command:
python3 tools/scrape.py --target universal --url "https://example.com"
Examples:
python3 tools/scrape.py --target universal --url "https://example.com"
python3 tools/scrape.py --target universal --url "https://news.ycombinator.com/"
Use this to get parsed data from an Amazon product (or other Amazon) page. Pass the product page URL as --url. The script sends parse: true and outputs the inner results object (e.g. ads, product details, etc.).
Command:
python3 tools/scrape.py --target amazon --url "https://www.amazon.com/dp/PRODUCT_ID"
Examples:
python3 tools/scrape.py --target amazon --url "https://www.amazon.com/dp/B09H74FXNW"
Use this to search Amazon and get parsed results (search results list, delivery_postcode, etc.). Pass the search query as --query.
Command:
python3 tools/scrape.py --target amazon_search --query "your search query"
Examples:
python3 tools/scrape.py --target amazon_search --query "laptop"
Use this to get subtitles/transcript for a YouTube video. Pass the video ID (e.g. from youtube.com/watch?v=VIDEO_ID) as --query.
Command:
python3 tools/scrape.py --target youtube_subtitles --query "VIDEO_ID"
Examples:
python3 tools/scrape.py --target youtube_subtitles --query "dFu9aKJoqGg"
Use this to get the content of a Reddit post (thread). Pass the full post URL as --url.
Command:
python3 tools/scrape.py --target reddit_post --url "https://www.reddit.com/r/SUBREDDIT/comments/ID/..."
Examples:
python3 tools/scrape.py --target reddit_post --url "https://www.reddit.com/r/nba/comments/17jrqc5/serious_next_day_thread_postgame_discussion/"
Use this to get the listing (posts) of a Reddit subreddit. Pass the subreddit URL as --url.
Command:
python3 tools/scrape.py --target reddit_subreddit --url "https://www.reddit.com/r/SUBREDDIT/"
Examples:
python3 tools/scrape.py --target reddit_subreddit --url "https://www.reddit.com/r/nba/"
| Action | Target | Argument | Example command |
|--------------------|----------------------|------------|-----------------|
| Search | google_search | --query | python3 tools/scrape.py --target google_search --query "laptop" |
| Scrape page | universal | --url | python3 tools/scrape.py --target universal --url "https://example.com" |
| Amazon product | amazon | --url | python3 tools/scrape.py --target amazon --url "https://www.amazon.com/dp/B09H74FXNW" |
| Amazon search | amazon_search | --query | python3 tools/scrape.py --target amazon_search --query "laptop" |
| YouTube subtitles | youtube_subtitles | --query | python3 tools/scrape.py --target youtube_subtitles --query "dFu9aKJoqGg" |
| Reddit post | reddit_post | --url | python3 tools/scrape.py --target reddit_post --url "https://www.reddit.com/r/nba/comments/17jrqc5/..." |
| Reddit subreddit | reddit_subreddit | --url | python3 tools/scrape.py --target reddit_subreddit --url "https://www.reddit.com/r/nba/" |
Output: Search β JSON (sections). Scrape URL β markdown. Amazon / Amazon search β JSON (results e.g. ads, product info, delivery_postcode). YouTube β transcript. Reddit β JSON (content).
Generated Mar 1, 2026
An e-commerce business uses the Decodo Scraper to monitor competitor pricing and product details on Amazon by scraping product pages and search results. This helps in adjusting their own pricing strategies and identifying market trends, ensuring competitive advantage in real-time.
A news aggregator platform employs the skill to scrape various news websites and Reddit subreddits for trending topics and discussions. This enables automated content curation and summarization, providing users with up-to-date information from multiple sources efficiently.
A digital marketing agency utilizes Google search scraping to analyze organic and paid search results for clients' keywords. This data helps in optimizing SEO strategies, tracking ad performance, and generating insights for content creation and campaign adjustments.
An educational technology company uses the skill to scrape YouTube subtitles for video lectures, enabling automatic transcription and content indexing. This supports creating searchable databases and study materials for students, enhancing learning accessibility.
A brand management firm scrapes Reddit posts and subreddits to monitor public sentiment and discussions about their products. This helps in identifying customer feedback, managing reputation, and informing product development decisions based on community insights.
Offer the Decodo Scraper as a cloud-based service with tiered subscription plans based on usage limits (e.g., number of scrapes per month). This model provides recurring revenue and scalability, catering to businesses of all sizes needing automated data extraction.
Monetize the scraping capabilities by providing a paid API access to developers and enterprises. Charge based on API calls or data volume, with custom pricing for high-volume clients, enabling integration into various applications without infrastructure management.
Provide consulting services to help businesses integrate the Decodo Scraper into their workflows, offering custom solutions for specific data needs. This includes setup, training, and ongoing support, generating revenue through project-based fees and retainer contracts.
π¬ Integration Tip
Ensure the DECODO_AUTH_TOKEN is securely stored in environment variables or a .env file to avoid exposure in code, and test each target with sample queries to verify output formats before full deployment.
Automates browser interactions for web testing, form filling, screenshots, and data extraction. Use when the user needs to navigate websites, interact with w...
Playwright-based web scraping OpenClaw Skill with anti-bot protection. Successfully tested on complex sites like Discuss.com.hk.
Browser automation and web scraping with Playwright. Forms, screenshots, data extraction. Works standalone or via MCP. Testing included.
Performs deep scraping of complex sites like YouTube using containerized Crawlee, extracting validated, ad-free transcripts and content as JSON output.
Automate web tasks like form filling, data scraping, testing, monitoring, and scheduled jobs with multi-browser support and retry mechanisms.
Web scraping and content comprehension agent β multi-strategy extraction with cascade fallback, news detection, boilerplate removal, structured metadata, and...