Discover Agent Skills for web scraping & data collection. Browse 17skills for Claude, ChatGPT & Codex.
Automates the downloading and text extraction of academic PDFs to provide high-fidelity evidence for research pipelines.
Enables efficient searching and retrieval of life sciences preprints from the bioRxiv server for research and analysis.
Conducts structured, multi-threaded web research by coordinating subagents to gather and synthesize complex information into comprehensive reports.
Automates web content retrieval using a progressive four-tier fallback strategy to bypass bot detection and access restrictions.
Implements a four-tier progressive escalation strategy to reliably scrape web content and bypass advanced bot detection.
Implements a four-tier progressive scraping strategy to bypass bot detection and reliably extract web content.
Searches and retrieves life sciences preprints from the bioRxiv database with advanced filtering and PDF download capabilities.
Downloads high-quality videos and audio from YouTube and other platforms for offline access and archival.
Downloads high-quality video and audio content from YouTube and other platforms directly through your terminal workspace.
Automates multi-step information gathering and synthesis using structured planning and parallel subagents.
Conducts deep web investigations by delegating tasks to specialized subagents and synthesizing findings into organized reports.
Packs external GitHub or local repositories into a token-efficient format for deep context analysis within Claude Code.
Conducts systematic web research through autonomous subagent delegation and multi-source synthesis.
Retrieves web content through a four-tier progressive fallback strategy to bypass bot detection and access restrictions.
Automates web content extraction using a four-tier fallback strategy to bypass bot detection and CAPTCHAs.
Conducts deep, multi-faceted web research by orchestrating parallel subagents to plan, gather, and synthesize complex information.
Converts complex file formats including PDF, Office documents, and media into clean Markdown optimized for LLM processing.
Analyzes Reddit community discussions to identify feature requests, user pain points, and emerging use cases for Claude AI and Claude Code.
Extracts subtitles and transcripts from YouTube videos directly into local text files using command-line tools or browser automation.
Detects and ingests LLM-optimized documentation via the llms.txt standard to accelerate context gathering for autonomous agents.
Scrapes documentation websites and transforms them into organized, categorized reference files for AI context and offline archives.
Transforms documentation websites into structured, categorized reference files optimized for AI context and offline archives.
Enhances Claude with advanced vision analysis, real-time web searching, and deep GitHub repository exploration capabilities.
Refreshes specific event pages to synchronize and update event details within a local media database.
Downloads and formats YouTube video transcripts with precise timestamps for streamlined content analysis and text extraction.
Automates the gathering of AI industry trends, product launches, and developer insights from multiple high-signal web sources.
Translates web articles and blog posts into high-quality Chinese Markdown files while preserving original imagery and formatting.
Extracts web page content and converts it into clean, readable Markdown for seamless AI analysis and data collection.
Extracts YouTube subtitles and generates formatted Chinese transcripts with optional translation and timestamp support.
Manages YouTube channel tracking by automating video content collection, transcript retrieval, and structured summary generation.
Scroll for more results...