Discover Agent Skills for web scraping & data collection. Browse 17skills for Claude, ChatGPT & Codex.
Extract data, manipulate documents, and programmatically generate PDF files using specialized libraries and tools.
Searches the Google Places API for business details, location reviews, and geographic coordinates directly from the command line.
Extracts and analyzes competitor advertising strategies across platforms to provide actionable messaging and creative insights.
Downloads high-quality videos and audio from YouTube and other platforms for offline viewing, editing, or archival.
Ensures high-fidelity web research through structured source validation, error recovery protocols, and transparent reasoning cycles.
Orchestrates a structured, multi-agent workflow to conduct deep-dive research, synthesize information from multiple sources, and generate comprehensive reports.
Scrapes web content and extracts structured data from any URL or search query using the Firecrawl MCP.
Converts any live webpage into clean, structured Markdown format using Chrome CDP for full JavaScript rendering.
Converts X (Twitter) tweets, threads, and articles into clean Markdown files with YAML front matter.
Extracts and analyzes competitor advertisements from major ad libraries to provide actionable insights for messaging and creative strategy.
Downloads YouTube videos and audio files with customizable quality settings and format options directly within Claude Code.
Automates biomedical literature searches and programmatic data extraction from the PubMed database using E-utilities and advanced MeSH queries.
Accesses USPTO APIs to perform comprehensive patent and trademark searches, retrieve examination histories, and analyze intellectual property data.
Accesses the ClinicalTrials.gov API v2 to search, filter, and export clinical study data for medical research and patient matching.
Performs AI-powered web searches with real-time information and source citations to access data beyond the model's knowledge cutoff.
Fetches Reddit content and research data using the Gemini CLI to bypass web access restrictions and 403 errors.
Implements production-ready design patterns and best practices for FireCrawl SDK integrations in TypeScript and Python.
Extracts subtitles and transcripts from YouTube videos into local text files using CLI tools or browser automation.
Orchestrates multi-agent parallel workflows to perform comprehensive web research, competitive analysis, and data synthesis into structured reports.
Provides validated architectural blueprints for scaling FireCrawl integrations from MVPs to enterprise-grade microservices.
Implements production-ready project structures and architectural patterns for robust FireCrawl-based web scraping applications.
Optimizes FireCrawl operational costs through intelligent tier selection, usage monitoring, and budget-aware implementation strategies.
Executes secondary FireCrawl workflows to complement primary data collection and automated web scraping tasks.
Implements robust reliability patterns like circuit breakers, idempotency, and graceful degradation for production-grade FireCrawl integrations.
Optimizes FireCrawl API performance using advanced caching, request batching, and connection pooling strategies.
Resolves complex FireCrawl errors using systematic evidence collection and deep-layer diagnostic techniques.
Automates the installation and configuration of FireCrawl SDKs and API authentication for web scraping projects.
Executes optimized secondary search and data retrieval tasks using the Exa API to complement primary research workflows.
Executes the primary integration workflow for the Exa search engine to implement core search and data retrieval features.
Automates the primary web crawling and data extraction process using the FireCrawl API to generate LLM-ready content.
Scroll for more results...