Discover Agent Skills for web scraping & data collection. Browse 17skills for Claude, ChatGPT & Codex.
Lists and manages archived snapshots from the Wayback Machine to track website history and recover lost content.
Manages local API response caching for Wayback Machine operations to optimize performance and ensure data freshness.
Locates and retrieves the most recent archived version of any URL from the Internet Archive's Wayback Machine.
Retrieves and calculates the full historical archive span for any URL using the Wayback Machine.
Retrieves comprehensive GitHub user and organization profile data including repository counts, follower statistics, and account metadata.
Retrieves the earliest archived snapshot of any URL from the Wayback Machine to identify a website's original version.
Archives URLs to the Internet Archive's Wayback Machine for permanent digital preservation and snapshot tracking.
Retrieves and manages historical visual snapshots of websites using the Internet Archive's Wayback Machine.
Converts batches of images and scanned documents into structured markdown files using local DeepSeek-OCR models via Ollama.
Orchestrates a multi-source image pipeline to download, validate, and normalize fighter photos from Wikimedia, Sherdog, and Bing.
Extracts and analyzes large PDF documents locally with semantic chunking to minimize token usage and maximize context efficiency.
Performs neural, context-aware web searches and deep research tasks to find high-quality information that keyword matching misses.
Converts any webpage into clean, formatted Markdown using Chrome CDP for full JavaScript rendering and metadata extraction.
Enables autonomous web scraping and content extraction using shot-scraper to interact with and retrieve data from websites.
Extracts and analyzes competitor advertisements from ad libraries to uncover winning messaging, pain points, and creative strategies.
Crawls and scrapes websites to extract structured article content using the FireCrawl API.
Crawls websites using the Tavily API to convert web pages into local markdown files for offline analysis and documentation retrieval.
Extracts clean, clutter-free article and blog content from URLs by stripping away ads, navigation, and unnecessary UI elements.
Conducts comprehensive, web-grounded research with automatic citations and structured data output directly from your terminal.
Tracks recent Initial Public Offerings and generates ready-to-import TradingView watchlists with enriched market data.
Converts complex PDF documents into clean, structured Markdown while preserving tables, formatting, and images for AI context.
Enables Claude to perform AI-powered web searches with real-time information and source citations using Perplexity models.
Configures sources, relevance weights, and domain interests for the Pattern Radar discovery tool.
Accesses comprehensive USPTO APIs for patent and trademark searches, examination history, and intellectual property analysis.
Downloads and processes YouTube video transcripts, subtitles, and captions with automatic fallback to AI-powered transcription.
Extracts subtitles and transcripts from YouTube videos and saves them as local text files with timestamps.
Automates resilient web content extraction using a four-tier fallback strategy to bypass bot detection and JavaScript hurdles.
Orchestrates parallel research agents across Perplexity, Claude, and Gemini to deliver synthesized, multi-perspective reports with source attribution.
Downloads YouTube videos and audio with customizable quality and format settings using yt-dlp.
Downloads high-quality videos and audio from YouTube and other platforms for offline viewing, editing, and archival.
Scroll for more results...