docs/md_v2/index.md
</a>
</a>
<a href="https://github.com/unclecode/crawl4ai/network/members">
</a>
<a href="https://badge.fury.io/py/crawl4ai">
</a>
</a>
<a href="https://pepy.tech/project/crawl4ai">
</a>
<a href="https://github.com/unclecode/crawl4ai/blob/main/LICENSE">
</a>
</a>
<a href="https://www.linkedin.com/company/crawl4ai">
</a>
<a href="https://discord.gg/jP8KfhDhyN">
</a>
Reliable, large-scale web extraction, now built to be drastically more cost-effective than any of the existing solutions.
👉 Apply here for early access
We’ll be onboarding in phases and working closely with early users.
Limited slots.
Crawl4AI is the #1 trending GitHub repository, actively maintained by a vibrant community. It delivers blazing-fast, AI-ready web crawling tailored for large language models, AI agents, and data pipelines. Fully open source, flexible, and built for real-time performance, Crawl4AI empowers developers with unmatched speed, precision, and deployment ease.
Enjoy using Crawl4AI? Consider becoming a sponsor to support ongoing development and community growth!
Crawl4AI now features intelligent adaptive crawling that knows when to stop! Using advanced information foraging algorithms, it determines when sufficient information has been gathered to answer your query.
Learn more about Adaptive Crawling →
Here's a quick example to show you how easy it is to use Crawl4AI with its asynchronous capabilities:
import asyncio
from crawl4ai import AsyncWebCrawler
async def main():
# Create an instance of AsyncWebCrawler
async with AsyncWebCrawler() as crawler:
# Run the crawler on a URL
result = await crawler.arun(url="https://crawl4ai.com")
# Print the extracted content
print(result.markdown)
# Run the async main function
asyncio.run(main())
Crawl4AI is a feature-rich crawler and scraper that aims to:
1. Generate Clean Markdown: Perfect for RAG pipelines or direct ingestion into LLMs.
2. Structured Extraction: Parse repeated patterns with CSS, XPath, or LLM-based extraction.
3. Advanced Browser Control: Hooks, proxies, stealth modes, session re-use—fine-grained control.
4. High Performance: Parallel crawling, chunk-based extraction, real-time use cases.
5. Open Source: No forced API keys, no paywalls—everyone can access their data.
Core Philosophies:
To help you get started, we’ve organized our docs into clear sections:
AsyncWebCrawler, arun(), and CrawlResult.Throughout these sections, you’ll find code samples you can copy-paste into your environment. If something is missing or unclear, raise an issue or PR.
Our mission: to empower everyone—students, researchers, entrepreneurs, data scientists—to access, parse, and shape the world’s data with speed, cost-efficiency, and creative freedom.
Thank you for joining me on this journey. Let’s keep building an open, democratic approach to data extraction and AI together.
Happy Crawling!
— Unclecode, Founder & Maintainer of Crawl4AI