unclecode c8589f8da3 Update:
- Fix Spacy model issue
- Update Readme and requirements.txt
2024-05-16 19:50:20 +08:00
2024-05-16 19:50:20 +08:00
2024-05-16 19:50:20 +08:00
2024-05-12 22:37:21 +08:00
2024-05-16 17:31:44 +08:00
2024-05-14 21:27:41 +08:00
2024-05-14 21:27:41 +08:00
2024-05-09 19:10:25 +08:00
2024-05-09 19:10:25 +08:00
2024-05-09 19:10:25 +08:00
2024-05-16 17:31:44 +08:00
2024-05-16 19:50:20 +08:00
2024-05-16 19:50:20 +08:00
2024-05-13 00:39:06 +08:00
2024-05-09 19:10:25 +08:00

Crawl4AI 🕷️🤖

GitHub Stars GitHub Forks GitHub Issues GitHub Pull Requests License

Crawl4AI is a powerful, free web crawling service designed to extract useful information from web pages and make it accessible for large language models (LLMs) and AI applications. 🆓🌐

Recent Changes

  • 🚀 10x faster!!
  • 📜 Execute custom JavaScript before crawling!
  • 🤝 Colab friendly!
  • 📚 Chunking strategies: topic-based, regex, sentence, and more!
  • 🧠 Extraction strategies: cosine clustering, LLM, and more!
  • 🎯 CSS selector support
  • 📝 Pass instructions/keywords to refine extraction

Power and Simplicity of Crawl4AI 🚀

Crawl4AI makes even complex web crawling tasks simple and intuitive. Below is an example of how you can execute JavaScript, filter data using keywords, and use a CSS selector to extract specific content—all in one go!

Example Task:

  1. Execute custom JavaScript to click a "Load More" button.
  2. Filter the data to include only content related to "technology".
  3. Use a CSS selector to extract only paragraphs (<p> tags).

Example Code:

# Import necessary modules
from crawl4ai import WebCrawler
from crawl4ai.chunking_strategy import *
from crawl4ai.extraction_strategy import *
from crawl4ai.crawler_strategy import *

# Define the JavaScript code to click the "Load More" button
js_code = """
const loadMoreButton = Array.from(document.querySelectorAll('button')).find(button => button.textContent.includes('Load More'));
loadMoreButton && loadMoreButton.click();
"""

# Define the crawling strategy
crawler_strategy = LocalSeleniumCrawlerStrategy(js_code=js_code)

# Create the WebCrawler instance with the defined strategy
crawler = WebCrawler(crawler_strategy=crawler_strategy)

# Run the crawler with keyword filtering and CSS selector
result = crawler.run(
    url="https://www.example.com",
    extraction_strategy=CosineStrategy(
        semantic_filter="technology",
    ),
)

# Run the crawler with LLM extraction strategy
result = crawler.run(
    url="https://www.example.com",
    extraction_strategy=LLMExtractionStrategy(
        provider="openai/gpt-4o",
        api_token=os.getenv('OPENAI_API_KEY'),
        instruction="Extract only content related to technology"
    ),
    css_selector="p"
)

# Display the extracted result
print(result)

With Crawl4AI, you can perform advanced web crawling and data extraction tasks with just a few lines of code. This example demonstrates how you can harness the power of Crawl4AI to simplify your workflow and get the data you need efficiently.


Continue reading to learn more about the features, installation process, usage, and more.

Table of Contents

  1. Features
  2. Installation
  3. REST API/Local Server
  4. Python Library Usage
  5. Parameters
  6. Chunking Strategies
  7. Extraction Strategies
  8. Contributing
  9. License
  10. Contact

Features

  • 🕷️ Efficient web crawling to extract valuable data from websites
  • 🤖 LLM-friendly output formats (JSON, cleaned HTML, markdown)
  • 🌍 Supports crawling multiple URLs simultaneously
  • 🌃 Replace media tags with ALT.
  • 🆓 Completely free to use and open-source
  • 📜 Execute custom JavaScript before crawling
  • 📚 Chunking strategies: topic-based, regex, sentence, and more
  • 🧠 Extraction strategies: cosine clustering, LLM, and more
  • 🎯 CSS selector support
  • 📝 Pass instructions/keywords to refine extraction

Installation 💻

There are three ways to use Crawl4AI:

  1. As a library (Recommended)
  2. As a local server (Docker) or using the REST API
  3. As a Google Colab notebook. Open In Colab

To install Crawl4AI as a library, follow these steps:

  1. Install the package from GitHub:
virtualenv venv
source venv/bin/activate
pip install git+https://github.com/unclecode/crawl4ai.git
python docs/examples/quickstart.py
  1. Alternatively, you can clone the repository and install the package locally:
virtualenv venv
source venv/bin/activate
git clone https://github.com/unclecode/crawl4ai.git
cd crawl4ai
pip install -e .
  1. Use docker to run the local server:
docker build -t crawl4ai . 
# For Mac users
# docker build --platform linux/amd64 -t crawl4ai .
docker run -d -p 8000:80 crawl4ai

For more information about how to run Crawl4AI as a local server, please refer to the GitHub repository.

Using the Local server ot REST API 🌐

You can also use Crawl4AI through the REST API. This method allows you to send HTTP requests to the Crawl4AI server and receive structured data in response. The base URL for the API is https://crawl4ai.com/crawl. If you run the local server, you can use http://localhost:8000/crawl. (Port is dependent on your docker configuration)

Example Usage

To use the REST API, send a POST request to https://crawl4ai.com/crawl with the following parameters in the request body.

Example Request:

{
    "urls": ["https://www.example.com"],
    "include_raw_html": false,
    "bypass_cache": true,
    "word_count_threshold": 5,
    "extraction_strategy": "CosineStrategy",
    "chunking_strategy": "RegexChunking",
    "css_selector": "p",
    "verbose": true,
    "extraction_strategy_args": {
        "semantic_filter": "finance economy and stock market",
        "word_count_threshold": 20,
        "max_dist": 0.2,
        "linkage_method": "ward",
        "top_k": 3
    },
    "chunking_strategy_args": {
        "patterns": ["\n\n"]
    }
}

Example Response:

{
    "status": "success",
    "data": [
        {
            "url": "https://www.example.com",
            "extracted_content": "...",
            "html": "...",
            "markdown": "...",
            "metadata": {...}
        }
    ]
}

For more information about the available parameters and their descriptions, refer to the Parameters section.

Python Library Usage 🚀

🔥 A great way to try out Crawl4AI is to run `quickstart.py` in the `docs/examples` directory. This script demonstrates how to use Crawl4AI to crawl a website and extract content from it.

Quickstart Guide

Create an instance of WebCrawler and call the warmup() function.

crawler = WebCrawler()
crawler.warmup()

Understanding 'bypass_cache' and 'include_raw_html' parameters

First crawl (caches the result):

result = crawler.run(url="https://www.nbcnews.com/business")

Second crawl (Force to crawl again):

result = crawler.run(url="https://www.nbcnews.com/business", bypass_cache=True)
💡 Don't forget to set `bypass_cache` to True if you want to try different strategies for the same URL. Otherwise, the cached result will be returned. You can also set `always_by_pass_cache` in constructor to True to always bypass the cache.

Crawl result without raw HTML content:

result = crawler.run(url="https://www.nbcnews.com/business", include_raw_html=False)

Adding a chunking strategy: RegexChunking

Using RegexChunking:

result = crawler.run(
    url="https://www.nbcnews.com/business",
    chunking_strategy=RegexChunking(patterns=["\n\n"])
)

Using NlpSentenceChunking:

result = crawler.run(
    url="https://www.nbcnews.com/business",
    chunking_strategy=NlpSentenceChunking()
)

Extraction strategy: CosineStrategy

Using CosineStrategy:

result = crawler.run(
    url="https://www.nbcnews.com/business",
    extraction_strategy=CosineStrategy(
        semantic_filter="",
        word_count_threshold=10, 
        max_dist=0.2, 
        linkage_method="ward", 
        top_k=3
    )
)

You can set semantic_filter to filter relevant documents before clustering. Documents are filtered based on their cosine similarity to the keyword filter embedding.

result = crawler.run(
    url="https://www.nbcnews.com/business",
    extraction_strategy=CosineStrategy(
        semantic_filter="finance economy and stock market",
        word_count_threshold=10, 
        max_dist=0.2, 
        linkage_method="ward", 
        top_k=3
    )
)

Using LLMExtractionStrategy

Without instructions:

result = crawler.run(
    url="https://www.nbcnews.com/business",
    extraction_strategy=LLMExtractionStrategy(
        provider="openai/gpt-4o", 
        api_token=os.getenv('OPENAI_API_KEY')
    )
)

With instructions:

result = crawler.run(
    url="https://www.nbcnews.com/business",
    extraction_strategy=LLMExtractionStrategy(
        provider="openai/gpt-4o",
        api_token=os.getenv('OPENAI_API_KEY'),
        instruction="I am interested in only financial news"
    )
)

Targeted extraction using CSS selector

Extract only H2 tags:

result = crawler.run(
    url="https://www.nbcnews.com/business",
    css_selector="h2"
)

Passing JavaScript code to click 'Load More' button

Using JavaScript to click 'Load More' button:

js_code = """
const loadMoreButton = Array.from(document.querySelectorAll('button')).find(button => button.textContent.includes('Load More'));
loadMoreButton && loadMoreButton.click();
"""
crawler_strategy = LocalSeleniumCrawlerStrategy(js_code=js_code)
crawler = WebCrawler(crawler_strategy=crawler_strategy, always_by_pass_cache=True)
result = crawler.run(url="https://www.nbcnews.com/business")

Parameters 📖

Parameter Description Required Default Value
urls A list of URLs to crawl and extract data from. Yes -
include_raw_html Whether to include the raw HTML content in the response. No false
bypass_cache Whether to force a fresh crawl even if the URL has been previously crawled. No false
word_count_threshold The minimum number of words a block must contain to be considered meaningful (minimum value is 5). No 5
extraction_strategy The strategy to use for extracting content from the HTML (e.g., "CosineStrategy"). No CosineStrategy
chunking_strategy The strategy to use for chunking the text before processing (e.g., "RegexChunking"). No RegexChunking
css_selector The CSS selector to target specific parts of the HTML for extraction. No None
verbose Whether to enable verbose logging. No true

Chunking Strategies 📚

RegexChunking

RegexChunking is a text chunking strategy that splits a given text into smaller parts using regular expressions. This is useful for preparing large texts for processing by language models, ensuring they are divided into manageable segments.

Constructor Parameters:

  • patterns (list, optional): A list of regular expression patterns used to split the text. Default is to split by double newlines (['\n\n']).

Example usage:

chunker = RegexChunking(patterns=[r'\n\n', r'\. '])
chunks = chunker.chunk("This is a sample text. It will be split into chunks.")

NlpSentenceChunking

NlpSentenceChunking uses a natural language processing model to chunk a given text into sentences. This approach leverages SpaCy to accurately split text based on sentence boundaries.

Constructor Parameters:

  • model (str, optional): The SpaCy model to use for sentence detection. Default is 'en_core_web_sm'.

Example usage:

chunker = NlpSentenceChunking(model='en_core_web_sm')
chunks = chunker.chunk("This is a sample text. It will be split into sentences.")

TopicSegmentationChunking

TopicSegmentationChunking uses the TextTiling algorithm to segment a given text into topic-based chunks. This method identifies thematic boundaries in the text.

Constructor Parameters:

  • num_keywords (int, optional): The number of keywords to extract for each topic segment. Default is 3.

Example usage:

chunker = TopicSegmentationChunking(num_keywords=3)
chunks = chunker.chunk("This is a sample text. It will be split into topic-based segments.")

FixedLengthWordChunking

FixedLengthWordChunking splits a given text into chunks of fixed length, based on the number of words.

Constructor Parameters:

  • chunk_size (int, optional): The number of words in each chunk. Default is 100.

Example usage:

chunker = FixedLengthWordChunking(chunk_size=100)
chunks = chunker.chunk("This is a sample text. It will be split into fixed-length word chunks.")

SlidingWindowChunking

SlidingWindowChunking uses a sliding window approach to chunk a given text. Each chunk has a fixed length, and the window slides by a specified step size.

Constructor Parameters:

  • window_size (int, optional): The number of words in each chunk. Default is 100.
  • step (int, optional): The number of words to slide the window. Default is 50.

Example usage:

chunker = SlidingWindowChunking(window_size=100, step=50)
chunks = chunker.chunk("This is a sample text. It will be split using a sliding window approach.")

Extraction Strategies 🧠

NoExtractionStrategy

NoExtractionStrategy is a basic extraction strategy that returns the entire HTML content without any modification. It is useful for cases where no specific extraction is required.

Constructor Parameters: None.

Example usage:

extractor = NoExtractionStrategy()
extracted_content = extractor.extract(url, html)

LLMExtractionStrategy

LLMExtractionStrategy uses a Language Model (LLM) to extract meaningful blocks or chunks from the given HTML content. This strategy leverages an external provider for language model completions.

Constructor Parameters:

  • provider (str, optional): The provider to use for the language model completions. Default is DEFAULT_PROVIDER (e.g., openai/gpt-4).
  • api_token (str, optional): The API token for the provider. If not provided, it will try to load from the environment variable OPENAI_API_KEY.
  • instruction (str, optional): An instruction to guide the LLM on how to perform the extraction. This allows users to specify the type of data they are interested in or set the tone of the response. Default is None.

Example usage:

extractor = LLMExtractionStrategy(provider='openai', api_token='your_api_token', instruction='Extract only news about AI.')
extracted_content = extractor.extract(url, html)

CosineStrategy

CosineStrategy uses hierarchical clustering based on cosine similarity to extract clusters of text from the given HTML content. This strategy is suitable for identifying related content sections.

Constructor Parameters:

  • semantic_filter (str, optional): A string containing keywords for filtering relevant documents before clustering. If provided, documents are filtered based on their cosine similarity to the keyword filter embedding. Default is None.
  • word_count_threshold (int, optional): Minimum number of words per cluster. Default is 20.
  • max_dist (float, optional): The maximum cophenetic distance on the dendrogram to form clusters. Default is 0.2.
  • linkage_method (str, optional): The linkage method for hierarchical clustering. Default is 'ward'.
  • top_k (int, optional): Number of top categories to extract. Default is 3.
  • model_name (str, optional): The model name for embedding generation. Default is 'BAAI/bge-small-en-v1.5'.

Example usage:

extractor = CosineStrategy(semantic_filter='artificial intelligence', word_count_threshold=10, max_dist=0.2, linkage_method='ward', top_k=3, model_name='BAAI/bge-small-en-v1.5')
extracted_content = extractor.extract(url, html)

TopicExtractionStrategy

TopicExtractionStrategy uses the TextTiling algorithm to segment the HTML content into topics and extracts keywords for each segment. This strategy is useful for identifying and summarizing thematic content.

Constructor Parameters:

  • num_keywords (int, optional): Number of keywords to represent each topic segment. Default is 3.

Example usage:

extractor = TopicExtractionStrategy(num_keywords=3)
extracted_content = extractor.extract(url, html)

Contributing 🤝

We welcome contributions from the open-source community to help improve Crawl4AI and make it even more valuable for AI enthusiasts and developers. To contribute, please follow these steps:

  1. Fork the repository.
  2. Create a new branch for your feature or bug fix.
  3. Make your changes and commit them with descriptive messages.
  4. Push your changes to your forked repository.
  5. Submit a pull request to the main repository.

For more information on contributing, please see our contribution guidelines.

License 📄

Crawl4AI is released under the Apache 2.0 License.

Contact 📧

If you have any questions, suggestions, or feedback, please feel free to reach out to us:

Let's work together to make the web more accessible and useful for AI applications! 💪🌐🤖

Description
🚀🤖 Crawl4AI: Open-source LLM Friendly Web Crawler & Scraper. Don't be shy, join here: https://discord.gg/jP8KfhDhyN
Readme Pixar 141 MiB
Languages
Python 99.2%
JavaScript 0.5%
Dockerfile 0.3%