diff --git a/docs/_src/api/api/crawler.md b/docs/_src/api/api/crawler.md index 2f852cdde..fcc03f79d 100644 --- a/docs/_src/api/api/crawler.md +++ b/docs/_src/api/api/crawler.md @@ -73,7 +73,7 @@ List of paths where the crawled webpages got stored #### run ```python - | run(output_dir: Union[str, Path, None] = None, urls: Optional[List[str]] = None, crawler_depth: Optional[int] = None, filter_urls: Optional[List] = None, overwrite_existing_files: Optional[bool] = None, **kwargs) -> Tuple[Dict, str] + | run(output_dir: Union[str, Path, None] = None, urls: Optional[List[str]] = None, crawler_depth: Optional[int] = None, filter_urls: Optional[List] = None, overwrite_existing_files: Optional[bool] = None, return_documents: Optional[bool] = False, **kwargs) -> Tuple[Dict, str] ``` Method to be executed when the Crawler is used as a Node within a Haystack pipeline. @@ -88,6 +88,7 @@ Method to be executed when the Crawler is used as a Node within a Haystack pipel - `filter_urls`: Optional list of regular expressions that the crawled URLs must comply with. All URLs not matching at least one of the regular expressions will be dropped. - `overwrite_existing_files`: Whether to overwrite existing files in output_dir with new content +- `return_documents`: Return json files content **Returns**: