DocusaurusLoader#
- class langchain_community.document_loaders.docusaurus.DocusaurusLoader(
- url: str,
- custom_html_tags: List[str] | None = None,
- **kwargs: Any,
Load from Docusaurus Documentation.
It leverages the SitemapLoader to loop through the generated pages of a Docusaurus Documentation website and extracts the content by looking for specific HTML tags. By default, the parser searches for the main content of the Docusaurus page, which is normally the <article>. You can also define your own custom HTML tags by providing them as a list, for example: [“div”, “.main”, “a”].
Initialize DocusaurusLoader
- Parameters:
url (str) – The base URL of the Docusaurus website.
custom_html_tags (List[str] | None) – Optional custom html tags to extract content from pages.
kwargs (Any) – Additional args to extend the underlying SitemapLoader, for example: filter_urls, blocksize, meta_function, is_local, continue_on_failure
Attributes
web_path
Methods
__init__
(url[, custom_html_tags])Initialize DocusaurusLoader
Async lazy load text from the url(s) in web_path.
aload
()ascrape_all
(urls[, parser])Async fetch all urls, then return soups for all results.
fetch_all
(urls)Fetch all urls concurrently with rate limiting.
Load sitemap.
load
()Load data into Document objects.
load_and_split
([text_splitter])Load Documents and split into chunks.
parse_sitemap
(soup, *[, depth])Parse sitemap xml and load into a list of dicts.
scrape
([parser])Scrape data from webpage and return it in BeautifulSoup format.
scrape_all
(urls[, parser])Fetch all urls, then return soups for all results.
- __init__(
- url: str,
- custom_html_tags: List[str] | None = None,
- **kwargs: Any,
Initialize DocusaurusLoader
- Parameters:
url (str) – The base URL of the Docusaurus website.
custom_html_tags (List[str] | None) – Optional custom html tags to extract content from pages.
kwargs (Any) – Additional args to extend the underlying SitemapLoader, for example: filter_urls, blocksize, meta_function, is_local, continue_on_failure
- async alazy_load() AsyncIterator[Document] #
Async lazy load text from the url(s) in web_path.
- Return type:
AsyncIterator[Document]
- aload() List[Document] #
Deprecated since version 0.3.14: See API reference for updated usage: http://python.langchain.com/api_reference/community/document_loaders/langchain_community.document_loaders.web_base.WebBaseLoader.html It will not be removed until langchain-community==1.0.
Load text from the urls in web_path async into Documents.
- Return type:
List[Document]
- async ascrape_all(
- urls: List[str],
- parser: str | None = None,
Async fetch all urls, then return soups for all results.
- Parameters:
urls (List[str])
parser (str | None)
- Return type:
List[Any]
- async fetch_all(
- urls: List[str],
Fetch all urls concurrently with rate limiting.
- Parameters:
urls (List[str])
- Return type:
Any
- load_and_split(
- text_splitter: TextSplitter | None = None,
Load Documents and split into chunks. Chunks are returned as Documents.
Do not override this method. It should be considered to be deprecated!
- Parameters:
text_splitter (Optional[TextSplitter]) – TextSplitter instance to use for splitting documents. Defaults to RecursiveCharacterTextSplitter.
- Returns:
List of Documents.
- Return type:
list[Document]
- parse_sitemap(
- soup: Any,
- *,
- depth: int = 0,
Parse sitemap xml and load into a list of dicts.
- Parameters:
soup (Any) – BeautifulSoup object.
depth (int) – current depth of the sitemap. Default: 0
- Returns:
List of dicts.
- Return type:
List[dict]
- scrape(
- parser: str | None = None,
Scrape data from webpage and return it in BeautifulSoup format.
- Parameters:
parser (str | None)
- Return type:
Any
- scrape_all(
- urls: List[str],
- parser: str | None = None,
Fetch all urls, then return soups for all results.
- Parameters:
urls (List[str])
parser (str | None)
- Return type:
List[Any]
Examples using DocusaurusLoader