| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | from abc import ABC |
| | import asyncio |
| | from crawl4ai import AsyncWebCrawler |
| | from agent.component.base import ComponentBase, ComponentParamBase |
| | from api.utils.web_utils import is_valid_url |
| |
|
| |
|
| | class CrawlerParam(ComponentParamBase): |
| | """ |
| | Define the Crawler component parameters. |
| | """ |
| |
|
| | def __init__(self): |
| | super().__init__() |
| | self.proxy = None |
| | self.extract_type = "markdown" |
| | |
| | def check(self): |
| | self.check_valid_value(self.extract_type, "Type of content from the crawler", ['html', 'markdown', 'content']) |
| |
|
| |
|
| | class Crawler(ComponentBase, ABC): |
| | component_name = "Crawler" |
| |
|
| | def _run(self, history, **kwargs): |
| | ans = self.get_input() |
| | ans = " - ".join(ans["content"]) if "content" in ans else "" |
| | if not is_valid_url(ans): |
| | return Crawler.be_output("URL not valid") |
| | try: |
| | result = asyncio.run(self.get_web(ans)) |
| |
|
| | return Crawler.be_output(result) |
| | |
| | except Exception as e: |
| | return Crawler.be_output(f"An unexpected error occurred: {str(e)}") |
| |
|
| | async def get_web(self, url): |
| | proxy = self._param.proxy if self._param.proxy else None |
| | async with AsyncWebCrawler(verbose=True, proxy=proxy) as crawler: |
| | result = await crawler.arun( |
| | url=url, |
| | bypass_cache=True |
| | ) |
| | |
| | if self._param.extract_type == 'html': |
| | return result.cleaned_html |
| | elif self._param.extract_type == 'markdown': |
| | return result.markdown |
| | elif self._param.extract_type == 'content': |
| | result.extracted_content |
| | return result.markdown |
| |
|