from langchain.tools import tool from deerflow.config import get_app_config from deerflow.utils.readability import ReadabilityExtractor from .infoquest_client import InfoQuestClient readability_extractor = ReadabilityExtractor() def _get_infoquest_client() -> InfoQuestClient: search_config = get_app_config().get_tool_config("web_search") search_time_range = -1 if search_config is not None and "search_time_range" in search_config.model_extra: search_time_range = search_config.model_extra.get("search_time_range") fetch_config = get_app_config().get_tool_config("web_fetch") fetch_time = -1 if fetch_config is not None and "fetch_time" in fetch_config.model_extra: fetch_time = fetch_config.model_extra.get("fetch_time") fetch_timeout = -1 if fetch_config is not None and "timeout" in fetch_config.model_extra: fetch_timeout = fetch_config.model_extra.get("timeout") navigation_timeout = -1 if fetch_config is not None and "navigation_timeout" in fetch_config.model_extra: navigation_timeout = fetch_config.model_extra.get("navigation_timeout") return InfoQuestClient( search_time_range=search_time_range, fetch_timeout=fetch_timeout, fetch_navigation_timeout=navigation_timeout, fetch_time=fetch_time, ) @tool("web_search", parse_docstring=True) def web_search_tool(query: str) -> str: """Search the web. Args: query: The query to search for. """ client = _get_infoquest_client() return client.web_search(query) @tool("web_fetch", parse_docstring=True) def web_fetch_tool(url: str) -> str: """Fetch the contents of a web page at a given URL. Only fetch EXACT URLs that have been provided directly by the user or have been returned in results from the web_search and web_fetch tools. This tool can NOT access content that requires authentication, such as private Google Docs or pages behind login walls. Do NOT add www. to URLs that do NOT have them. URLs must include the schema: https://example.com is a valid URL while example.com is an invalid URL. Args: url: The URL to fetch the contents of. """ client = _get_infoquest_client() result = client.fetch(url) if result.startswith("Error: "): return result article = readability_extractor.extract_article(result) return article.to_markdown()[:4096]