Files
deepagent/deepagents_sourcecode/libs/deepagents-cli/deepagents_cli/tools.py
HyunjunJeon af5fbfabec 문서 추가: Context Engineering 문서 추가 및 deepagents_sourcecode 한국어 번역
- Context_Engineering.md: 에이전트 컨텍스트 엔지니어링 개념 정리 문서 추가
- Context_Engineering_Research.ipynb: 연구 노트북 업데이트
- deepagents_sourcecode/: docstring과 주석을 한국어로 번역
2026-01-11 17:55:52 +09:00

208 lines
6.8 KiB
Python

"""CLI 에이전트용 커스텀 도구 모음입니다.
Custom tools for the CLI agent.
"""
# NOTE(KR): 이 파일의 `http_request` / `web_search` / `fetch_url` 함수 docstring은
# LangChain tool description으로 사용될 수 있으므로 번역/수정하지 마세요(영어 유지).
from typing import Any, Literal
import requests
from markdownify import markdownify
from tavily import TavilyClient
from deepagents_cli.config import settings
_HTTP_ERROR_STATUS_CODE_MIN = 400
# Initialize Tavily client if API key is available
tavily_client = TavilyClient(api_key=settings.tavily_api_key) if settings.has_tavily else None
def http_request(
url: str,
method: str = "GET",
headers: dict[str, str] | None = None,
data: str | dict | None = None,
params: dict[str, str] | None = None,
timeout: int = 30,
) -> dict[str, Any]:
"""Make HTTP requests to APIs and web services.
Args:
url: Target URL
method: HTTP method (GET, POST, PUT, DELETE, etc.)
headers: HTTP headers to include
data: Request body data (string or dict)
params: URL query parameters
timeout: Request timeout in seconds
Returns:
Dictionary with response data including status, headers, and content
"""
try:
json_data: dict[str, Any] | None = None
body_data: str | None = None
if data is not None:
if isinstance(data, dict):
json_data = data
else:
body_data = data
response = requests.request(
method=method.upper(),
url=url,
headers=headers,
params=params,
data=body_data,
json=json_data,
timeout=timeout,
)
try:
content = response.json()
except ValueError:
content = response.text
return {
"success": response.status_code < _HTTP_ERROR_STATUS_CODE_MIN,
"status_code": response.status_code,
"headers": dict(response.headers),
"content": content,
"url": response.url,
}
except requests.exceptions.Timeout:
return {
"success": False,
"status_code": 0,
"headers": {},
"content": f"Request timed out after {timeout} seconds",
"url": url,
}
except requests.exceptions.RequestException as e:
return {
"success": False,
"status_code": 0,
"headers": {},
"content": f"Request error: {e!s}",
"url": url,
}
except Exception as e: # noqa: BLE001
return {
"success": False,
"status_code": 0,
"headers": {},
"content": f"Error making request: {e!s}",
"url": url,
}
def web_search(
query: str,
*,
max_results: int = 5,
topic: Literal["general", "news", "finance"] = "general",
include_raw_content: bool = False,
) -> dict[str, Any]:
"""Search the web using Tavily for current information and documentation.
This tool searches the web and returns relevant results. After receiving results,
you MUST synthesize the information into a natural, helpful response for the user.
Args:
query: The search query (be specific and detailed)
max_results: Number of results to return (default: 5)
topic: Search topic type - "general" for most queries, "news" for current events
include_raw_content: Include full page content (warning: uses more tokens)
Returns:
Dictionary containing:
- results: List of search results, each with:
- title: Page title
- url: Page URL
- content: Relevant excerpt from the page
- score: Relevance score (0-1)
- query: The original search query
IMPORTANT: After using this tool:
1. Read through the 'content' field of each result
2. Extract relevant information that answers the user's question
3. Synthesize this into a clear, natural language response
4. Cite sources by mentioning the page titles or URLs
5. NEVER show the raw JSON to the user - always provide a formatted response
"""
if tavily_client is None:
return {
"error": (
"Tavily API key not configured. Please set TAVILY_API_KEY environment variable."
),
"query": query,
}
try:
return tavily_client.search(
query,
max_results=max_results,
include_raw_content=include_raw_content,
topic=topic,
)
except Exception as e: # noqa: BLE001
return {"error": f"Web search error: {e!s}", "query": query}
def fetch_url(url: str, timeout: int = 30) -> dict[str, Any]:
"""Fetch content from a URL and convert HTML to markdown format.
This tool fetches web page content and converts it to clean markdown text,
making it easy to read and process HTML content. After receiving the markdown,
you MUST synthesize the information into a natural, helpful response for the user.
Args:
url: The URL to fetch (must be a valid HTTP/HTTPS URL)
timeout: Request timeout in seconds (default: 30)
Returns:
Dictionary containing:
- success: Whether the request succeeded
- url: The final URL after redirects
- markdown_content: The page content converted to markdown
- status_code: HTTP status code
- content_length: Length of the markdown content in characters
IMPORTANT: After using this tool:
1. Read through the markdown content
2. Extract relevant information that answers the user's question
3. Synthesize this into a clear, natural language response
4. NEVER show the raw markdown to the user unless specifically requested
"""
try:
response = requests.get(
url,
timeout=timeout,
headers={"User-Agent": "Mozilla/5.0 (compatible; DeepAgents/1.0)"},
)
response.raise_for_status()
# Convert HTML content to markdown
markdown_content = markdownify(response.text)
return {
"success": True,
"url": str(response.url),
"markdown_content": markdown_content,
"status_code": response.status_code,
"content_length": len(markdown_content),
}
except requests.exceptions.Timeout:
return {
"success": False,
"error": f"Fetch URL timed out after {timeout} seconds",
"url": url,
}
except requests.exceptions.RequestException as e:
return {"success": False, "error": f"Fetch URL request error: {e!s}", "url": url}
except Exception as e: # noqa: BLE001
return {"success": False, "error": f"Fetch URL error: {e!s}", "url": url}