mirror of
https://gitee.com/wanwujie/deer-flow
synced 2026-04-26 15:24:48 +08:00
Replace all bare print() calls with proper logging using Python's standard logging module across the deerflow harness package. Changes across 8 files (16 print statements replaced): - agents/middlewares/clarification_middleware.py: use logger.info/debug - agents/middlewares/memory_middleware.py: use logger.debug - agents/middlewares/thread_data_middleware.py: use logger.debug - agents/middlewares/view_image_middleware.py: use logger.debug - agents/memory/queue.py: use logger.info/debug/warning/error - agents/lead_agent/prompt.py: use logger.error - skills/loader.py: use logger.warning - skills/parser.py: use logger.error Each file follows the established codebase convention: import logging logger = logging.getLogger(__name__) Log levels chosen based on message semantics: - debug: routine operational details (directory creation, timer resets) - info: significant state changes (memory queued, updates processed) - warning: recoverable issues (config load failures, skipped updates) - error: unexpected failures (parsing errors, memory update errors) Note: client.py is intentionally excluded as it uses print() for CLI output, which is the correct behavior for a command-line client. Co-authored-by: moose-lab <moose-lab@users.noreply.github.com>
226 lines
8.5 KiB
Python
226 lines
8.5 KiB
Python
"""Middleware for injecting image details into conversation before LLM call."""
|
|
|
|
import logging
|
|
from typing import NotRequired, override
|
|
|
|
from langchain.agents import AgentState
|
|
from langchain.agents.middleware import AgentMiddleware
|
|
from langchain_core.messages import AIMessage, HumanMessage, ToolMessage
|
|
from langgraph.runtime import Runtime
|
|
|
|
from deerflow.agents.thread_state import ViewedImageData
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
class ViewImageMiddlewareState(AgentState):
|
|
"""Compatible with the `ThreadState` schema."""
|
|
|
|
viewed_images: NotRequired[dict[str, ViewedImageData] | None]
|
|
|
|
|
|
class ViewImageMiddleware(AgentMiddleware[ViewImageMiddlewareState]):
|
|
"""Injects image details as a human message before LLM calls when view_image tools have completed.
|
|
|
|
This middleware:
|
|
1. Runs before each LLM call
|
|
2. Checks if the last assistant message contains view_image tool calls
|
|
3. Verifies all tool calls in that message have been completed (have corresponding ToolMessages)
|
|
4. If conditions are met, creates a human message with all viewed image details (including base64 data)
|
|
5. Adds the message to state so the LLM can see and analyze the images
|
|
|
|
This enables the LLM to automatically receive and analyze images that were loaded via view_image tool,
|
|
without requiring explicit user prompts to describe the images.
|
|
"""
|
|
|
|
state_schema = ViewImageMiddlewareState
|
|
|
|
def _get_last_assistant_message(self, messages: list) -> AIMessage | None:
|
|
"""Get the last assistant message from the message list.
|
|
|
|
Args:
|
|
messages: List of messages
|
|
|
|
Returns:
|
|
Last AIMessage or None if not found
|
|
"""
|
|
for msg in reversed(messages):
|
|
if isinstance(msg, AIMessage):
|
|
return msg
|
|
return None
|
|
|
|
def _has_view_image_tool(self, message: AIMessage) -> bool:
|
|
"""Check if the assistant message contains view_image tool calls.
|
|
|
|
Args:
|
|
message: Assistant message to check
|
|
|
|
Returns:
|
|
True if message contains view_image tool calls
|
|
"""
|
|
if not hasattr(message, "tool_calls") or not message.tool_calls:
|
|
return False
|
|
|
|
return any(tool_call.get("name") == "view_image" for tool_call in message.tool_calls)
|
|
|
|
def _all_tools_completed(self, messages: list, assistant_msg: AIMessage) -> bool:
|
|
"""Check if all tool calls in the assistant message have been completed.
|
|
|
|
Args:
|
|
messages: List of all messages
|
|
assistant_msg: The assistant message containing tool calls
|
|
|
|
Returns:
|
|
True if all tool calls have corresponding ToolMessages
|
|
"""
|
|
if not hasattr(assistant_msg, "tool_calls") or not assistant_msg.tool_calls:
|
|
return False
|
|
|
|
# Get all tool call IDs from the assistant message
|
|
tool_call_ids = {tool_call.get("id") for tool_call in assistant_msg.tool_calls if tool_call.get("id")}
|
|
|
|
# Find the index of the assistant message
|
|
try:
|
|
assistant_idx = messages.index(assistant_msg)
|
|
except ValueError:
|
|
return False
|
|
|
|
# Get all ToolMessages after the assistant message
|
|
completed_tool_ids = set()
|
|
for msg in messages[assistant_idx + 1 :]:
|
|
if isinstance(msg, ToolMessage) and msg.tool_call_id:
|
|
completed_tool_ids.add(msg.tool_call_id)
|
|
|
|
# Check if all tool calls have been completed
|
|
return tool_call_ids.issubset(completed_tool_ids)
|
|
|
|
def _create_image_details_message(self, state: ViewImageMiddlewareState) -> list[str | dict]:
|
|
"""Create a formatted message with all viewed image details.
|
|
|
|
Args:
|
|
state: Current state containing viewed_images
|
|
|
|
Returns:
|
|
List of content blocks (text and images) for the HumanMessage
|
|
"""
|
|
viewed_images = state.get("viewed_images", {})
|
|
if not viewed_images:
|
|
# Return a properly formatted text block, not a plain string array
|
|
return [{"type": "text", "text": "No images have been viewed."}]
|
|
|
|
# Build the message with image information
|
|
content_blocks: list[str | dict] = [{"type": "text", "text": "Here are the images you've viewed:"}]
|
|
|
|
for image_path, image_data in viewed_images.items():
|
|
mime_type = image_data.get("mime_type", "unknown")
|
|
base64_data = image_data.get("base64", "")
|
|
|
|
# Add text description
|
|
content_blocks.append({"type": "text", "text": f"\n- **{image_path}** ({mime_type})"})
|
|
|
|
# Add the actual image data so LLM can "see" it
|
|
if base64_data:
|
|
content_blocks.append(
|
|
{
|
|
"type": "image_url",
|
|
"image_url": {"url": f"data:{mime_type};base64,{base64_data}"},
|
|
}
|
|
)
|
|
|
|
return content_blocks
|
|
|
|
def _should_inject_image_message(self, state: ViewImageMiddlewareState) -> bool:
|
|
"""Determine if we should inject an image details message.
|
|
|
|
Args:
|
|
state: Current state
|
|
|
|
Returns:
|
|
True if we should inject the message
|
|
"""
|
|
messages = state.get("messages", [])
|
|
if not messages:
|
|
return False
|
|
|
|
# Get the last assistant message
|
|
last_assistant_msg = self._get_last_assistant_message(messages)
|
|
if not last_assistant_msg:
|
|
return False
|
|
|
|
# Check if it has view_image tool calls
|
|
if not self._has_view_image_tool(last_assistant_msg):
|
|
return False
|
|
|
|
# Check if all tools have been completed
|
|
if not self._all_tools_completed(messages, last_assistant_msg):
|
|
return False
|
|
|
|
# Check if we've already added an image details message
|
|
# Look for a human message after the last assistant message that contains image details
|
|
assistant_idx = messages.index(last_assistant_msg)
|
|
for msg in messages[assistant_idx + 1 :]:
|
|
if isinstance(msg, HumanMessage):
|
|
content_str = str(msg.content)
|
|
if "Here are the images you've viewed" in content_str or "Here are the details of the images you've viewed" in content_str:
|
|
# Already added, don't add again
|
|
return False
|
|
|
|
return True
|
|
|
|
def _inject_image_message(self, state: ViewImageMiddlewareState) -> dict | None:
|
|
"""Internal helper to inject image details message.
|
|
|
|
Args:
|
|
state: Current state
|
|
|
|
Returns:
|
|
State update with additional human message, or None if no update needed
|
|
"""
|
|
if not self._should_inject_image_message(state):
|
|
return None
|
|
|
|
# Create the image details message with text and image content
|
|
image_content = self._create_image_details_message(state)
|
|
|
|
# Create a new human message with mixed content (text + images)
|
|
human_msg = HumanMessage(content=image_content)
|
|
|
|
logger.debug("Injecting image details message with images before LLM call")
|
|
|
|
# Return state update with the new message
|
|
return {"messages": [human_msg]}
|
|
|
|
@override
|
|
def before_model(self, state: ViewImageMiddlewareState, runtime: Runtime) -> dict | None:
|
|
"""Inject image details message before LLM call if view_image tools have completed (sync version).
|
|
|
|
This runs before each LLM call, checking if the previous turn included view_image
|
|
tool calls that have all completed. If so, it injects a human message with the image
|
|
details so the LLM can see and analyze the images.
|
|
|
|
Args:
|
|
state: Current state
|
|
runtime: Runtime context (unused but required by interface)
|
|
|
|
Returns:
|
|
State update with additional human message, or None if no update needed
|
|
"""
|
|
return self._inject_image_message(state)
|
|
|
|
@override
|
|
async def abefore_model(self, state: ViewImageMiddlewareState, runtime: Runtime) -> dict | None:
|
|
"""Inject image details message before LLM call if view_image tools have completed (async version).
|
|
|
|
This runs before each LLM call, checking if the previous turn included view_image
|
|
tool calls that have all completed. If so, it injects a human message with the image
|
|
details so the LLM can see and analyze the images.
|
|
|
|
Args:
|
|
state: Current state
|
|
runtime: Runtime context (unused but required by interface)
|
|
|
|
Returns:
|
|
State update with additional human message, or None if no update needed
|
|
"""
|
|
return self._inject_image_message(state)
|