8ba01dfd83
* refactor: thread app config through lead prompt * fix: honor explicit app config across runtime paths * style: format subagent executor tests * fix: thread resolved app config and guard subagents-only fallback Address two PR review findings: 1. _create_summarization_middleware passed the original (possibly None) app_config into create_chat_model, forcing the model factory back to ambient get_app_config() and risking config drift between the middleware's resolved view and the model's view. Pass the resolved AppConfig instance through end-to-end. 2. get_available_subagent_names accepted Any-typed config and forwarded it to is_host_bash_allowed, which reads ``.sandbox``. A SubagentsAppConfig (also accepted upstream as a sum-type input) has no ``.sandbox`` attribute and would be silently treated as "no sandbox configured", incorrectly disabling the bash subagent. Guard on hasattr and fall back to ambient lookup otherwise. Adds regression tests for both paths. * chore: simplify hasattr guard and tighten regression tests - Collapse if/else into ternary in get_available_subagent_names; hasattr(None, ...) is False so the explicit None check was redundant. - Drop comments that narrate the change rather than explain non-obvious WHY (test names already convey intent). - Replace stringly-typed sentinel "no-arg" in regression test with direct args tuple comparison. --------- Co-authored-by: greatmengqi <chenmengqi.0376@bytedance.com>
412 lines
18 KiB
Python
412 lines
18 KiB
Python
import logging
|
|
|
|
from langchain.agents import create_agent
|
|
from langchain.agents.middleware import AgentMiddleware
|
|
from langchain_core.runnables import RunnableConfig
|
|
|
|
from deerflow.agents.lead_agent.prompt import apply_prompt_template
|
|
from deerflow.agents.memory.summarization_hook import memory_flush_hook
|
|
from deerflow.agents.middlewares.clarification_middleware import ClarificationMiddleware
|
|
from deerflow.agents.middlewares.loop_detection_middleware import LoopDetectionMiddleware
|
|
from deerflow.agents.middlewares.memory_middleware import MemoryMiddleware
|
|
from deerflow.agents.middlewares.subagent_limit_middleware import SubagentLimitMiddleware
|
|
from deerflow.agents.middlewares.summarization_middleware import BeforeSummarizationHook, DeerFlowSummarizationMiddleware
|
|
from deerflow.agents.middlewares.title_middleware import TitleMiddleware
|
|
from deerflow.agents.middlewares.todo_middleware import TodoMiddleware
|
|
from deerflow.agents.middlewares.token_usage_middleware import TokenUsageMiddleware
|
|
from deerflow.agents.middlewares.tool_error_handling_middleware import build_lead_runtime_middlewares
|
|
from deerflow.agents.middlewares.view_image_middleware import ViewImageMiddleware
|
|
from deerflow.agents.thread_state import ThreadState
|
|
from deerflow.config.agents_config import load_agent_config, validate_agent_name
|
|
from deerflow.config.app_config import AppConfig, get_app_config
|
|
from deerflow.models import create_chat_model
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
def _get_runtime_config(config: RunnableConfig) -> dict:
|
|
"""Merge legacy configurable options with LangGraph runtime context."""
|
|
cfg = dict(config.get("configurable", {}) or {})
|
|
context = config.get("context", {}) or {}
|
|
if isinstance(context, dict):
|
|
cfg.update(context)
|
|
return cfg
|
|
|
|
|
|
def _resolve_model_name(requested_model_name: str | None = None, *, app_config: AppConfig | None = None) -> str:
|
|
"""Resolve a runtime model name safely, falling back to default if invalid. Returns None if no models are configured."""
|
|
app_config = app_config or get_app_config()
|
|
default_model_name = app_config.models[0].name if app_config.models else None
|
|
if default_model_name is None:
|
|
raise ValueError("No chat models are configured. Please configure at least one model in config.yaml.")
|
|
|
|
if requested_model_name and app_config.get_model_config(requested_model_name):
|
|
return requested_model_name
|
|
|
|
if requested_model_name and requested_model_name != default_model_name:
|
|
logger.warning(f"Model '{requested_model_name}' not found in config; fallback to default model '{default_model_name}'.")
|
|
return default_model_name
|
|
|
|
|
|
def _create_summarization_middleware(*, app_config: AppConfig | None = None) -> DeerFlowSummarizationMiddleware | None:
|
|
"""Create and configure the summarization middleware from config."""
|
|
resolved_app_config = app_config or get_app_config()
|
|
config = resolved_app_config.summarization
|
|
|
|
if not config.enabled:
|
|
return None
|
|
|
|
# Prepare trigger parameter
|
|
trigger = None
|
|
if config.trigger is not None:
|
|
if isinstance(config.trigger, list):
|
|
trigger = [t.to_tuple() for t in config.trigger]
|
|
else:
|
|
trigger = config.trigger.to_tuple()
|
|
|
|
# Prepare keep parameter
|
|
keep = config.keep.to_tuple()
|
|
|
|
# Prepare model parameter.
|
|
# Bind "middleware:summarize" tag so RunJournal identifies these LLM calls
|
|
# as middleware rather than lead_agent (SummarizationMiddleware is a
|
|
# LangChain built-in, so we tag the model at creation time).
|
|
if config.model_name:
|
|
model = create_chat_model(name=config.model_name, thinking_enabled=False, app_config=resolved_app_config)
|
|
else:
|
|
model = create_chat_model(thinking_enabled=False, app_config=resolved_app_config)
|
|
model = model.with_config(tags=["middleware:summarize"])
|
|
|
|
# Prepare kwargs
|
|
kwargs = {
|
|
"model": model,
|
|
"trigger": trigger,
|
|
"keep": keep,
|
|
}
|
|
|
|
if config.trim_tokens_to_summarize is not None:
|
|
kwargs["trim_tokens_to_summarize"] = config.trim_tokens_to_summarize
|
|
|
|
if config.summary_prompt is not None:
|
|
kwargs["summary_prompt"] = config.summary_prompt
|
|
|
|
hooks: list[BeforeSummarizationHook] = []
|
|
if resolved_app_config.memory.enabled:
|
|
hooks.append(memory_flush_hook)
|
|
|
|
# The logic below relies on two assumptions holding true: this factory is
|
|
# the sole entry point for DeerFlowSummarizationMiddleware, and the runtime
|
|
# config is not expected to change after startup.
|
|
skills_container_path = resolved_app_config.skills.container_path or "/mnt/skills"
|
|
|
|
return DeerFlowSummarizationMiddleware(
|
|
**kwargs,
|
|
skills_container_path=skills_container_path,
|
|
skill_file_read_tool_names=config.skill_file_read_tool_names,
|
|
before_summarization=hooks,
|
|
preserve_recent_skill_count=config.preserve_recent_skill_count,
|
|
preserve_recent_skill_tokens=config.preserve_recent_skill_tokens,
|
|
preserve_recent_skill_tokens_per_skill=config.preserve_recent_skill_tokens_per_skill,
|
|
)
|
|
|
|
|
|
def _create_todo_list_middleware(is_plan_mode: bool) -> TodoMiddleware | None:
|
|
"""Create and configure the TodoList middleware.
|
|
|
|
Args:
|
|
is_plan_mode: Whether to enable plan mode with TodoList middleware.
|
|
|
|
Returns:
|
|
TodoMiddleware instance if plan mode is enabled, None otherwise.
|
|
"""
|
|
if not is_plan_mode:
|
|
return None
|
|
|
|
# Custom prompts matching DeerFlow's style
|
|
system_prompt = """
|
|
<todo_list_system>
|
|
You have access to the `write_todos` tool to help you manage and track complex multi-step objectives.
|
|
|
|
**CRITICAL RULES:**
|
|
- Mark todos as completed IMMEDIATELY after finishing each step - do NOT batch completions
|
|
- Keep EXACTLY ONE task as `in_progress` at any time (unless tasks can run in parallel)
|
|
- Update the todo list in REAL-TIME as you work - this gives users visibility into your progress
|
|
- DO NOT use this tool for simple tasks (< 3 steps) - just complete them directly
|
|
|
|
**When to Use:**
|
|
This tool is designed for complex objectives that require systematic tracking:
|
|
- Complex multi-step tasks requiring 3+ distinct steps
|
|
- Non-trivial tasks needing careful planning and execution
|
|
- User explicitly requests a todo list
|
|
- User provides multiple tasks (numbered or comma-separated list)
|
|
- The plan may need revisions based on intermediate results
|
|
|
|
**When NOT to Use:**
|
|
- Single, straightforward tasks
|
|
- Trivial tasks (< 3 steps)
|
|
- Purely conversational or informational requests
|
|
- Simple tool calls where the approach is obvious
|
|
|
|
**Best Practices:**
|
|
- Break down complex tasks into smaller, actionable steps
|
|
- Use clear, descriptive task names
|
|
- Remove tasks that become irrelevant
|
|
- Add new tasks discovered during implementation
|
|
- Don't be afraid to revise the todo list as you learn more
|
|
|
|
**Task Management:**
|
|
Writing todos takes time and tokens - use it when helpful for managing complex problems, not for simple requests.
|
|
</todo_list_system>
|
|
"""
|
|
|
|
tool_description = """Use this tool to create and manage a structured task list for complex work sessions.
|
|
|
|
**IMPORTANT: Only use this tool for complex tasks (3+ steps). For simple requests, just do the work directly.**
|
|
|
|
## When to Use
|
|
|
|
Use this tool in these scenarios:
|
|
1. **Complex multi-step tasks**: When a task requires 3 or more distinct steps or actions
|
|
2. **Non-trivial tasks**: Tasks requiring careful planning or multiple operations
|
|
3. **User explicitly requests todo list**: When the user directly asks you to track tasks
|
|
4. **Multiple tasks**: When users provide a list of things to be done
|
|
5. **Dynamic planning**: When the plan may need updates based on intermediate results
|
|
|
|
## When NOT to Use
|
|
|
|
Skip this tool when:
|
|
1. The task is straightforward and takes less than 3 steps
|
|
2. The task is trivial and tracking provides no benefit
|
|
3. The task is purely conversational or informational
|
|
4. It's clear what needs to be done and you can just do it
|
|
|
|
## How to Use
|
|
|
|
1. **Starting a task**: Mark it as `in_progress` BEFORE beginning work
|
|
2. **Completing a task**: Mark it as `completed` IMMEDIATELY after finishing
|
|
3. **Updating the list**: Add new tasks, remove irrelevant ones, or update descriptions as needed
|
|
4. **Multiple updates**: You can make several updates at once (e.g., complete one task and start the next)
|
|
|
|
## Task States
|
|
|
|
- `pending`: Task not yet started
|
|
- `in_progress`: Currently working on (can have multiple if tasks run in parallel)
|
|
- `completed`: Task finished successfully
|
|
|
|
## Task Completion Requirements
|
|
|
|
**CRITICAL: Only mark a task as completed when you have FULLY accomplished it.**
|
|
|
|
Never mark a task as completed if:
|
|
- There are unresolved issues or errors
|
|
- Work is partial or incomplete
|
|
- You encountered blockers preventing completion
|
|
- You couldn't find necessary resources or dependencies
|
|
- Quality standards haven't been met
|
|
|
|
If blocked, keep the task as `in_progress` and create a new task describing what needs to be resolved.
|
|
|
|
## Best Practices
|
|
|
|
- Create specific, actionable items
|
|
- Break complex tasks into smaller, manageable steps
|
|
- Use clear, descriptive task names
|
|
- Update task status in real-time as you work
|
|
- Mark tasks complete IMMEDIATELY after finishing (don't batch completions)
|
|
- Remove tasks that are no longer relevant
|
|
- **IMPORTANT**: When you write the todo list, mark your first task(s) as `in_progress` immediately
|
|
- **IMPORTANT**: Unless all tasks are completed, always have at least one task `in_progress` to show progress
|
|
|
|
Being proactive with task management demonstrates thoroughness and ensures all requirements are completed successfully.
|
|
|
|
**Remember**: If you only need a few tool calls to complete a task and it's clear what to do, it's better to just do the task directly and NOT use this tool at all.
|
|
"""
|
|
|
|
return TodoMiddleware(system_prompt=system_prompt, tool_description=tool_description)
|
|
|
|
|
|
# ThreadDataMiddleware must be before SandboxMiddleware to ensure thread_id is available
|
|
# UploadsMiddleware should be after ThreadDataMiddleware to access thread_id
|
|
# DanglingToolCallMiddleware patches missing ToolMessages before model sees the history
|
|
# SummarizationMiddleware should be early to reduce context before other processing
|
|
# TodoListMiddleware should be before ClarificationMiddleware to allow todo management
|
|
# TitleMiddleware generates title after first exchange
|
|
# MemoryMiddleware queues conversation for memory update (after TitleMiddleware)
|
|
# ViewImageMiddleware should be before ClarificationMiddleware to inject image details before LLM
|
|
# ToolErrorHandlingMiddleware should be before ClarificationMiddleware to convert tool exceptions to ToolMessages
|
|
# ClarificationMiddleware should be last to intercept clarification requests after model calls
|
|
def _build_middlewares(
|
|
config: RunnableConfig,
|
|
model_name: str | None,
|
|
agent_name: str | None = None,
|
|
custom_middlewares: list[AgentMiddleware] | None = None,
|
|
*,
|
|
app_config: AppConfig | None = None,
|
|
):
|
|
"""Build middleware chain based on runtime configuration.
|
|
|
|
Args:
|
|
config: Runtime configuration containing configurable options like is_plan_mode.
|
|
agent_name: If provided, MemoryMiddleware will use per-agent memory storage.
|
|
custom_middlewares: Optional list of custom middlewares to inject into the chain.
|
|
|
|
Returns:
|
|
List of middleware instances.
|
|
"""
|
|
resolved_app_config = app_config or get_app_config()
|
|
middlewares = build_lead_runtime_middlewares(app_config=resolved_app_config, lazy_init=True)
|
|
|
|
# Add summarization middleware if enabled
|
|
summarization_middleware = _create_summarization_middleware(app_config=resolved_app_config)
|
|
if summarization_middleware is not None:
|
|
middlewares.append(summarization_middleware)
|
|
|
|
# Add TodoList middleware if plan mode is enabled
|
|
cfg = _get_runtime_config(config)
|
|
is_plan_mode = cfg.get("is_plan_mode", False)
|
|
todo_list_middleware = _create_todo_list_middleware(is_plan_mode)
|
|
if todo_list_middleware is not None:
|
|
middlewares.append(todo_list_middleware)
|
|
|
|
# Add TokenUsageMiddleware when token_usage tracking is enabled
|
|
if resolved_app_config.token_usage.enabled:
|
|
middlewares.append(TokenUsageMiddleware())
|
|
|
|
# Add TitleMiddleware
|
|
middlewares.append(TitleMiddleware(app_config=resolved_app_config))
|
|
|
|
# Add MemoryMiddleware (after TitleMiddleware)
|
|
middlewares.append(MemoryMiddleware(agent_name=agent_name, memory_config=resolved_app_config.memory))
|
|
|
|
# Add ViewImageMiddleware only if the current model supports vision.
|
|
# Use the resolved runtime model_name from make_lead_agent to avoid stale config values.
|
|
model_config = resolved_app_config.get_model_config(model_name) if model_name else None
|
|
if model_config is not None and model_config.supports_vision:
|
|
middlewares.append(ViewImageMiddleware())
|
|
|
|
# Add DeferredToolFilterMiddleware to hide deferred tool schemas from model binding
|
|
if resolved_app_config.tool_search.enabled:
|
|
from deerflow.agents.middlewares.deferred_tool_filter_middleware import DeferredToolFilterMiddleware
|
|
|
|
middlewares.append(DeferredToolFilterMiddleware())
|
|
|
|
# Add SubagentLimitMiddleware to truncate excess parallel task calls
|
|
subagent_enabled = cfg.get("subagent_enabled", False)
|
|
if subagent_enabled:
|
|
max_concurrent_subagents = cfg.get("max_concurrent_subagents", 3)
|
|
middlewares.append(SubagentLimitMiddleware(max_concurrent=max_concurrent_subagents))
|
|
|
|
# LoopDetectionMiddleware — detect and break repetitive tool call loops
|
|
middlewares.append(LoopDetectionMiddleware())
|
|
|
|
# Inject custom middlewares before ClarificationMiddleware
|
|
if custom_middlewares:
|
|
middlewares.extend(custom_middlewares)
|
|
|
|
# ClarificationMiddleware should always be last
|
|
middlewares.append(ClarificationMiddleware())
|
|
return middlewares
|
|
|
|
|
|
def make_lead_agent(config: RunnableConfig):
|
|
"""LangGraph graph factory; keep the signature compatible with LangGraph Server."""
|
|
runtime_config = _get_runtime_config(config)
|
|
runtime_app_config = runtime_config.get("app_config")
|
|
return _make_lead_agent(config, app_config=runtime_app_config or get_app_config())
|
|
|
|
|
|
def _make_lead_agent(config: RunnableConfig, *, app_config: AppConfig):
|
|
# Lazy import to avoid circular dependency
|
|
from deerflow.tools import get_available_tools
|
|
from deerflow.tools.builtins import setup_agent
|
|
|
|
cfg = _get_runtime_config(config)
|
|
resolved_app_config = app_config
|
|
|
|
thinking_enabled = cfg.get("thinking_enabled", True)
|
|
reasoning_effort = cfg.get("reasoning_effort", None)
|
|
requested_model_name: str | None = cfg.get("model_name") or cfg.get("model")
|
|
is_plan_mode = cfg.get("is_plan_mode", False)
|
|
subagent_enabled = cfg.get("subagent_enabled", False)
|
|
max_concurrent_subagents = cfg.get("max_concurrent_subagents", 3)
|
|
is_bootstrap = cfg.get("is_bootstrap", False)
|
|
agent_name = validate_agent_name(cfg.get("agent_name"))
|
|
|
|
agent_config = load_agent_config(agent_name) if not is_bootstrap else None
|
|
# Custom agent model from agent config (if any), or None to let _resolve_model_name pick the default
|
|
agent_model_name = agent_config.model if agent_config and agent_config.model else None
|
|
|
|
# Final model name resolution: request → agent config → global default, with fallback for unknown names
|
|
model_name = _resolve_model_name(requested_model_name or agent_model_name, app_config=resolved_app_config)
|
|
|
|
model_config = resolved_app_config.get_model_config(model_name)
|
|
|
|
if model_config is None:
|
|
raise ValueError("No chat model could be resolved. Please configure at least one model in config.yaml or provide a valid 'model_name'/'model' in the request.")
|
|
if thinking_enabled and not model_config.supports_thinking:
|
|
logger.warning(f"Thinking mode is enabled but model '{model_name}' does not support it; fallback to non-thinking mode.")
|
|
thinking_enabled = False
|
|
|
|
logger.info(
|
|
"Create Agent(%s) -> thinking_enabled: %s, reasoning_effort: %s, model_name: %s, is_plan_mode: %s, subagent_enabled: %s, max_concurrent_subagents: %s",
|
|
agent_name or "default",
|
|
thinking_enabled,
|
|
reasoning_effort,
|
|
model_name,
|
|
is_plan_mode,
|
|
subagent_enabled,
|
|
max_concurrent_subagents,
|
|
)
|
|
|
|
# Inject run metadata for LangSmith trace tagging
|
|
if "metadata" not in config:
|
|
config["metadata"] = {}
|
|
|
|
config["metadata"].update(
|
|
{
|
|
"agent_name": agent_name or "default",
|
|
"model_name": model_name or "default",
|
|
"thinking_enabled": thinking_enabled,
|
|
"reasoning_effort": reasoning_effort,
|
|
"is_plan_mode": is_plan_mode,
|
|
"subagent_enabled": subagent_enabled,
|
|
"tool_groups": agent_config.tool_groups if agent_config else None,
|
|
"available_skills": ["bootstrap"] if is_bootstrap else (agent_config.skills if agent_config and agent_config.skills is not None else None),
|
|
}
|
|
)
|
|
|
|
if is_bootstrap:
|
|
# Special bootstrap agent with minimal prompt for initial custom agent creation flow
|
|
return create_agent(
|
|
model=create_chat_model(name=model_name, thinking_enabled=thinking_enabled, app_config=resolved_app_config),
|
|
tools=get_available_tools(model_name=model_name, subagent_enabled=subagent_enabled, app_config=resolved_app_config) + [setup_agent],
|
|
middleware=_build_middlewares(config, model_name=model_name, app_config=resolved_app_config),
|
|
system_prompt=apply_prompt_template(
|
|
subagent_enabled=subagent_enabled,
|
|
max_concurrent_subagents=max_concurrent_subagents,
|
|
available_skills=set(["bootstrap"]),
|
|
app_config=resolved_app_config,
|
|
),
|
|
state_schema=ThreadState,
|
|
)
|
|
|
|
# Default lead agent (unchanged behavior)
|
|
return create_agent(
|
|
model=create_chat_model(name=model_name, thinking_enabled=thinking_enabled, reasoning_effort=reasoning_effort, app_config=resolved_app_config),
|
|
tools=get_available_tools(
|
|
model_name=model_name,
|
|
groups=agent_config.tool_groups if agent_config else None,
|
|
subagent_enabled=subagent_enabled,
|
|
app_config=resolved_app_config,
|
|
),
|
|
middleware=_build_middlewares(config, model_name=model_name, agent_name=agent_name, app_config=resolved_app_config),
|
|
system_prompt=apply_prompt_template(
|
|
subagent_enabled=subagent_enabled,
|
|
max_concurrent_subagents=max_concurrent_subagents,
|
|
agent_name=agent_name,
|
|
available_skills=set(agent_config.skills) if agent_config and agent_config.skills is not None else None,
|
|
app_config=resolved_app_config,
|
|
),
|
|
state_schema=ThreadState,
|
|
)
|