mirror of
https://github.com/bytedance/deer-flow.git
synced 2026-04-25 11:18:22 +00:00
* feat(uploads): guide agent to use grep/glob/read_file for uploaded documents Add workflow guidance to the <uploaded_files> context block so the agent knows to use grep and glob (added in #1784) alongside read_file when working with uploaded documents, rather than falling back to web search. This is the final piece of the three-PR PDF agentic search pipeline: - PR1 (#1727): pymupdf4llm converter produces structured Markdown with headings - PR2 (#1738): document outline injected into agent context with line numbers - PR3 (this): agent guided to use outline + grep + read_file workflow * feat(uploads): add file-first priority and fallback guidance to uploaded_files context * fix(uploads): handle split-bold headings and ** ** artefacts in extract_outline - Add _clean_bold_title() to merge adjacent bold spans (** **) produced by pymupdf4llm when bold text crosses span boundaries - Add _SPLIT_BOLD_HEADING_RE (Style 3) to recognise **<num>** **<title>** headings common in academic papers; excludes pure-number table headers and rows with more than 4 bold blocks - When outline is empty, read first 5 non-empty lines of the .md as a content preview and surface a grep hint in the agent context - Update _format_file_entry to render the preview + grep hint instead of silently omitting the outline section - Add 3 new extract_outline tests and 2 new middleware tests (65 total) * fix(uploads): address Copilot review comments on extract_outline regex - Replace ASCII [A-Za-z] guard with negative lookahead to support non-ASCII titles (e.g. **1** **概述**); pure-numeric/punctuation blocks still excluded - Replace .+ with [^*]+ and cap repetition at {0,2} (four blocks total) to keep _SPLIT_BOLD_HEADING_RE linear and avoid ReDoS on malformed input - Remove now-redundant len(blocks) <= 4 code-level check (enforced by regex) - Log debug message with exc_info when preview extraction fails
290 lines
12 KiB
Python
290 lines
12 KiB
Python
"""Middleware to inject uploaded files information into agent context."""
|
|
|
|
import logging
|
|
from pathlib import Path
|
|
from typing import NotRequired, override
|
|
|
|
from langchain.agents import AgentState
|
|
from langchain.agents.middleware import AgentMiddleware
|
|
from langchain_core.messages import HumanMessage
|
|
from langgraph.runtime import Runtime
|
|
|
|
from deerflow.config.paths import Paths, get_paths
|
|
from deerflow.utils.file_conversion import extract_outline
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
|
|
_OUTLINE_PREVIEW_LINES = 5
|
|
|
|
|
|
def _extract_outline_for_file(file_path: Path) -> tuple[list[dict], list[str]]:
|
|
"""Return the document outline and fallback preview for *file_path*.
|
|
|
|
Looks for a sibling ``<stem>.md`` file produced by the upload conversion
|
|
pipeline.
|
|
|
|
Returns:
|
|
(outline, preview) where:
|
|
- outline: list of ``{title, line}`` dicts (plus optional sentinel).
|
|
Empty when no headings are found or no .md exists.
|
|
- preview: first few non-empty lines of the .md, used as a content
|
|
anchor when outline is empty so the agent has some context.
|
|
Empty when outline is non-empty (no fallback needed).
|
|
"""
|
|
md_path = file_path.with_suffix(".md")
|
|
if not md_path.is_file():
|
|
return [], []
|
|
|
|
outline = extract_outline(md_path)
|
|
if outline:
|
|
logger.debug("Extracted %d outline entries from %s", len(outline), file_path.name)
|
|
return outline, []
|
|
|
|
# outline is empty — read the first few non-empty lines as a content preview
|
|
preview: list[str] = []
|
|
try:
|
|
with md_path.open(encoding="utf-8") as f:
|
|
for line in f:
|
|
stripped = line.strip()
|
|
if stripped:
|
|
preview.append(stripped)
|
|
if len(preview) >= _OUTLINE_PREVIEW_LINES:
|
|
break
|
|
except Exception:
|
|
logger.debug("Failed to read preview lines from %s", md_path, exc_info=True)
|
|
return [], preview
|
|
|
|
|
|
class UploadsMiddlewareState(AgentState):
|
|
"""State schema for uploads middleware."""
|
|
|
|
uploaded_files: NotRequired[list[dict] | None]
|
|
|
|
|
|
class UploadsMiddleware(AgentMiddleware[UploadsMiddlewareState]):
|
|
"""Middleware to inject uploaded files information into the agent context.
|
|
|
|
Reads file metadata from the current message's additional_kwargs.files
|
|
(set by the frontend after upload) and prepends an <uploaded_files> block
|
|
to the last human message so the model knows which files are available.
|
|
"""
|
|
|
|
state_schema = UploadsMiddlewareState
|
|
|
|
def __init__(self, base_dir: str | None = None):
|
|
"""Initialize the middleware.
|
|
|
|
Args:
|
|
base_dir: Base directory for thread data. Defaults to Paths resolution.
|
|
"""
|
|
super().__init__()
|
|
self._paths = Paths(base_dir) if base_dir else get_paths()
|
|
|
|
def _format_file_entry(self, file: dict, lines: list[str]) -> None:
|
|
"""Append a single file entry (name, size, path, optional outline) to lines."""
|
|
size_kb = file["size"] / 1024
|
|
size_str = f"{size_kb:.1f} KB" if size_kb < 1024 else f"{size_kb / 1024:.1f} MB"
|
|
lines.append(f"- {file['filename']} ({size_str})")
|
|
lines.append(f" Path: {file['path']}")
|
|
outline = file.get("outline") or []
|
|
if outline:
|
|
truncated = outline[-1].get("truncated", False)
|
|
visible = [e for e in outline if not e.get("truncated")]
|
|
lines.append(" Document outline (use `read_file` with line ranges to read sections):")
|
|
for entry in visible:
|
|
lines.append(f" L{entry['line']}: {entry['title']}")
|
|
if truncated:
|
|
lines.append(f" ... (showing first {len(visible)} headings; use `read_file` to explore further)")
|
|
else:
|
|
preview = file.get("outline_preview") or []
|
|
if preview:
|
|
lines.append(" No structural headings detected. Document begins with:")
|
|
for text in preview:
|
|
lines.append(f" > {text}")
|
|
lines.append(" Use `grep` to search for keywords (e.g. `grep(pattern='keyword', path='/mnt/user-data/uploads/')`).")
|
|
lines.append("")
|
|
|
|
def _create_files_message(self, new_files: list[dict], historical_files: list[dict]) -> str:
|
|
"""Create a formatted message listing uploaded files.
|
|
|
|
Args:
|
|
new_files: Files uploaded in the current message.
|
|
historical_files: Files uploaded in previous messages.
|
|
Each file dict may contain an optional ``outline`` key — a list of
|
|
``{title, line}`` dicts extracted from the converted Markdown file.
|
|
|
|
Returns:
|
|
Formatted string inside <uploaded_files> tags.
|
|
"""
|
|
lines = ["<uploaded_files>"]
|
|
|
|
lines.append("The following files were uploaded in this message:")
|
|
lines.append("")
|
|
if new_files:
|
|
for file in new_files:
|
|
self._format_file_entry(file, lines)
|
|
else:
|
|
lines.append("(empty)")
|
|
lines.append("")
|
|
|
|
if historical_files:
|
|
lines.append("The following files were uploaded in previous messages and are still available:")
|
|
lines.append("")
|
|
for file in historical_files:
|
|
self._format_file_entry(file, lines)
|
|
|
|
lines.append("To work with these files:")
|
|
lines.append("- Read from the file first — use the outline line numbers and `read_file` to locate relevant sections.")
|
|
lines.append("- Use `grep` to search for keywords when you are not sure which section to look at")
|
|
lines.append(" (e.g. `grep(pattern='revenue', path='/mnt/user-data/uploads/')`).")
|
|
lines.append("- Use `glob` to find files by name pattern")
|
|
lines.append(" (e.g. `glob(pattern='**/*.md', path='/mnt/user-data/uploads/')`).")
|
|
lines.append("- Only fall back to web search if the file content is clearly insufficient to answer the question.")
|
|
lines.append("</uploaded_files>")
|
|
|
|
return "\n".join(lines)
|
|
|
|
def _files_from_kwargs(self, message: HumanMessage, uploads_dir: Path | None = None) -> list[dict] | None:
|
|
"""Extract file info from message additional_kwargs.files.
|
|
|
|
The frontend sends uploaded file metadata in additional_kwargs.files
|
|
after a successful upload. Each entry has: filename, size (bytes),
|
|
path (virtual path), status.
|
|
|
|
Args:
|
|
message: The human message to inspect.
|
|
uploads_dir: Physical uploads directory used to verify file existence.
|
|
When provided, entries whose files no longer exist are skipped.
|
|
|
|
Returns:
|
|
List of file dicts with virtual paths, or None if the field is absent or empty.
|
|
"""
|
|
kwargs_files = (message.additional_kwargs or {}).get("files")
|
|
if not isinstance(kwargs_files, list) or not kwargs_files:
|
|
return None
|
|
|
|
files = []
|
|
for f in kwargs_files:
|
|
if not isinstance(f, dict):
|
|
continue
|
|
filename = f.get("filename") or ""
|
|
if not filename or Path(filename).name != filename:
|
|
continue
|
|
if uploads_dir is not None and not (uploads_dir / filename).is_file():
|
|
continue
|
|
files.append(
|
|
{
|
|
"filename": filename,
|
|
"size": int(f.get("size") or 0),
|
|
"path": f"/mnt/user-data/uploads/{filename}",
|
|
"extension": Path(filename).suffix,
|
|
}
|
|
)
|
|
return files if files else None
|
|
|
|
@override
|
|
def before_agent(self, state: UploadsMiddlewareState, runtime: Runtime) -> dict | None:
|
|
"""Inject uploaded files information before agent execution.
|
|
|
|
New files come from the current message's additional_kwargs.files.
|
|
Historical files are scanned from the thread's uploads directory,
|
|
excluding the new ones.
|
|
|
|
Prepends <uploaded_files> context to the last human message content.
|
|
The original additional_kwargs (including files metadata) is preserved
|
|
on the updated message so the frontend can read it from the stream.
|
|
|
|
Args:
|
|
state: Current agent state.
|
|
runtime: Runtime context containing thread_id.
|
|
|
|
Returns:
|
|
State updates including uploaded files list.
|
|
"""
|
|
messages = list(state.get("messages", []))
|
|
if not messages:
|
|
return None
|
|
|
|
last_message_index = len(messages) - 1
|
|
last_message = messages[last_message_index]
|
|
|
|
if not isinstance(last_message, HumanMessage):
|
|
return None
|
|
|
|
# Resolve uploads directory for existence checks
|
|
thread_id = (runtime.context or {}).get("thread_id")
|
|
if thread_id is None:
|
|
try:
|
|
from langgraph.config import get_config
|
|
|
|
thread_id = get_config().get("configurable", {}).get("thread_id")
|
|
except RuntimeError:
|
|
pass # get_config() raises outside a runnable context (e.g. unit tests)
|
|
uploads_dir = self._paths.sandbox_uploads_dir(thread_id) if thread_id else None
|
|
|
|
# Get newly uploaded files from the current message's additional_kwargs.files
|
|
new_files = self._files_from_kwargs(last_message, uploads_dir) or []
|
|
|
|
# Collect historical files from the uploads directory (all except the new ones)
|
|
new_filenames = {f["filename"] for f in new_files}
|
|
historical_files: list[dict] = []
|
|
if uploads_dir and uploads_dir.exists():
|
|
for file_path in sorted(uploads_dir.iterdir()):
|
|
if file_path.is_file() and file_path.name not in new_filenames:
|
|
stat = file_path.stat()
|
|
outline, preview = _extract_outline_for_file(file_path)
|
|
historical_files.append(
|
|
{
|
|
"filename": file_path.name,
|
|
"size": stat.st_size,
|
|
"path": f"/mnt/user-data/uploads/{file_path.name}",
|
|
"extension": file_path.suffix,
|
|
"outline": outline,
|
|
"outline_preview": preview,
|
|
}
|
|
)
|
|
|
|
# Attach outlines to new files as well
|
|
if uploads_dir:
|
|
for file in new_files:
|
|
phys_path = uploads_dir / file["filename"]
|
|
outline, preview = _extract_outline_for_file(phys_path)
|
|
file["outline"] = outline
|
|
file["outline_preview"] = preview
|
|
|
|
if not new_files and not historical_files:
|
|
return None
|
|
|
|
logger.debug(f"New files: {[f['filename'] for f in new_files]}, historical: {[f['filename'] for f in historical_files]}")
|
|
|
|
# Create files message and prepend to the last human message content
|
|
files_message = self._create_files_message(new_files, historical_files)
|
|
|
|
# Extract original content - handle both string and list formats
|
|
original_content = ""
|
|
if isinstance(last_message.content, str):
|
|
original_content = last_message.content
|
|
elif isinstance(last_message.content, list):
|
|
text_parts = []
|
|
for block in last_message.content:
|
|
if isinstance(block, dict) and block.get("type") == "text":
|
|
text_parts.append(block.get("text", ""))
|
|
original_content = "\n".join(text_parts)
|
|
|
|
# Create new message with combined content.
|
|
# Preserve additional_kwargs (including files metadata) so the frontend
|
|
# can read structured file info from the streamed message.
|
|
updated_message = HumanMessage(
|
|
content=f"{files_message}\n\n{original_content}",
|
|
id=last_message.id,
|
|
additional_kwargs=last_message.additional_kwargs,
|
|
)
|
|
|
|
messages[last_message_index] = updated_message
|
|
|
|
return {
|
|
"uploaded_files": new_files,
|
|
"messages": messages,
|
|
}
|