deer-flow/backend/tests/test_persistence_scaffold.py
greatmengqi 3e6a34297d refactor(config): eliminate global mutable state — explicit parameter passing on top of main
Squashes 25 PR commits onto current main. AppConfig becomes a pure value
object with no ambient lookup. Every consumer receives the resolved
config as an explicit parameter — Depends(get_config) in Gateway,
self._app_config in DeerFlowClient, runtime.context.app_config in agent
runs, AppConfig.from_file() at the LangGraph Server registration
boundary.

Phase 1 — frozen data + typed context

- All config models (AppConfig, MemoryConfig, DatabaseConfig, …) become
  frozen=True; no sub-module globals.
- AppConfig.from_file() is pure (no side-effect singleton loaders).
- Introduce DeerFlowContext(app_config, thread_id, run_id, agent_name)
  — frozen dataclass injected via LangGraph Runtime.
- Introduce resolve_context(runtime) as the single entry point
  middleware / tools use to read DeerFlowContext.

Phase 2 — pure explicit parameter passing

- Gateway: app.state.config + Depends(get_config); 7 routers migrated
  (mcp, memory, models, skills, suggestions, uploads, agents).
- DeerFlowClient: __init__(config=...) captures config locally.
- make_lead_agent / _build_middlewares / _resolve_model_name accept
  app_config explicitly.
- RunContext.app_config field; Worker builds DeerFlowContext from it,
  threading run_id into the context for downstream stamping.
- Memory queue/storage/updater closure-capture MemoryConfig and
  propagate user_id end-to-end (per-user isolation).
- Sandbox/skills/community/factories/tools thread app_config.
- resolve_context() rejects non-typed runtime.context.
- Test suite migrated off AppConfig.current() monkey-patches.
- AppConfig.current() classmethod deleted.

Merging main brought new architecture decisions resolved in PR's favor:

- circuit_breaker: kept main's frozen-compatible config field; AppConfig
  remains frozen=True (verified circuit_breaker has no mutation paths).
- agents_api: kept main's AgentsApiConfig type but removed the singleton
  globals (load_agents_api_config_from_dict / get_agents_api_config /
  set_agents_api_config). 8 routes in agents.py now read via
  Depends(get_config).
- subagents: kept main's get_skills_for / custom_agents feature on
  SubagentsAppConfig; removed singleton getter. registry.py now reads
  app_config.subagents directly.
- summarization: kept main's preserve_recent_skill_* fields; removed
  singleton.
- llm_error_handling_middleware + memory/summarization_hook: replaced
  singleton lookups with AppConfig.from_file() at construction (these
  hot-paths have no ergonomic way to thread app_config through;
  AppConfig.from_file is a pure load).
- worker.py + thread_data_middleware.py: DeerFlowContext.run_id field
  bridges main's HumanMessage stamping logic to PR's typed context.

Trade-offs (follow-up work):

- main's #2138 (async memory updater) reverted to PR's sync
  implementation. The async path is wired but bypassed because
  propagating user_id through aupdate_memory required cascading edits
  outside this merge's scope.
- tests/test_subagent_skills_config.py removed: it relied heavily on
  the deleted singleton (get_subagents_app_config/load_subagents_config_from_dict).
  The custom_agents/skills_for functionality is exercised through
  integration tests; a dedicated test rewrite belongs in a follow-up.

Verification: backend test suite — 2560 passed, 4 skipped, 84 failures.
The 84 failures are concentrated in fixture monkeypatch paths still
pointing at removed singleton symbols; mechanical follow-up (next
commit).
2026-04-26 21:45:02 +08:00

234 lines
8.4 KiB
Python

"""Tests for the persistence layer scaffolding.
Tests:
1. DatabaseConfig property derivation (paths, URLs)
2. MemoryRunStore CRUD + user_id filtering
3. Base.to_dict() via inspect mixin
4. Engine init/close lifecycle (memory + SQLite)
5. Postgres missing-dep error message
"""
from datetime import UTC, datetime
import pytest
from deerflow.config.database_config import DatabaseConfig
from deerflow.runtime.runs.store.memory import MemoryRunStore
# -- DatabaseConfig --
class TestDatabaseConfig:
def test_defaults(self):
c = DatabaseConfig()
assert c.backend == "memory"
assert c.pool_size == 5
def test_sqlite_paths_unified(self):
c = DatabaseConfig(backend="sqlite", sqlite_dir="./mydata")
assert c.sqlite_path.endswith("deerflow.db")
assert "mydata" in c.sqlite_path
# Backward-compatible aliases point to the same file
assert c.checkpointer_sqlite_path == c.sqlite_path
assert c.app_sqlite_path == c.sqlite_path
def test_app_sqlalchemy_url_sqlite(self):
c = DatabaseConfig(backend="sqlite", sqlite_dir="./data")
url = c.app_sqlalchemy_url
assert url.startswith("sqlite+aiosqlite:///")
assert "deerflow.db" in url
def test_app_sqlalchemy_url_postgres(self):
c = DatabaseConfig(
backend="postgres",
postgres_url="postgresql://u:p@h:5432/db",
)
url = c.app_sqlalchemy_url
assert url.startswith("postgresql+asyncpg://")
assert "u:p@h:5432/db" in url
def test_app_sqlalchemy_url_postgres_already_asyncpg(self):
c = DatabaseConfig(
backend="postgres",
postgres_url="postgresql+asyncpg://u:p@h:5432/db",
)
url = c.app_sqlalchemy_url
assert url.count("asyncpg") == 1
def test_memory_has_no_url(self):
c = DatabaseConfig(backend="memory")
with pytest.raises(ValueError, match="No SQLAlchemy URL"):
_ = c.app_sqlalchemy_url
# -- MemoryRunStore --
class TestMemoryRunStore:
@pytest.fixture
def store(self):
return MemoryRunStore()
@pytest.mark.anyio
async def test_put_and_get(self, store):
await store.put("r1", thread_id="t1", status="pending")
row = await store.get("r1")
assert row is not None
assert row["run_id"] == "r1"
assert row["status"] == "pending"
@pytest.mark.anyio
async def test_get_missing_returns_none(self, store):
assert await store.get("nope") is None
@pytest.mark.anyio
async def test_update_status(self, store):
await store.put("r1", thread_id="t1")
await store.update_status("r1", "running")
assert (await store.get("r1"))["status"] == "running"
@pytest.mark.anyio
async def test_update_status_with_error(self, store):
await store.put("r1", thread_id="t1")
await store.update_status("r1", "error", error="boom")
row = await store.get("r1")
assert row["status"] == "error"
assert row["error"] == "boom"
@pytest.mark.anyio
async def test_list_by_thread(self, store):
await store.put("r1", thread_id="t1")
await store.put("r2", thread_id="t1")
await store.put("r3", thread_id="t2")
rows = await store.list_by_thread("t1")
assert len(rows) == 2
assert all(r["thread_id"] == "t1" for r in rows)
@pytest.mark.anyio
async def test_list_by_thread_owner_filter(self, store):
await store.put("r1", thread_id="t1", user_id="alice")
await store.put("r2", thread_id="t1", user_id="bob")
rows = await store.list_by_thread("t1", user_id="alice")
assert len(rows) == 1
assert rows[0]["user_id"] == "alice"
@pytest.mark.anyio
async def test_owner_none_returns_all(self, store):
await store.put("r1", thread_id="t1", user_id="alice")
await store.put("r2", thread_id="t1", user_id="bob")
rows = await store.list_by_thread("t1", user_id=None)
assert len(rows) == 2
@pytest.mark.anyio
async def test_delete(self, store):
await store.put("r1", thread_id="t1")
await store.delete("r1")
assert await store.get("r1") is None
@pytest.mark.anyio
async def test_delete_nonexistent_is_noop(self, store):
await store.delete("nope") # should not raise
@pytest.mark.anyio
async def test_list_pending(self, store):
await store.put("r1", thread_id="t1", status="pending")
await store.put("r2", thread_id="t1", status="running")
await store.put("r3", thread_id="t2", status="pending")
pending = await store.list_pending()
assert len(pending) == 2
assert all(r["status"] == "pending" for r in pending)
@pytest.mark.anyio
async def test_list_pending_respects_before(self, store):
past = "2020-01-01T00:00:00+00:00"
future = "2099-01-01T00:00:00+00:00"
await store.put("r1", thread_id="t1", status="pending", created_at=past)
await store.put("r2", thread_id="t1", status="pending", created_at=future)
pending = await store.list_pending(before=datetime.now(UTC).isoformat())
assert len(pending) == 1
assert pending[0]["run_id"] == "r1"
@pytest.mark.anyio
async def test_list_pending_fifo_order(self, store):
await store.put("r2", thread_id="t1", status="pending", created_at="2024-01-02T00:00:00+00:00")
await store.put("r1", thread_id="t1", status="pending", created_at="2024-01-01T00:00:00+00:00")
pending = await store.list_pending()
assert pending[0]["run_id"] == "r1"
# -- Base.to_dict mixin --
class TestBaseToDictMixin:
@pytest.mark.anyio
async def test_to_dict_and_exclude(self, tmp_path):
"""Create a temp SQLite DB with a minimal model, verify to_dict."""
from sqlalchemy import String
from sqlalchemy.ext.asyncio import async_sessionmaker, create_async_engine
from sqlalchemy.orm import Mapped, mapped_column
from deerflow.persistence.base import Base
class _Tmp(Base):
__tablename__ = "_tmp_test"
id: Mapped[str] = mapped_column(String(64), primary_key=True)
name: Mapped[str] = mapped_column(String(128))
engine = create_async_engine(f"sqlite+aiosqlite:///{tmp_path / 'test.db'}")
async with engine.begin() as conn:
await conn.run_sync(Base.metadata.create_all)
sf = async_sessionmaker(engine, expire_on_commit=False)
async with sf() as session:
session.add(_Tmp(id="1", name="hello"))
await session.commit()
obj = await session.get(_Tmp, "1")
assert obj.to_dict() == {"id": "1", "name": "hello"}
assert obj.to_dict(exclude={"name"}) == {"id": "1"}
assert "_Tmp" in repr(obj)
await engine.dispose()
# -- Engine lifecycle --
class TestEngineLifecycle:
@pytest.mark.anyio
async def test_memory_is_noop(self):
from deerflow.persistence.engine import close_engine, get_session_factory, init_engine
await init_engine("memory")
assert get_session_factory() is None
await close_engine()
@pytest.mark.anyio
async def test_sqlite_creates_engine(self, tmp_path):
from deerflow.persistence.engine import close_engine, get_session_factory, init_engine
url = f"sqlite+aiosqlite:///{tmp_path / 'test.db'}"
await init_engine("sqlite", url=url, sqlite_dir=str(tmp_path))
sf = get_session_factory()
assert sf is not None
async with sf() as session:
assert session is not None
await close_engine()
assert get_session_factory() is None
@pytest.mark.anyio
async def test_postgres_without_asyncpg_gives_actionable_error(self):
"""If asyncpg is not installed, error message tells user what to do."""
from deerflow.persistence.engine import init_engine
try:
import asyncpg # noqa: F401
pytest.skip("asyncpg is installed -- cannot test missing-dep path")
except ImportError:
# asyncpg is not installed — this is the expected state for this test.
# We proceed to verify that init_engine raises an actionable ImportError.
pass # noqa: S110 — intentionally ignored
with pytest.raises(ImportError, match="uv sync --extra postgres"):
await init_engine("postgres", url="postgresql+asyncpg://x:x@localhost/x")