mirror of
https://github.com/bytedance/deer-flow.git
synced 2026-04-29 05:08:26 +00:00
- config.example.yaml: deprecate standalone checkpointer section, activate unified database:sqlite as default (drives both checkpointer + app data) - New: test_thread_meta_repo.py (14 tests) — full ThreadMetaRepository coverage including check_access owner logic, list_by_owner pagination - Extended test_run_repository.py (+4 tests) — completion preserves fields, list ordering desc, limit, owner_none returns all - Extended test_run_journal.py (+8 tests) — on_chain_error, track_tokens=false, middleware no ai_message, unknown caller tokens, convenience fields, tool_error, non-summarization custom event - Extended test_run_event_store.py (+7 tests) — DB batch seq continuity, make_run_event_store factory (memory/db/jsonl/fallback/unknown) - Extended test_phase2b_integration.py (+4 tests) — create_or_reject persists, follow-up metadata, summarization in history, full DB-backed lifecycle - Fixed DB integration test to use proper fake objects (not MagicMock) for JSON-serializable metadata - 157 total Phase 2 tests pass, zero regressions Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
197 lines
7.0 KiB
Python
197 lines
7.0 KiB
Python
"""Tests for RunRepository (SQLAlchemy-backed RunStore).
|
|
|
|
Uses a temp SQLite DB to test ORM-backed CRUD operations.
|
|
"""
|
|
|
|
import pytest
|
|
|
|
from deerflow.persistence.repositories.run_repo import RunRepository
|
|
|
|
|
|
async def _make_repo(tmp_path):
|
|
from deerflow.persistence.engine import get_session_factory, init_engine
|
|
|
|
url = f"sqlite+aiosqlite:///{tmp_path / 'test.db'}"
|
|
await init_engine("sqlite", url=url, sqlite_dir=str(tmp_path))
|
|
return RunRepository(get_session_factory())
|
|
|
|
|
|
async def _cleanup():
|
|
from deerflow.persistence.engine import close_engine
|
|
|
|
await close_engine()
|
|
|
|
|
|
class TestRunRepository:
|
|
@pytest.mark.anyio
|
|
async def test_put_and_get(self, tmp_path):
|
|
repo = await _make_repo(tmp_path)
|
|
await repo.put("r1", thread_id="t1", status="pending")
|
|
row = await repo.get("r1")
|
|
assert row is not None
|
|
assert row["run_id"] == "r1"
|
|
assert row["thread_id"] == "t1"
|
|
assert row["status"] == "pending"
|
|
await _cleanup()
|
|
|
|
@pytest.mark.anyio
|
|
async def test_get_missing_returns_none(self, tmp_path):
|
|
repo = await _make_repo(tmp_path)
|
|
assert await repo.get("nope") is None
|
|
await _cleanup()
|
|
|
|
@pytest.mark.anyio
|
|
async def test_update_status(self, tmp_path):
|
|
repo = await _make_repo(tmp_path)
|
|
await repo.put("r1", thread_id="t1")
|
|
await repo.update_status("r1", "running")
|
|
row = await repo.get("r1")
|
|
assert row["status"] == "running"
|
|
await _cleanup()
|
|
|
|
@pytest.mark.anyio
|
|
async def test_update_status_with_error(self, tmp_path):
|
|
repo = await _make_repo(tmp_path)
|
|
await repo.put("r1", thread_id="t1")
|
|
await repo.update_status("r1", "error", error="boom")
|
|
row = await repo.get("r1")
|
|
assert row["status"] == "error"
|
|
assert row["error"] == "boom"
|
|
await _cleanup()
|
|
|
|
@pytest.mark.anyio
|
|
async def test_list_by_thread(self, tmp_path):
|
|
repo = await _make_repo(tmp_path)
|
|
await repo.put("r1", thread_id="t1")
|
|
await repo.put("r2", thread_id="t1")
|
|
await repo.put("r3", thread_id="t2")
|
|
rows = await repo.list_by_thread("t1")
|
|
assert len(rows) == 2
|
|
assert all(r["thread_id"] == "t1" for r in rows)
|
|
await _cleanup()
|
|
|
|
@pytest.mark.anyio
|
|
async def test_list_by_thread_owner_filter(self, tmp_path):
|
|
repo = await _make_repo(tmp_path)
|
|
await repo.put("r1", thread_id="t1", owner_id="alice")
|
|
await repo.put("r2", thread_id="t1", owner_id="bob")
|
|
rows = await repo.list_by_thread("t1", owner_id="alice")
|
|
assert len(rows) == 1
|
|
assert rows[0]["owner_id"] == "alice"
|
|
await _cleanup()
|
|
|
|
@pytest.mark.anyio
|
|
async def test_delete(self, tmp_path):
|
|
repo = await _make_repo(tmp_path)
|
|
await repo.put("r1", thread_id="t1")
|
|
await repo.delete("r1")
|
|
assert await repo.get("r1") is None
|
|
await _cleanup()
|
|
|
|
@pytest.mark.anyio
|
|
async def test_delete_nonexistent_is_noop(self, tmp_path):
|
|
repo = await _make_repo(tmp_path)
|
|
await repo.delete("nope") # should not raise
|
|
await _cleanup()
|
|
|
|
@pytest.mark.anyio
|
|
async def test_list_pending(self, tmp_path):
|
|
repo = await _make_repo(tmp_path)
|
|
await repo.put("r1", thread_id="t1", status="pending")
|
|
await repo.put("r2", thread_id="t1", status="running")
|
|
await repo.put("r3", thread_id="t2", status="pending")
|
|
pending = await repo.list_pending()
|
|
assert len(pending) == 2
|
|
assert all(r["status"] == "pending" for r in pending)
|
|
await _cleanup()
|
|
|
|
@pytest.mark.anyio
|
|
async def test_update_run_completion(self, tmp_path):
|
|
repo = await _make_repo(tmp_path)
|
|
await repo.put("r1", thread_id="t1", status="running")
|
|
await repo.update_run_completion(
|
|
"r1",
|
|
status="success",
|
|
total_input_tokens=100,
|
|
total_output_tokens=50,
|
|
total_tokens=150,
|
|
llm_call_count=2,
|
|
lead_agent_tokens=120,
|
|
subagent_tokens=20,
|
|
middleware_tokens=10,
|
|
message_count=3,
|
|
last_ai_message="The answer is 42",
|
|
first_human_message="What is the meaning?",
|
|
)
|
|
row = await repo.get("r1")
|
|
assert row["status"] == "success"
|
|
assert row["total_tokens"] == 150
|
|
assert row["llm_call_count"] == 2
|
|
assert row["lead_agent_tokens"] == 120
|
|
assert row["message_count"] == 3
|
|
assert row["last_ai_message"] == "The answer is 42"
|
|
assert row["first_human_message"] == "What is the meaning?"
|
|
await _cleanup()
|
|
|
|
@pytest.mark.anyio
|
|
async def test_metadata_preserved(self, tmp_path):
|
|
repo = await _make_repo(tmp_path)
|
|
await repo.put("r1", thread_id="t1", metadata={"key": "value"})
|
|
row = await repo.get("r1")
|
|
assert row["metadata"] == {"key": "value"}
|
|
await _cleanup()
|
|
|
|
@pytest.mark.anyio
|
|
async def test_kwargs_with_non_serializable(self, tmp_path):
|
|
"""kwargs containing non-JSON-serializable objects should be safely handled."""
|
|
repo = await _make_repo(tmp_path)
|
|
|
|
class Dummy:
|
|
pass
|
|
|
|
await repo.put("r1", thread_id="t1", kwargs={"obj": Dummy()})
|
|
row = await repo.get("r1")
|
|
assert "obj" in row["kwargs"]
|
|
await _cleanup()
|
|
|
|
@pytest.mark.anyio
|
|
async def test_update_run_completion_preserves_existing_fields(self, tmp_path):
|
|
"""update_run_completion does not overwrite thread_id or assistant_id."""
|
|
repo = await _make_repo(tmp_path)
|
|
await repo.put("r1", thread_id="t1", assistant_id="agent1", status="running")
|
|
await repo.update_run_completion("r1", status="success", total_tokens=100)
|
|
row = await repo.get("r1")
|
|
assert row["thread_id"] == "t1"
|
|
assert row["assistant_id"] == "agent1"
|
|
assert row["total_tokens"] == 100
|
|
await _cleanup()
|
|
|
|
@pytest.mark.anyio
|
|
async def test_list_by_thread_ordered_desc(self, tmp_path):
|
|
"""list_by_thread returns newest first."""
|
|
repo = await _make_repo(tmp_path)
|
|
await repo.put("r1", thread_id="t1", created_at="2024-01-01T00:00:00+00:00")
|
|
await repo.put("r2", thread_id="t1", created_at="2024-01-02T00:00:00+00:00")
|
|
rows = await repo.list_by_thread("t1")
|
|
assert rows[0]["run_id"] == "r2"
|
|
assert rows[1]["run_id"] == "r1"
|
|
await _cleanup()
|
|
|
|
@pytest.mark.anyio
|
|
async def test_list_by_thread_limit(self, tmp_path):
|
|
repo = await _make_repo(tmp_path)
|
|
for i in range(5):
|
|
await repo.put(f"r{i}", thread_id="t1")
|
|
rows = await repo.list_by_thread("t1", limit=2)
|
|
assert len(rows) == 2
|
|
await _cleanup()
|
|
|
|
@pytest.mark.anyio
|
|
async def test_owner_none_returns_all(self, tmp_path):
|
|
repo = await _make_repo(tmp_path)
|
|
await repo.put("r1", thread_id="t1", owner_id="alice")
|
|
await repo.put("r2", thread_id="t1", owner_id="bob")
|
|
rows = await repo.list_by_thread("t1", owner_id=None)
|
|
assert len(rows) == 2
|
|
await _cleanup()
|