chrysopedia/backend/pipeline
jlightner d58194ff96 feat: Store LLM request params (max_tokens, model, modality) in pipeline events
- _make_llm_callback now accepts request_params dict
- All 6 LLM call sites pass max_tokens, model_override, modality, response_model, hard_limit
- request_params stored in payload JSONB on every llm_call event (always, not just debug mode)
- Frontend JSON export includes full payload + request_params at top level
- DebugPayloadViewer shows 'Request Params' section even with debug mode off
- Answers whether max_tokens is actually being sent on pipeline requests
2026-04-01 07:01:57 +00:00
..
__init__.py fix: restore complete project tree from ub01 canonical state 2026-03-31 02:10:41 +00:00
embedding_client.py fix: restore complete project tree from ub01 canonical state 2026-03-31 02:10:41 +00:00
llm_client.py feat: Truncation detection, batched classification, and pipeline auto-resume 2026-03-31 17:48:19 -05:00
qdrant_client.py fix: Parallel search with match_context, deterministic Qdrant IDs, raised embedding timeout 2026-04-01 06:54:34 +00:00
schemas.py fix: Moment-to-page linking via moment_indices in stage 5 synthesis 2026-04-01 00:34:37 -05:00
stages.py feat: Store LLM request params (max_tokens, model, modality) in pipeline events 2026-04-01 07:01:57 +00:00