src/dazzle_back/runtime/llm_queue.py¶
Generated: 2026-03-30T04:38:47.530192+00:00
Commit: 615f1354a103
Language: python
Scope: dazzle_back
Purpose¶
Background job queue for async LLM intent execution.
Provides: - In-process async queue with configurable workers - Per-model semaphore (concurrency cap) - Per-model token bucket (rate limiting) - AIJob lifecycle management (pending → running → completed/failed) - LLM event emission on completion/f...
Key Exports¶
TokenBucketLLMJobLLMJobQueueacquiresubmitstartshutdownpending_count
Dependencies¶
Imports¶
__future__.annotationsasynciologgingtimecollections.abc.Awaitablecollections.abc.Callabledataclasses.dataclasstyping.TYPE_CHECKINGtyping.Anydazzle_back.runtime.event_bus.LLMEventTypedazzle_back.runtime.llm_executor.ExecutionResultdazzle_back.runtime.llm_executor.LLMIntentExecutoruuiddazzle_back.runtime.event_bus.EntityEventdazzle_back.runtime.event_bus.EntityEventType
Event Interactions¶
Consumes: EntityEvent
Invariants and Assumptions¶
No invariants documented.
Tests¶
tests/unit/test_llm_queue.pytests/integration/test_runtime_e2e.pytests/integration/test_runtime_pipeline.pytests/unit/test_runtime_test_routes.pytests/unit/test_runtime_services.py
Notes¶
This page was automatically generated by tools/gen_reference_docs.py.