fix: Tweak prompting behavior of LLMSummarizingCondenser (#7695)

Co-authored-by: Calvin Smith <calvin@all-hands.dev>
This commit is contained in:
Calvin Smith
2025-04-03 16:06:10 -06:00
committed by GitHub
parent 8bceee9e42
commit cc1aadaba5
2 changed files with 58 additions and 15 deletions

View File

@@ -75,6 +75,10 @@ class LLMSummarizingCondenserConfig(BaseModel):
description='Maximum size of the condensed history before triggering forgetting.',
ge=2,
)
max_event_length: int = Field(
default=10_000,
description='Maximum length of the event representations to be passed to the LLM.',
)
model_config = {'extra': 'forbid'}