Bump condenser defaults: max_size 120->240 (#12267)

Co-authored-by: openhands <openhands@all-hands.dev>
This commit is contained in:
Xingyao Wang
2026-01-06 16:43:23 -05:00
committed by GitHub
parent 1907ebeaa8
commit af5c22700c
6 changed files with 71 additions and 10 deletions

View File

@@ -380,7 +380,7 @@ class AppConversationServiceBase(AppConversationService, ABC):
Returns:
Configured LLMSummarizingCondenser instance
"""
# LLMSummarizingCondenser has defaults: max_size=120, keep_first=4
# LLMSummarizingCondenser SDK defaults: max_size=240, keep_first=2
condenser_kwargs = {
'llm': llm.model_copy(
update={

View File

@@ -234,7 +234,7 @@ class WebSession:
# The order matters: with the browser output first, the summarizer
# will only see the most recent browser output, which should keep
# the summarization cost down.
max_events_for_condenser = settings.condenser_max_size or 120
max_events_for_condenser = settings.condenser_max_size or 240
default_condenser_config = CondenserPipelineConfig(
condensers=[
ConversationWindowCondenserConfig(),