mirror of
https://github.com/All-Hands-AI/OpenHands.git
synced 2026-01-08 22:38:05 -05:00
Bump condenser defaults: max_size 120->240 (#12267)
Co-authored-by: openhands <openhands@all-hands.dev>
This commit is contained in:
@@ -380,7 +380,7 @@ class AppConversationServiceBase(AppConversationService, ABC):
|
||||
Returns:
|
||||
Configured LLMSummarizingCondenser instance
|
||||
"""
|
||||
# LLMSummarizingCondenser has defaults: max_size=120, keep_first=4
|
||||
# LLMSummarizingCondenser SDK defaults: max_size=240, keep_first=2
|
||||
condenser_kwargs = {
|
||||
'llm': llm.model_copy(
|
||||
update={
|
||||
|
||||
@@ -234,7 +234,7 @@ class WebSession:
|
||||
# The order matters: with the browser output first, the summarizer
|
||||
# will only see the most recent browser output, which should keep
|
||||
# the summarization cost down.
|
||||
max_events_for_condenser = settings.condenser_max_size or 120
|
||||
max_events_for_condenser = settings.condenser_max_size or 240
|
||||
default_condenser_config = CondenserPipelineConfig(
|
||||
condensers=[
|
||||
ConversationWindowCondenserConfig(),
|
||||
|
||||
Reference in New Issue
Block a user