diff --git a/frontend/src/utils/verified-models.ts b/frontend/src/utils/verified-models.ts index 12453c6c86..5b2e19e9ef 100644 --- a/frontend/src/utils/verified-models.ts +++ b/frontend/src/utils/verified-models.ts @@ -59,6 +59,7 @@ export const VERIFIED_ANTHROPIC_MODELS = [ "claude-haiku-4-5-20251001", "claude-opus-4-20250514", "claude-opus-4-1-20250805", + "claude-opus-4-5-20251101", ]; // LiteLLM does not return the compatible Mistral models with the provider, so we list them here to set them ourselves diff --git a/openhands/llm/llm.py b/openhands/llm/llm.py index b94ed3bc2b..150fa54925 100644 --- a/openhands/llm/llm.py +++ b/openhands/llm/llm.py @@ -188,12 +188,14 @@ class LLM(RetryMixin, DebugMixin): if 'claude-opus-4-1' in self.config.model.lower(): kwargs['thinking'] = {'type': 'disabled'} - # Anthropic constraint: Opus 4.1 and Sonnet 4 models cannot accept both temperature and top_p + # Anthropic constraint: Opus 4.1, Opus 4.5, and Sonnet 4 models cannot accept both temperature and top_p # Prefer temperature (drop top_p) if both are specified. _model_lower = self.config.model.lower() - # Apply to Opus 4.1 and Sonnet 4 models to avoid API errors + # Apply to Opus 4.1, Opus 4.5, and Sonnet 4 models to avoid API errors if ( - ('claude-opus-4-1' in _model_lower) or ('claude-sonnet-4' in _model_lower) + ('claude-opus-4-1' in _model_lower) + or ('claude-opus-4-5' in _model_lower) + or ('claude-sonnet-4' in _model_lower) ) and ('temperature' in kwargs and 'top_p' in kwargs): kwargs.pop('top_p', None) diff --git a/tests/unit/llm/test_llm.py b/tests/unit/llm/test_llm.py index dfdb4e05b4..b04425e631 100644 --- a/tests/unit/llm/test_llm.py +++ b/tests/unit/llm/test_llm.py @@ -1255,6 +1255,25 @@ def test_opus_41_keeps_temperature_top_p(mock_completion): assert 'top_p' not in call_kwargs +@patch('openhands.llm.llm.litellm_completion') +def test_opus_45_keeps_temperature_drops_top_p(mock_completion): + mock_completion.return_value = { + 'choices': [{'message': {'content': 'ok'}}], + } + config = LLMConfig( + model='anthropic/claude-opus-4-5-20251101', + api_key='k', + temperature=0.7, + top_p=0.9, + ) + llm = LLM(config, service_id='svc') + llm.completion(messages=[{'role': 'user', 'content': 'hi'}]) + call_kwargs = mock_completion.call_args[1] + assert call_kwargs.get('temperature') == 0.7 + # Anthropic rejects both temperature and top_p together on Opus 4.5; we keep temperature and drop top_p + assert 'top_p' not in call_kwargs + + @patch('openhands.llm.llm.litellm_completion') def test_sonnet_4_keeps_temperature_drops_top_p(mock_completion): mock_completion.return_value = {