mirror of
https://github.com/All-Hands-AI/OpenHands.git
synced 2026-01-10 07:18:10 -05:00
Add top_k (#8480)
This commit is contained in:
@@ -28,6 +28,7 @@ class LLMConfig(BaseModel):
|
||||
max_message_chars: The approximate max number of characters in the content of an event included in the prompt to the LLM. Larger observations are truncated.
|
||||
temperature: The temperature for the API.
|
||||
top_p: The top p for the API.
|
||||
top_k: The top k for the API.
|
||||
custom_llm_provider: The custom LLM provider to use. This is undocumented in openhands, and normally not used. It is documented on the litellm side.
|
||||
max_input_tokens: The maximum number of input tokens. Note that this is currently unused, and the value at runtime is actually the total tokens in OpenAI (e.g. 128,000 tokens for GPT-4).
|
||||
max_output_tokens: The maximum number of output tokens. This is sent to the LLM.
|
||||
@@ -66,6 +67,7 @@ class LLMConfig(BaseModel):
|
||||
) # maximum number of characters in an observation's content when sent to the llm
|
||||
temperature: float = Field(default=0.0)
|
||||
top_p: float = Field(default=1.0)
|
||||
top_k: float | None = Field(default=None)
|
||||
custom_llm_provider: str | None = Field(default=None)
|
||||
max_input_tokens: int | None = Field(default=None)
|
||||
max_output_tokens: int | None = Field(default=None)
|
||||
|
||||
Reference in New Issue
Block a user