change top_p default value to 1.0 (#3983)

This commit is contained in:
Xingyao Wang
2024-09-21 13:00:18 -05:00
committed by GitHub
parent 01462e11d7
commit 402a03cb9a
3 changed files with 3 additions and 4 deletions

View File

@@ -204,7 +204,6 @@ class CodeActAgent(Agent):
'</execute_bash>',
'</execute_browse>',
],
'temperature': 0.0,
}
if self.llm.is_caching_prompt_active():

View File

@@ -159,7 +159,7 @@ model = "gpt-4o"
#timeout = 0
# Top p for the API
#top_p = 0.5
#top_p = 1.0
# If model is vision capable, this option allows to disable image processing (useful for cost reduction).
#disable_vision = true

View File

@@ -71,8 +71,8 @@ class LLMConfig:
retry_max_wait: int = 120
timeout: int | None = None
max_message_chars: int = 10_000 # maximum number of characters in an observation's content when sent to the llm
temperature: float = 0
top_p: float = 0.5
temperature: float = 0.0
top_p: float = 1.0
custom_llm_provider: str | None = None
max_input_tokens: int | None = None
max_output_tokens: int | None = None