Make drop_params default in llm_config (#4012)

This commit is contained in:
Graham Neubig
2024-09-23 16:57:10 -04:00
committed by GitHub
parent 3f6aa0d1f1
commit 73ded7de10
2 changed files with 3 additions and 5 deletions

View File

@@ -64,7 +64,7 @@ class LLMConfig:
input_cost_per_token: float | None = None
output_cost_per_token: float | None = None
ollama_base_url: str | None = None
drop_params: bool | None = None
drop_params: bool = True
disable_vision: bool | None = None
caching_prompt: bool = False
log_completions: bool = False

View File

@@ -132,9 +132,6 @@ class LLM:
):
self.config.max_output_tokens = self.model_info['max_tokens']
if self.config.drop_params:
litellm.drop_params = self.config.drop_params
# This only seems to work with Google as the provider, not with OpenRouter!
gemini_safety_settings = (
[
@@ -170,6 +167,7 @@ class LLM:
timeout=self.config.timeout,
temperature=self.config.temperature,
top_p=self.config.top_p,
drop_params=self.config.drop_params,
**(
{'safety_settings': gemini_safety_settings}
if gemini_safety_settings is not None
@@ -298,7 +296,7 @@ class LLM:
timeout=self.config.timeout,
temperature=self.config.temperature,
top_p=self.config.top_p,
drop_params=True,
drop_params=self.config.drop_params,
**(
{'safety_settings': gemini_safety_settings}
if gemini_safety_settings is not None