From 0823371ecbfbf91c19806a0e3c5578a040b27fe9 Mon Sep 17 00:00:00 2001 From: Devin AI <158243242+devin-ai-integration[bot]@users.noreply.github.com> Date: Wed, 5 Feb 2025 09:51:51 +0000 Subject: [PATCH] chore: increase max_tokens to 8192 for local LLM Co-Authored-By: Han Xiao --- src/utils/llm-client.ts | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/utils/llm-client.ts b/src/utils/llm-client.ts index fd8ad31..1a529d8 100644 --- a/src/utils/llm-client.ts +++ b/src/utils/llm-client.ts @@ -84,7 +84,7 @@ export class LocalLLMClient implements LLMClient { type: 'json_schema', json_schema: config.generationConfig?.responseSchema, }, - max_tokens: 1000, + max_tokens: 8192, stream: false, }), });