fix: improve json parsing resilience and disable gemini thinking

- add jsonrepair fallback for truncated LLM output
- disable gemini built-in thinking mode (thinkingBudget: 0)
- increase token limits for errorAnalyzer, queryRewriter, serpCluster
- switch production default to gemini-2.5-flash-lite
- fix normalizeHostName to handle wildcard patterns
This commit is contained in:
Han Xiao
2025-12-13 12:07:37 +01:00
parent 579fd95fff
commit d44cec6524
6 changed files with 91 additions and 25 deletions

View File

@@ -34,31 +34,35 @@
"models": {
"gemini": {
"default": {
"model": "gemini-2.5-flash",
"model": "gemini-2.5-flash-lite",
"temperature": 0.6,
"maxTokens": 8000
},
"tools": {
"coder": {
"maxTokens": 2000,
"model": "gemini-2.5-flash-lite"
"maxTokens": 2000
},
"researchPlanner": {},
"evaluator": {
"maxTokens": 2000
},
"serpCluster": {},
"serpCluster": {
"maxTokens": 4000
},
"errorAnalyzer": {
"maxTokens": 1000
"maxTokens": 4000
},
"queryRewriter": {
"maxTokens": 2000
"maxTokens": 4000
},
"agent": {
"model": "gemini-2.5-flash"
},
"agentBeastMode": {
"model": "gemini-2.5-flash"
},
"agent": {},
"agentBeastMode": {},
"fallback": {
"maxTokens": 8000,
"model": "gemini-2.5-flash-lite"
"maxTokens": 8000
},
"finalizer": {},
"reducer": {