From cacb06de3147ee83881624c2f96efabb22896d45 Mon Sep 17 00:00:00 2001 From: Han Xiao Date: Fri, 7 Feb 2025 10:13:31 +0800 Subject: [PATCH] fix: #40 --- README.md | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/README.md b/README.md index 4573c70..dee4f3a 100644 --- a/README.md +++ b/README.md @@ -1,6 +1,6 @@ # DeepResearch -Keep searching, reading webpages, reasoning until it finds the answer (or exceeding the token budget). +Keep searching, reading webpages, reasoning until an answer is found (or the token budget is exceeded). ```mermaid --- @@ -95,6 +95,7 @@ If you use Ollama or LMStudio, you can redirect the reasoning request to your lo ```bash export LLM_PROVIDER=openai # yes, that's right - for local llm we still use openai client export OPENAI_BASE_URL=http://127.0.0.1:1234/v1 # your local llm endpoint +export OPENAI_API_KEY=whatever # random string would do, as we don't use it (unless your local LLM has authentication) export DEFAULT_MODEL_NAME=qwen2.5-7b # your local llm model name ```