mirror of
https://github.com/dzhng/deep-research.git
synced 2026-03-22 07:57:16 +08:00
docker and readme tweaks
This commit is contained in:
11
README.md
11
README.md
@@ -111,10 +111,17 @@ To use local LLM, comment out `OPENAI_KEY` and instead uncomment `OPENAI_ENDPOIN
|
||||
1. Clone the repository
|
||||
2. Rename `.env.example` to `.env.local` and set your API keys
|
||||
|
||||
3. Run the Docker image:
|
||||
3. Run `npm install`
|
||||
|
||||
4. Run the Docker image:
|
||||
|
||||
```bash
|
||||
docker compose run --rm deep-research
|
||||
docker compose up -d
|
||||
```
|
||||
|
||||
5. Execute `npm run docker` in the docker service:
|
||||
```bash
|
||||
docker exec -it deep-research npm run docker
|
||||
```
|
||||
|
||||
## Usage
|
||||
|
||||
@@ -1,7 +1,10 @@
|
||||
services:
|
||||
deep-research:
|
||||
container_name: deep-research
|
||||
build: .
|
||||
env_file:
|
||||
- .env.local
|
||||
volumes:
|
||||
- ./:/app/
|
||||
tty: true
|
||||
stdin_open: true
|
||||
|
||||
@@ -13,7 +13,7 @@ const openai = createOpenAI({
|
||||
baseURL: process.env.OPENAI_ENDPOINT || 'https://api.openai.com/v1',
|
||||
} as CustomOpenAIProviderSettings);
|
||||
|
||||
const customModel = process.env.OPENAI_MODEL || 'o3-mini';
|
||||
const customModel = process.env.OPENAI_MODEL || 'gpt-4o-mini';
|
||||
|
||||
// Models
|
||||
|
||||
|
||||
Reference in New Issue
Block a user