From 1d8402a14a8daa7ffba5049a92bee7ce97e733c0 Mon Sep 17 00:00:00 2001 From: mamoodi Date: Tue, 14 May 2024 13:36:53 -0400 Subject: [PATCH] doc: Small fixes to documentation (#1783) --- docs/modules/usage/about.md | 2 +- docs/modules/usage/intro.mdx | 2 +- docs/modules/usage/llms/llms.md | 2 +- docs/modules/usage/llms/localLLMs.md | 10 +++++----- docs/modules/usage/troubleshooting/windows.md | 4 ++-- 5 files changed, 10 insertions(+), 10 deletions(-) diff --git a/docs/modules/usage/about.md b/docs/modules/usage/about.md index edd31efc86..82c1d5a6f8 100644 --- a/docs/modules/usage/about.md +++ b/docs/modules/usage/about.md @@ -34,7 +34,7 @@ Now we have both Slack workspace for the collaboration on building OpenDevin and - [Slack workspace](https://join.slack.com/t/opendevin/shared_invite/zt-2ggtwn3k5-PvAA2LUmqGHVZ~XzGq~ILw) - [Discord server](https://discord.gg/ESHStjSjD4) -If you would love to contribute, feel free to join our community (note that now there is no need to fill in the [form](https://forms.gle/758d5p6Ve8r2nxxq6)). Let's simplify software engineering together! +If you would love to contribute, feel free to join our community. Let's simplify software engineering together! 🐚 **Code less, make more with OpenDevin.** diff --git a/docs/modules/usage/intro.mdx b/docs/modules/usage/intro.mdx index a08eb6aa52..40a4a5ddd7 100644 --- a/docs/modules/usage/intro.mdx +++ b/docs/modules/usage/intro.mdx @@ -60,7 +60,7 @@ Explore the codebase of OpenDevin on [GitHub](https://github.com/OpenDevin/OpenD The easiest way to run OpenDevin is inside a Docker container. -To start the app, run these commands, replacing `$(pwd)/workspace` with the path to the code you want OpenDevin to work with. +To start the app, run these commands, replacing `$(pwd)/workspace` with the directory you want OpenDevin to work with. ``` # Your OpenAI API key, or any other LLM API key diff --git a/docs/modules/usage/llms/llms.md b/docs/modules/usage/llms/llms.md index 43fdc01275..46bac01b29 100644 --- a/docs/modules/usage/llms/llms.md +++ b/docs/modules/usage/llms/llms.md @@ -44,4 +44,4 @@ are actively working on building better open source models! Some LLMs have rate limits and may require retries. OpenDevin will automatically retry requests if it receives a 429 error or API connection error. You can set `LLM_NUM_RETRIES`, `LLM_RETRY_MIN_WAIT`, `LLM_RETRY_MAX_WAIT` environment variables to control the number of retries and the time between retries. -By default, `LLM_NUM_RETRIES` is 5 and `LLM_RETRY_MIN_WAIT`, `LLM_RETRY_MAX_WAIT` are 3 seconds and respectively 60 seconds. +By default, `LLM_NUM_RETRIES` is 5 and `LLM_RETRY_MIN_WAIT`, `LLM_RETRY_MAX_WAIT` are 3 seconds and 60 seconds respectively. diff --git a/docs/modules/usage/llms/localLLMs.md b/docs/modules/usage/llms/localLLMs.md index 46bae30307..5da05d9839 100644 --- a/docs/modules/usage/llms/localLLMs.md +++ b/docs/modules/usage/llms/localLLMs.md @@ -1,9 +1,9 @@ # Local LLM with Ollama Ensure that you have the Ollama server up and running. -For detailed startup instructions, refer to the [here](https://github.com/ollama/ollama) +For detailed startup instructions, refer to [here](https://github.com/ollama/ollama) -This guide assumes you've started ollama with `ollama serve`. If you're running ollama differently (e.g. inside docker), the instructions might need to be modified. Please note that if you're running wsl the default ollama configuration blocks requests from docker containers. See [here](#4-configuring-the-ollama-service-wsl). +This guide assumes you've started ollama with `ollama serve`. If you're running ollama differently (e.g. inside docker), the instructions might need to be modified. Please note that if you're running WSL the default ollama configuration blocks requests from docker containers. See [here](#configuring-the-ollama-service-wsl). ## Pull Models @@ -87,7 +87,7 @@ And now you're ready to go! ## Configuring the ollama service (WSL) -The default configuration for ollama in wsl only serves localhost. This means you can't reach it from a docker container. eg. it wont work with OpenDevin. First let's test that ollama is running correctly. +The default configuration for ollama in WSL only serves localhost. This means you can't reach it from a docker container. eg. it wont work with OpenDevin. First let's test that ollama is running correctly. ```bash ollama list # get list of installed models @@ -96,7 +96,7 @@ curl http://localhost:11434/api/generate -d '{"model":"[NAME]","prompt":"hi"}' #ex. curl http://localhost:11434/api/generate -d '{"model":"codellama","prompt":"hi"}' #the tag is optional if there is only one ``` -Once that is done test that it allows "outside" requests, like those from inside a docker container. +Once that is done, test that it allows "outside" requests, like those from inside a docker container. ```bash docker ps # get list of running docker containers, for most accurate test choose the open devin sandbox container. @@ -106,7 +106,7 @@ docker exec [CONTAINER ID] curl http://host.docker.internal:11434/api/generate - ## Fixing it -Now let's make it work, edit /etc/systemd/system/ollama.service with sudo privileges. (Path may vary depending on linux flavor) +Now let's make it work. Edit /etc/systemd/system/ollama.service with sudo privileges. (Path may vary depending on linux flavor) ```bash sudo vi /etc/systemd/system/ollama.service diff --git a/docs/modules/usage/troubleshooting/windows.md b/docs/modules/usage/troubleshooting/windows.md index 066d33445d..0bb904b007 100644 --- a/docs/modules/usage/troubleshooting/windows.md +++ b/docs/modules/usage/troubleshooting/windows.md @@ -7,7 +7,7 @@ Please be sure to run all commands inside your WSL terminal. ### Failed to create opendevin user -If you encounter the following error during setup: `Exception: Failed to create opendevin user in sandbox: b'useradd: UID 0 is not unique\n'` +If you encounter the following error during setup: `Exception: Failed to create opendevin user in sandbox: b'useradd: UID 0 is not unique\n'`. You can resolve it by running: ` export SANDBOX_USER_ID=1000 ` @@ -20,7 +20,7 @@ If you face issues running Poetry even after installing it during the build proc ### NoneType object has no attribute 'request' -If you experiencing issues related to networking, such as `NoneType object has no attribute 'request'` when executing `make run`, you may need to configure your WSL2 networking settings. Follow these steps: +If you are experiencing issues related to networking, such as `NoneType object has no attribute 'request'` when executing `make run`, you may need to configure your WSL2 networking settings. Follow these steps: - Open or create the `.wslconfig` file located at `C:\Users\%username%\.wslconfig` on your Windows host machine. - Add the following configuration to the `.wslconfig` file: