This guide is for people working on OpenDevin and editing the source code.
- Linux, Mac OS, or WSL on Windows
- Docker(For those on MacOS, make sure to allow the default Docker socket to be used from advanced settings!)
- Python >= 3.11
- NodeJS >= 18.17.1
- Poetry >= 1.8
Make sure you have all these dependencies installed before moving on to make build
.
If you want to develop without system admin/sudo access to upgrade/install Python
and/or NodeJs
, you can use conda
or mamba
to manage the packages for you:
# Download and install Mamba (a faster version of conda)
curl -L -O "https://github.com/conda-forge/miniforge/releases/latest/download/Miniforge3-$(uname)-$(uname -m).sh"
bash Miniforge3-$(uname)-$(uname -m).sh
# Install Python 3.11, nodejs, and poetry
mamba install python=3.11
mamba install conda-forge::nodejs
mamba install conda-forge::poetry
- Build the Project: Begin by building the project, which includes setting up the environment and installing dependencies. This step ensures that OpenDevin is ready to run smoothly on your system.
make build
OpenDevin supports a diverse array of Language Models (LMs) through the powerful litellm library. By default, we've chosen the mighty GPT-4 from OpenAI as our go-to model, but the world is your oyster! You can unleash the potential of Anthropic's suave Claude, the enigmatic Llama, or any other LM that piques your interest.
To configure the LM of your choice, follow these steps:
- Using the Makefile: The Effortless Approach
With a single command, you can have a smooth LM setup for your OpenDevin experience. Simply run:
This command will prompt you to enter the LLM API key, model name, and other variables ensuring that OpenDevin is tailored to your specific needs. Note that the model name will apply only when you run headless. If you use the UI, please set the model in the UI.
make setup-config
Note on Alternative Models: Some alternative models may prove more challenging to tame than others. Fear not, brave adventurer! We shall soon unveil LLM-specific documentation to guide you on your quest. And if you've already mastered the art of wielding a model other than OpenAI's GPT, we encourage you to share your setup instructions with us.
For a full list of the LM providers and models available, please consult the litellm documentation.
There is also documentation for running with local models using ollama.
- Run the Application: Once the setup is complete, launching OpenDevin is as simple as running a single command. This command starts both the backend and frontend servers seamlessly, allowing you to interact with OpenDevin without any hassle.
make run
-
Start the Backend Server: If you prefer, you can start the backend server independently to focus on backend-related tasks or configurations.
make start-backend
-
Start the Frontend Server: Similarly, you can start the frontend server on its own to work on frontend-related components or interface enhancements.
make start-frontend
If you encounter any issues with the Language Model (LM) or you're simply curious, you can inspect the actual LLM prompts and responses. To do so, export DEBUG=1 in the environment and restart the backend. OpenDevin will then log the prompts and responses in the logs/llm/CURRENT_DATE directory, allowing you to identify the causes.
- Get Some Help: Need assistance or information on available targets and commands? The help command provides all the necessary guidance to ensure a smooth experience with OpenDevin.
make help
poetry run pytest ./tests/unit/test_sandbox.py
Please refer to this README for details.