-
Notifications
You must be signed in to change notification settings - Fork 13
Setup
Cielo edited this page Jul 29, 2022
·
3 revisions
This "guide" assumes that you cloned the repository on your machine and that you have Docker, docker-compose and appropriate CUDA drivers installed if you wish to leverage your GPU for inference.
- On the first run,
docker-compose up
should configure the FastAPI app and PostgreSQL. Uvicorn options are set toreload=True
, so any code changes you make on your host machine will be automatically applied to the container.
NOTE: If you have an appropriate NVIDIA GPU that you want to use, run docker-compose -f docker-compose_nvidia-gpu.yaml up
instead. Remember to install nvidia-container-runtime on your host machine.
- Run
docker-compose run app alembic upgrade head
to create db tables if you're setting up, or to apply migrations if you made any db schema changes.
kubectl apply -f k8s
- TODO
This is assuming that you're using Docker for Desktop Windows.
- Follow the steps from the Linux guide.
- Curl up into a ball because you're not using a Unix system.
- If you want to use your GPU, good luck. Refer to Docker and NVIDIA's guide to set up CUDA for WSL. If I were you, I'd just skip the pain and use distilgpt2 or gpt-j-random-tinier models for testing.