docker-compose-wsl2-and-linux-nvidia-gpu-ai.yml #7
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Description
GPU Enabled AI Assistant Container for Linux and Windows OS with Docker under WSL2 (for Windows).
Explanation:
For this to work, you will need Linux and/or Windows with WSL2 (updated),
both OS must be equiped with the dedicated GPU which is usable for LLM
(with CUDA cores).
If this is the case, then ollama service will start with the GPU and AI
assistant will be super fast in responses compared to the CPU only.
Notebook users, be careful about the nVIDIA Optimus, that can break things.
If you have Optimus enabled notebook, just switch to dedicated GPU (sometimes
it may be required to do that in the BIOS directly).
Updated drivers must be installed.
Functional check can be done using: https://docs.docker.com/desktop/gpu/
More details (Linux): https://ollama.com/blog/ollama-is-now-available-as-an-official-docker-image
Type of change
GPU Enabled AI Assistant Container for Linux and Windows with Docker under WSL2.
How Has This Been Tested?
Several times tested with many different scenarios. Optimus enabled notebook issue described in explanation.