Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Provide a method of caching for the large wheels. #6591

Open
jfmherokiller opened this issue Dec 19, 2024 · 0 comments
Open

Provide a method of caching for the large wheels. #6591

jfmherokiller opened this issue Dec 19, 2024 · 0 comments
Labels
enhancement New feature or request

Comments

@jfmherokiller
Copy link

jfmherokiller commented Dec 19, 2024

Description

Setup a download setup (if possible) wherein the big wheels:
llama_cpp_python_cuda_tensorcores
exllamav2
llama_cpp_python_cuda
are downloaded once and then hash checked locally instead of being redownloaded every single time.

This should be done because these combined wheels are about 1GB and fail to be cached by pip. This is seemingly due to them being hosted by GitHub.

I will admit this seems to be a general issue with pip as seen here https://discuss.python.org/t/what-are-the-caching-rules-for-wheels-installed-from-urls/21594/2

@jfmherokiller jfmherokiller added the enhancement New feature or request label Dec 19, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant