Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Provide a method of caching for the large wheels. #6591

Open
jfmherokiller opened this issue Dec 19, 2024 · 1 comment
Open

Provide a method of caching for the large wheels. #6591

jfmherokiller opened this issue Dec 19, 2024 · 1 comment
Labels
enhancement New feature or request

Comments

@jfmherokiller
Copy link

jfmherokiller commented Dec 19, 2024

Description

Setup a download setup (if possible) wherein the big wheels:
llama_cpp_python_cuda_tensorcores
exllamav2
llama_cpp_python_cuda
are downloaded once and then hash checked locally instead of being redownloaded every single time.

This should be done because these combined wheels are about 1GB and fail to be cached by pip. This is seemingly due to them being hosted by GitHub.

I will admit this seems to be a general issue with pip as seen here https://discuss.python.org/t/what-are-the-caching-rules-for-wheels-installed-from-urls/21594/2

@jfmherokiller jfmherokiller added the enhancement New feature or request label Dec 19, 2024
@jfmherokiller
Copy link
Author

while also admit this may be slightly streching the application design goal. Due to the fact that this application seems to want to be an everything ai gen adding some accomations for those who dont have unstable internet. Because I dread any time i need to update this application due to the download speeds.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant