-
-
Notifications
You must be signed in to change notification settings - Fork 2.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Used docker image to implement the localGPT not work. #480
Comments
Yes, I get the same error message. python3.10 -m venv /mnt/yhteinen/docker-localgpt ..... 2023-09-17 05:54:52,701 - INFO - run_localGPT.py:212 - Running on: cuda |
Thanks for the report! I haven't tested the Dockerfile with the new GGUF models and will investigate! |
found. |
Oh I see, that used to be inside requirements.txt.
I will add it back in the Dockerfile |
Created PR at #484 but I currently don't have a GPU to test, please report here and in the PR if it works @sp1d3rino, @chiehpower, @finnishbroccoli. |
Finally I tested docker image on runpod.io and GPU worked but only with this model set in constants.py MODEL_ID = "TheBloke/WizardLM-13B-V1.2-GPTQ" |
Sure! thanks for updating. I will test it later and update here soon. |
@sp1d3rino What is the error when you use a GGUF model? |
@KonradHoeffner actually no errors, simply no GPU is used. Only if I use GPTQ model I can see GPU working |
Can you paste the log? |
@KonradHoeffner there is no log ... it works but uses only CPU. |
I mean the console output. |
this is GGUF and it doesn't work with GPU, but only with CPU. 2023-09-19 06:49:37,377 - INFO - run_localGPT.py:221 - Running on: cuda Enter a query: |
@sp1d3rino hm this shows BLAS=0, are you sure you checked out the correct branch and rebuild the Dockerfile? |
hmm but I compiled requirements with CMAKE_ARGS="-DLLAMA_CUBLAS=on" FORCE_CMAKE=1 |
hi @KonradHoeffner Although I noticed the GPU memory raising by almost ~2GB, I also found CPU memory almost full. The one query time takes sooooo long. |
What do you mean? Your locally installed python dependencies don't have any effect on Docker images. |
? in docker file there are that two env vars that you have to set before installl requirements with pip |
Sorry guys but, I just downloaded and executed the Docker file from master branch and GPU is not used in my case.
no signal os GPU usage same computer but executing another test (llama-cpp-python) it shows the usage of the GPU and returns faster:
|
Closed the issue that the bug was fixed. |
hi @KonradHoeffner and everyone,
I followed the
docker section
to build the docker image first, and start the container. However, it encountered this error below.Steps
docker build . -t localgpt
docker run -it --mount src="$HOME/.cache",target=/root/.cache,type=bind --gpus=all localgpt
Error message
Does anyone also encounter this problem?
The text was updated successfully, but these errors were encountered: