-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Segmentation fault Windows 11 Docker #254
Comments
I do have the exact same problem. |
This causes long prompts to parse very slowly.
Just downloaded the repo and installed the 30B model, having the same issue. `/root/dalai/alpaca/main --seed -1 --threads 4 --n_predict 200 --model models/30B/ggml-model-q4_0.bin --top_k 40 --top_p 0.9 --temp 0.8 --repeat_last_n 64 --repeat_penalty 1.3 -p "Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request. Instruction:
Response:"
|
I also have this issue with alpaca 30B and llama 30B, exactly the same error (but the I have 32Go RAM, docker seems to consume a lot of it sometimes (via the So I bet 32Go of ram in not enough for running the 30B model using docker ? 🤔 How much do you have ? |
Here i described my experience running models on Windows 10 |
I have the assumption that the issue comes from the fact that this models requires a lot of RAM in your machine. Can anybody confirm or dismiss this? I believe when the model is loaded it is loaded in the RAM, that is the reason it breaks. |
In my case the context size was causing this issue, I fixed it by adding new config to the UI which allows me to play with context size. I was using 6 gb ram server to try it, in my case context size below 1024 seems to work without any errors. PR for the same: #424 |
I tried installing dalai with docker on windows. Currently I am getting the following error when I try generating a response with debug mode on:
Looking at the llama.cpp project it seems that they have tried to fix some segmentation problems but where unsuccessful. Perhaps this is the issue I am facing but I do not know. ggerganov/llama.cpp@3cd8dde
Any tips on how to debug this or to get a better error would be appreciated.
The text was updated successfully, but these errors were encountered: