-
Notifications
You must be signed in to change notification settings - Fork 947
Issues: abetlen/llama-cpp-python
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
low level examples broken after [feat: Update sampling API for llama.cpp (#1742)]
#1803
opened Oct 20, 2024 by
mite51
Llama.from_pretrained
should work with HF_HUB_OFFLINE=1
#1801
opened Oct 16, 2024 by
davidgilbertson
server: chat completions returns wrong logprobs model
#1787
opened Oct 6, 2024 by
domdomegg
4 tasks done
Tool parser cannot analysis tool calls string from qwen2.5.
#1784
opened Oct 5, 2024 by
hpx502766238
Why is this not working for the current release. UNABLE TO USE GPU
#1781
opened Oct 2, 2024 by
AnirudhJM24
Setting temperature to 100000000000000000 does not affect output.
#1773
opened Oct 1, 2024 by
ivanstepanovftw
Error when passing model to deepcopy in llama_cpp_python>=0.3.0
#1769
opened Sep 28, 2024 by
sergey21000
Inference Speed is Extremely Slow for 72B Model with Long Contexts
#1767
opened Sep 27, 2024 by
wrench1997
FileNotFoundError: Shared library with base name 'llama' not found
#1764
opened Sep 26, 2024 by
HAOYON-666
Feature request: ability to tokenize a list of strings _or_ keep the tokenizer warm
enhancement
New feature or request
#1763
opened Sep 25, 2024 by
lsorber
Llama.embed
crashes when n_batch
> 512
bug
#1762
opened Sep 25, 2024 by
lsorber
Server crash with exceed context | lib version >= v0.2.81
#1759
opened Sep 25, 2024 by
carlostomazin
4 tasks done
chatml-function-calling chat format fails to generate multi calls to the same tool
#1756
opened Sep 23, 2024 by
jeffmaury
4 tasks done
Previous Next
ProTip!
What’s not been updated in a month: updated:<2024-09-20.