-
Notifications
You must be signed in to change notification settings - Fork 11.6k
Issues: ggml-org/llama.cpp
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
sampling: Integrate Top-nσ into main sampling chain (and add it to the server)
examples
server
#13264
opened May 2, 2025 by
oobabooga
Loading…
Support start strings, the opposite of stop tokens.
examples
python
python script changes
server
#13214
opened Apr 30, 2025 by
matteoserva
Loading…
Support jinja extra template kwargs (Qwen3 enable_thinking feature), from command line and from client
examples
server
#13196
opened Apr 29, 2025 by
matteoserva
Loading…
set b = ub when b > ub with embedding
examples
server
#12940
opened Apr 14, 2025 by
ahmedshakill
Loading…
server : crash when -b > -ub with embeddings
bug
Something isn't working
embeddings
embedding related topics
good first issue
Good for newcomers
server
#12836
opened Apr 8, 2025 by
ggerganov
common
: add partial regex support
examples
server
testing
#12808
opened Apr 7, 2025 by
ochafik
Loading…
server
: inject date_string in llama 3.x template + fix date for firefunction v2
examples
python
#12802
opened Apr 7, 2025 by
ochafik
Loading…
WIP: Add support for CogAgent
examples
python
python script changes
server
#12679
opened Mar 31, 2025 by
Tianyue-Zhao
•
Draft
llama-server : implement universal assisted decoding
examples
server
#12635
opened Mar 28, 2025 by
g2mt
Loading…
server
: streaming of tool calls and thoughts when --jinja
is on
documentation
tool-call
: Phi-4 support
android
#12288
opened Mar 9, 2025 by
jpohhhh
Loading…
server webui easy config selection
demo
Demonstrate some concept or idea, not intended to be merged
examples
server
#12031
opened Feb 22, 2025 by
poulphunter
Loading…
sampling: add Top-nσ sampler to
llama-server
examples
server
#11896
opened Feb 15, 2025 by
CasualAutopsy
Loading…
server (webui): Fix issue with muliple
<think>
tags in response
examples
server
#11779
opened Feb 10, 2025 by
stduhpf
Loading…
When llama_chat_apply_template doesn't work
examples
server
#11687
opened Feb 5, 2025 by
ericcurtin
Loading…
Previous Next
ProTip!
What’s not been updated in a month: updated:<2025-04-02.