-
Notifications
You must be signed in to change notification settings - Fork 8.8k
Issues: ggerganov/llama.cpp
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
How to evaluate my converted gguf model ? What all benchmark i can run and how to run on my converted model?
bug-unconfirmed
low severity
Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#8409
opened Jul 10, 2024 by
RakshitAralimatti
Bug: InternLM 2.5 Chat Tool Calls: Incorrect and Inconsistent Formatting
bug-unconfirmed
low severity
Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#8405
opened Jul 10, 2024 by
apresence
Bug: tokenizer.chat_template missing from key/values
bug-unconfirmed
medium severity
Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#8403
opened Jul 9, 2024 by
bandoti
Bug: Failed to create server socket
bug-unconfirmed
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#8401
opened Jul 9, 2024 by
oldmanjk
Bug: Renaming of executables breaks tutorials
bug-unconfirmed
low severity
Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#8397
opened Jul 9, 2024 by
gpacix
Bug: Model ignores system prompt when use Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
/completion
endpoint
bug-unconfirmed
medium severity
#8393
opened Jul 9, 2024 by
andreys42
Bug: [SYCL] Qwen2 MoE: 0 layers offloaded to GPU
bug-unconfirmed
medium severity
Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#8387
opened Jul 9, 2024 by
ch1y0q
k cache quantization
enhancement
New feature or request
#8385
opened Jul 9, 2024 by
yancaoweidaode
4 tasks done
Bug: Misplaced docs/token_generation_performance_tips.md or link broken
bug-unconfirmed
low severity
Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#8381
opened Jul 9, 2024 by
dspasyuk
Bug: ggml/src/ggml.c: In function 'ggml_vec_mad_f16':
bug-unconfirmed
medium severity
Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#8378
opened Jul 8, 2024 by
oldmanjk
Bug: Yi 1.5 segmentation fault
bug-unconfirmed
medium severity
Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#8369
opened Jul 8, 2024 by
SimplyCorbett
server : support content array in OAI chat API
enhancement
New feature or request
good first issue
Good for newcomers
server/api
#8367
opened Jul 8, 2024 by
ggerganov
Bug: Build llama.cpp with DGGML_VULKAN=ON on Ubuntu 20.04 of arch64, the result of the program is not correct.
bug-unconfirmed
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#8365
opened Jul 8, 2024 by
warren-lei
Feature Request: support dynamic ntk rope scaling
enhancement
New feature or request
#8361
opened Jul 8, 2024 by
RunningLeon
4 tasks done
Steering sampling and decoding strategy with
num_beams
and do_sample
#8353
opened Jul 7, 2024 by
d-kleine
Feature Request: add indicator for which of MMQ vs cuBLAS is used
enhancement
New feature or request
#8350
opened Jul 7, 2024 by
isaac-mcfadyen
4 tasks done
Encounter the "newline in constant" error while compiling with MSVC
bug-unconfirmed
medium severity
Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#8334
opened Jul 6, 2024 by
Yan-Xiangjun
Bug: Falcon 2 11B GGUF conversion works, but fails to read 8K context length from config.json
bug-unconfirmed
medium severity
Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#8330
opened Jul 5, 2024 by
jim-plus
Adding NVPL BLAS support
enhancement
New feature or request
#8329
opened Jul 5, 2024 by
nicholaiTukanov
4 tasks done
Huge performance degradation using latest branch on Intel Core Ultra 7 155H
bug-unconfirmed
medium severity
Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#8328
opened Jul 5, 2024 by
aahouzi
Bug: make error
bug-unconfirmed
low severity
Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#8313
opened Jul 5, 2024 by
lorihuang
Feature Request: Support for Meta: Multi Token Prediction Models
enhancement
New feature or request
#8297
opened Jul 4, 2024 by
sorasoras
4 tasks done
Bug: [SYCL] Inference not working correctly on multiple GPUs
bug-unconfirmed
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#8294
opened Jul 4, 2024 by
ch1y0q
Feature Request: support for Gemini Nano?
enhancement
New feature or request
#8289
opened Jul 4, 2024 by
flatsiedatsie
4 tasks done
Add support for InternLM 2.5 1M context. Should be as good as command r+
enhancement
New feature or request
#8285
opened Jul 4, 2024 by
mirek190
4 tasks done
Previous Next
ProTip!
Exclude everything labeled
bug
with -label:bug.