Skip to content

Issues: ggerganov/llama.cpp

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

How to evaluate my converted gguf model ? What all benchmark i can run and how to run on my converted model? bug-unconfirmed low severity Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#8409 opened Jul 10, 2024 by RakshitAralimatti
Bug: InternLM 2.5 Chat Tool Calls: Incorrect and Inconsistent Formatting bug-unconfirmed low severity Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#8405 opened Jul 10, 2024 by apresence
Bug: tokenizer.chat_template missing from key/values bug-unconfirmed medium severity Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#8403 opened Jul 9, 2024 by bandoti
Bug: Failed to create server socket bug-unconfirmed high severity Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#8401 opened Jul 9, 2024 by oldmanjk
Bug: Renaming of executables breaks tutorials bug-unconfirmed low severity Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#8397 opened Jul 9, 2024 by gpacix
Bug: Model ignores system prompt when use /completion endpoint bug-unconfirmed medium severity Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#8393 opened Jul 9, 2024 by andreys42
Bug: [SYCL] Qwen2 MoE: 0 layers offloaded to GPU bug-unconfirmed medium severity Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#8387 opened Jul 9, 2024 by ch1y0q
k cache quantization enhancement New feature or request
#8385 opened Jul 9, 2024 by yancaoweidaode
4 tasks done
Bug: Misplaced docs/token_generation_performance_tips.md or link broken bug-unconfirmed low severity Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#8381 opened Jul 9, 2024 by dspasyuk
Bug: ggml/src/ggml.c: In function 'ggml_vec_mad_f16': bug-unconfirmed medium severity Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#8378 opened Jul 8, 2024 by oldmanjk
Bug: Yi 1.5 segmentation fault bug-unconfirmed medium severity Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#8369 opened Jul 8, 2024 by SimplyCorbett
server : support content array in OAI chat API enhancement New feature or request good first issue Good for newcomers server/api
#8367 opened Jul 8, 2024 by ggerganov
Bug: Build llama.cpp with DGGML_VULKAN=ON on Ubuntu 20.04 of arch64, the result of the program is not correct. bug-unconfirmed high severity Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#8365 opened Jul 8, 2024 by warren-lei
Feature Request: support dynamic ntk rope scaling enhancement New feature or request
#8361 opened Jul 8, 2024 by RunningLeon
4 tasks done
Feature Request: add indicator for which of MMQ vs cuBLAS is used enhancement New feature or request
#8350 opened Jul 7, 2024 by isaac-mcfadyen
4 tasks done
Encounter the "newline in constant" error while compiling with MSVC bug-unconfirmed medium severity Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#8334 opened Jul 6, 2024 by Yan-Xiangjun
Bug: Falcon 2 11B GGUF conversion works, but fails to read 8K context length from config.json bug-unconfirmed medium severity Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#8330 opened Jul 5, 2024 by jim-plus
Adding NVPL BLAS support enhancement New feature or request
#8329 opened Jul 5, 2024 by nicholaiTukanov
4 tasks done
Huge performance degradation using latest branch on Intel Core Ultra 7 155H bug-unconfirmed medium severity Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#8328 opened Jul 5, 2024 by aahouzi
Bug: make error bug-unconfirmed low severity Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#8313 opened Jul 5, 2024 by lorihuang
Feature Request: Support for Meta: Multi Token Prediction Models enhancement New feature or request
#8297 opened Jul 4, 2024 by sorasoras
4 tasks done
Bug: [SYCL] Inference not working correctly on multiple GPUs bug-unconfirmed high severity Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#8294 opened Jul 4, 2024 by ch1y0q
Feature Request: support for Gemini Nano? enhancement New feature or request
#8289 opened Jul 4, 2024 by flatsiedatsie
4 tasks done
Add support for InternLM 2.5 1M context. Should be as good as command r+ enhancement New feature or request
#8285 opened Jul 4, 2024 by mirek190
4 tasks done
ProTip! Exclude everything labeled bug with -label:bug.