-
Notifications
You must be signed in to change notification settings - Fork 10.2k
Issues: ggerganov/llama.cpp
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Misc. bug: llama-server - shared libraries after build 4409, last working 4406
bug-unconfirmed
#11144
opened Jan 8, 2025 by
DaLiV
server : add support for multiple responses
server/api
server
#11142
opened Jan 8, 2025 by
ggerganov
DeepSeek Models (V2/V3) Hang with ROCm Backend
bug-unconfirmed
#11141
opened Jan 8, 2025 by
emuchogu
Feature Request: How can I use SAM GUFF model?
enhancement
New feature or request
#11133
opened Jan 7, 2025 by
zydjohnHotmail
4 tasks done
Eval bug: asymmetric layer splitting of reduced models on multiple CUDA GPUs
bug-unconfirmed
#11132
opened Jan 7, 2025 by
recallmenot
Feature Request: Vulkan: Implement CPY op for quantized types
enhancement
New feature or request
#11127
opened Jan 7, 2025 by
stduhpf
4 tasks done
Eval bug: llama-server is toooo slow when inference int8 model in reranker
bug-unconfirmed
#11114
opened Jan 7, 2025 by
Tian14267
Misc. bug: dylib is not allow for iOS and Vision OS release
bug-unconfirmed
#11113
opened Jan 6, 2025 by
Animaxx
Compile bug: parameter packs not expanded with ‘...’:
bug-unconfirmed
#11112
opened Jan 6, 2025 by
TheKoakuma
Eval bug: Bug: Infinite text generation with Meta-llama-8B GGUF model when using llama-cli
bug-unconfirmed
#11111
opened Jan 6, 2025 by
NKU-Yang
Feature Request: RuntimeError: Unsloth: The file 'llama.cpp/llama-quantize' or 'llama.cpp/quantize' does not exist. But we expect this file to exist! Maybe the llama.cpp developers changed the name?
enhancement
New feature or request
#11100
opened Jan 6, 2025 by
San-Jain18
4 tasks done
Vulkan related question: what's the different between server and cli?
#11099
opened Jan 6, 2025 by
FNsi
Misc. bug: ggml_backend_sycl_graph_compute: error: op not supported node_1586 (FLASH_ATTN_EXT)
bug-unconfirmed
#11084
opened Jan 5, 2025 by
alfrentgen
Misc. bug: [Mac M4]llama-server cannot run in release-4409 but can run in 4406
bug-unconfirmed
#11083
opened Jan 5, 2025 by
bobleer
Compile bug: Trying to compile on a raspi w v2 and failing to compile
bug-unconfirmed
#11079
opened Jan 4, 2025 by
RichNeese
Eval bug: ggml_sycl_cpy: unsupported type combination (q8_0 to f32)
bug-unconfirmed
#11078
opened Jan 4, 2025 by
paoletto
Feature Request: Add Parameters Section To change parameters in llama-android
enhancement
New feature or request
#11073
opened Jan 4, 2025 by
Dhruvanand24
4 tasks done
Compile bug: Error Domain=MTLLibraryErrorDomain Code=3
bug-unconfirmed
#11071
opened Jan 3, 2025 by
p-w-rs
Previous Next
ProTip!
Add no:assignee to see everything that’s not assigned.