-
Notifications
You must be signed in to change notification settings - Fork 8.5k
Issues: ggerganov/llama.cpp
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Bug: Phi-3 mini output get weird after 2048 tokens
bug-unconfirmed
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#7709
opened Jun 3, 2024 by
Amadeus-AI
Could NOT find BLAS (missing: BLAS_LIBRARIES)Bug:
bug-unconfirmed
medium severity
Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#7708
opened Jun 3, 2024 by
vt-alt
Bug: convert-ht-to-gguf-update.py breaks in Windows Python 3.11.5
bug-unconfirmed
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#7706
opened Jun 3, 2024 by
jim-plus
Feature Request: Prevent server.exe from being detected as ? Trojan:Win32/Wacatac.B!ml
enhancement
New feature or request
#7704
opened Jun 2, 2024 by
ArEnSc
4 tasks done
Bug: JSON Schema Not Respected ?
bug-unconfirmed
medium severity
Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#7703
opened Jun 2, 2024 by
ArEnSc
Feature Request: GGUF 2 BIN
enhancement
New feature or request
#7695
opened Jun 2, 2024 by
0wwafa
4 tasks done
Bug: value of keep alive max count in cpp-httplib hardcoded too low
bug-unconfirmed
medium severity
Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#7694
opened Jun 2, 2024 by
skoulik
Bug: Incorrect memory allocation when mixing Nvidia and AMD GPU's
bug-unconfirmed
medium severity
Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#7674
opened May 31, 2024 by
C-monC
Bug: Phi-2 model tokenizer not recognized
bug-unconfirmed
medium severity
Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#7667
opened May 31, 2024 by
saeid93
Bug: The output of llama.cpp with Phi-3 contains Non-sense/meaningless words, Does anyone encounter the similar problem?
bug-unconfirmed
low severity
Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#7666
opened May 31, 2024 by
Ross-Fan
When using GPU (OpenCL), the reply speed is slower and all replies are incorrect??
bug-unconfirmed
medium severity
Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#7661
opened May 31, 2024 by
QIANXUNZDL123
Why is convert.py missing?
documentation
Improvements or additions to documentation
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
python
python script changes
script
Script related
#7658
opened May 31, 2024 by
David-AU-github
Refactor: Add CONTRIBUTING.md and/or update PR template with [no ci] tips
devops
improvements to build systems and github actions
documentation
Improvements or additions to documentation
enhancement
New feature or request
help wanted
Extra attention is needed
low severity
Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
#7657
opened May 30, 2024 by
mofosyne
Bug: No longer makes with w64devkit
bug-unconfirmed
critical severity
Used to report critical severity bugs in llama.cpp (e.g. Crashing, Corrupted, Dataloss)
#7655
opened May 30, 2024 by
gjnave
Add Support for Solidity Model
enhancement
New feature or request
#7653
opened May 30, 2024 by
kanema20
4 tasks done
Bug: DeepSeek-V2-Lite GGML_ASSERT: ggml-metal.m:1857: dst_rows <= 2048 and aborts
bug-unconfirmed
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#7652
opened May 30, 2024 by
hnfong
Bug: cant finetune
bug-unconfirmed
critical severity
Used to report critical severity bugs in llama.cpp (e.g. Crashing, Corrupted, Dataloss)
#7643
opened May 30, 2024 by
cabfile
Bug: server crashes on startup is ckt ctv specified.
bug
Something isn't working
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#7639
opened May 30, 2024 by
0wwafa
Bug: server crashed today for the first time.
bug-unconfirmed
high severity
Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
#7637
opened May 30, 2024 by
0wwafa
Feature Request: change model and lora from server api
enhancement
New feature or request
#7635
opened May 30, 2024 by
stygmate
4 tasks done
Feature Request: Support for Yuan2-M32
enhancement
New feature or request
#7632
opened May 30, 2024 by
matteoserva
4 tasks done
Bug: SPM tokenization breaks in at least one specific case.
bug-unconfirmed
medium severity
Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
#7629
opened May 29, 2024 by
snichols
Question: How to convert Yi-34B-Chat-4bits to gguf?
question
Further information is requested
#7623
opened May 29, 2024 by
goodmaney
2 tasks done
Feature Request: codestral support
enhancement
New feature or request
#7622
opened May 29, 2024 by
bachittle
4 tasks done
Feature: support Vulkan devices that don't support 16-bit storage
enhancement
New feature or request
#7620
opened May 29, 2024 by
grego
4 tasks done
Previous Next
ProTip!
Type g p on any issue or pull request to go back to the pull request listing page.