Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Xsn/llama batch remove compat #316

Merged

Conversation

Nexesenex
Copy link
Owner

No description provided.

ngxson and others added 19 commits October 11, 2024 11:48
* ggml : move more prints to the ggml log system

* show BLAS OpenMP warnings in all builds using debug print
…#9798)

* llama : improve infill support

ggml-ci

* llama : add more FIM token strings

ggml-ci

* server : update prompt on slot restore (ggerganov#9800)

* gguf : deprecate old FIM token KVs
* server : remove legacy system_prompt feature

ggml-ci

* readme : update [no ci]

* server : fix non-transformer logic + remove response from /props
* server : remove self-extend

ggml-ci

* server : fix context limit check to use slot.n_past

ggml-ci
@Nexesenex Nexesenex merged commit bbb1ca9 into Nexesenex:lcpp_pr_llama_batch Oct 15, 2024
6 of 7 checks passed
@github-actions github-actions bot added documentation Improvements or additions to documentation examples python server ggml android labels Oct 15, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
android documentation Improvements or additions to documentation examples ggml python server
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants