-
Notifications
You must be signed in to change notification settings - Fork 10.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
examples : add configuration presets #10932
Labels
documentation
Improvements or additions to documentation
enhancement
New feature or request
examples
good first issue
Good for newcomers
help wanted
Extra attention is needed
Comments
ggerganov
added
documentation
Improvements or additions to documentation
enhancement
New feature or request
help wanted
Extra attention is needed
good first issue
Good for newcomers
examples
labels
Dec 21, 2024
Hi! I'm interested in contributing to this issue as a first-time contributor I'd like to work on implementing the chat server preset for commonly used models. |
IMO having a flag llama-server --preset qwen-fim-7b
llama-server --preset embd-bert
... Or we can even introduce positional parameters (I steal the idea from ollama): llama-server launch qwen-fim-7b |
This comment was marked as off-topic.
This comment was marked as off-topic.
danbev
added a commit
to danbev/llama.cpp
that referenced
this issue
Feb 5, 2025
This commit adds default embeddings presets for the following models: - bge-small-en-v1.5 - e5-small-v2 - gte-small These can be used with llama-embedding and llama-server. For example, with llama-embedding: ```console ./build/bin/llama-embedding --embd-gte-small-default -p "Hello, how are you?" ``` And with llama-server: ```console ./build/bin/llama-server --embd-gte-small-default ``` And the embeddings endpoint can then be called with a POST request: ```console curl --request POST \ --url http://localhost:8080/embeddings \ --header "Content-Type: application/json" \ --data '{"input": "Hello, how are you?"}' ``` I'm not sure if these are the most common embedding models but hopefully this can be a good starting point for discussion and further improvements. Refs: ggerganov#10932
danbev
added a commit
that referenced
this issue
Feb 7, 2025
* common : add default embeddings presets This commit adds default embeddings presets for the following models: - bge-small-en-v1.5 - e5-small-v2 - gte-small These can be used with llama-embedding and llama-server. For example, with llama-embedding: ```console ./build/bin/llama-embedding --embd-gte-small-default -p "Hello, how are you?" ``` And with llama-server: ```console ./build/bin/llama-server --embd-gte-small-default ``` And the embeddings endpoint can then be called with a POST request: ```console curl --request POST \ --url http://localhost:8080/embeddings \ --header "Content-Type: application/json" \ --data '{"input": "Hello, how are you?"}' ``` I'm not sure if these are the most common embedding models but hopefully this can be a good starting point for discussion and further improvements. Refs: #10932
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Labels
documentation
Improvements or additions to documentation
enhancement
New feature or request
examples
good first issue
Good for newcomers
help wanted
Extra attention is needed
Description
I was recently looking for ways to demonstrate some of the functionality of the
llama.cpp
examples and some of the commands can become very cumbersome. For example, here is what I use for thellama.vim
FIM server:It would be much cleaner if I could just run, for example:
Or if I could turn this embedding server command into something simpler:
Implementation
There is already an initial example of how we can create such configuration presets:
llama.cpp/common/arg.cpp
Lines 2208 to 2220 in 5cd85b5
This preset configures the model urls so that they would be automatically downloaded from HF when the example runs and thus simplifies the command significantly. It can additionally set various default values, such as context size, batch size, pooling type, etc.
Goal
The goal of this issue is to create such presets for various common tasks:
llama.vim
The list of configuration presets would require curation and proper documentation.
I think this is a great task for new contributors to help and to get involved in the project.
The text was updated successfully, but these errors were encountered: