Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add support for GLM-Edge and GLM-Edge-V series models #10573

Merged
merged 34 commits into from
Feb 2, 2025
Merged
Show file tree
Hide file tree
Changes from 1 commit
Commits
Show all changes
34 commits
Select commit Hold shift + click to select a range
677058f
add glm edge chat model
piDack Nov 8, 2024
a249dc0
Merge branch 'master' of https://github.com/piDack/llama.cpp into sup…
piDack Nov 8, 2024
4f69662
use config partial_rotary_factor as rope ratio
piDack Nov 26, 2024
6fc90cb
support for glm edge model
piDack Nov 26, 2024
ae41d3e
Merge branch 'master' of https://github.com/piDack/llama.cpp into sup…
Nov 26, 2024
3b27041
vision model support
Nov 26, 2024
55a6f95
remove debug info
Nov 29, 2024
7d80a4a
fix format
Nov 29, 2024
6c50e9c
llava.cpp trailing whitespace
Nov 29, 2024
816d93d
Merge branch 'master' of https://github.com/piDack/llama.cpp into sup…
Nov 29, 2024
6928805
remove unused AutoTokenizer
Nov 29, 2024
5ff5632
Update src/llama.cpp for not contain <|end|> or </s>
piDack Nov 30, 2024
82cbfda
Merge branch 'master' into support_glm_edge_model
piDack Dec 3, 2024
3b409c1
add edge template
piDack Dec 3, 2024
6e9fdb0
fix chat template
piDack Dec 4, 2024
bc93d2a
Merge branch 'ggerganov:master' into support_glm_edge_model
piDack Dec 4, 2024
f91cf62
Merge branch 'master' into support_glm_edge_model
piDack Dec 19, 2024
24bad77
fix confict
piDack Dec 19, 2024
86bce2b
merge
piDack Jan 26, 2025
d9db092
fix confict
piDack Jan 26, 2025
593cc86
fix ci err
piDack Jan 26, 2025
f077b03
fix format err
piDack Jan 26, 2025
9f5d809
fix template err
piDack Jan 26, 2025
1099ef2
9b hf chat support
piDack Jan 26, 2025
96bde6f
format
piDack Jan 29, 2025
a7054a1
format clip.cpp
piDack Jan 29, 2025
13961b3
fix format
piDack Jan 30, 2025
0d3ad16
Merge branch 'ggerganov:master' into support_glm_edge_model
piDack Jan 30, 2025
0536d00
Apply suggestions from code review
ngxson Jan 30, 2025
6108d4c
Apply suggestions from code review
ngxson Jan 30, 2025
5f13c24
Update examples/llava/clip.cpp
ngxson Jan 30, 2025
77b0b28
fix format
piDack Feb 1, 2025
c8ec97d
Merge branch 'ggerganov:master' into support_glm_edge_model
piDack Feb 1, 2025
31191ad
minor : style
ggerganov Feb 2, 2025
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
fix confict
  • Loading branch information
piDack committed Jan 26, 2025
commit d9db0929b55b87b9ab33c8cb8622e00d14bbafed
3 changes: 3 additions & 0 deletions src/llama-arch.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -1024,6 +1024,9 @@ static const std::map<llm_arch, std::map<llm_tensor, const char *>> LLM_TENSOR_N
{ LLM_TENSOR_OUTPUT, "output" },
{ LLM_TENSOR_ATTN_NORM, "blk.%d.attn_norm" },
{ LLM_TENSOR_ATTN_QKV, "blk.%d.attn_qkv" },
{ LLM_TENSOR_ATTN_Q, "blk.%d.attn_q" },
{ LLM_TENSOR_ATTN_K, "blk.%d.attn_k" },
{ LLM_TENSOR_ATTN_V, "blk.%d.attn_v" },
{ LLM_TENSOR_ATTN_OUT, "blk.%d.attn_output" },
{ LLM_TENSOR_FFN_NORM, "blk.%d.ffn_norm" },
{ LLM_TENSOR_FFN_UP, "blk.%d.ffn_up" },
Expand Down
3 changes: 3 additions & 0 deletions src/llama-model.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -3068,6 +3068,7 @@ bool llama_model::load_tensors(llama_model_loader & ml) {
} break;
case LLM_ARCH_CHATGLM:
{
printf("Loading ChatGLM model...\n");
tok_embd = create_tensor(tn(LLM_TENSOR_TOKEN_EMBD, "weight"), {n_embd, n_vocab}, 0);

// output
Expand All @@ -3076,12 +3077,14 @@ bool llama_model::load_tensors(llama_model_loader & ml) {

for (int i = 0; i < n_layer; ++i) {
auto & layer = layers[i];
printf("Loading layer %d...\n",i);

layer.attn_norm = create_tensor(tn(LLM_TENSOR_ATTN_NORM, "weight", i), {n_embd}, 0);
layer.wqkv = create_tensor(tn(LLM_TENSOR_ATTN_QKV, "weight", i), {n_embd, n_embd + 2*n_embd_gqa}, llama_model_loader::TENSOR_NOT_REQUIRED);
layer.bqkv = create_tensor(tn(LLM_TENSOR_ATTN_QKV, "bias", i), {n_embd + 2*n_embd_gqa}, llama_model_loader::TENSOR_NOT_REQUIRED);

if(layer.wqkv == nullptr){
printf("Loading W q k v %d...\n",i);
layer.wq = create_tensor(tn(LLM_TENSOR_ATTN_Q, "weight", i), {n_embd, n_embd_head_k * n_head}, 0);
layer.wk = create_tensor(tn(LLM_TENSOR_ATTN_K, "weight", i), {n_embd, n_embd_k_gqa}, 0);
layer.wv = create_tensor(tn(LLM_TENSOR_ATTN_V, "weight", i), {n_embd, n_embd_v_gqa}, 0);
Expand Down
Loading