Xuan-Son Nguyen
1466621e73
llama : Support llama 4 text-only (#12791)
* llama4 conversion
* initial support, no chat template
* clean up a bit
* fix tokenizer conversion
* correct hparams
* try this
* fix shexp
* ffn_inp_normed
* chat template
* clean up model conversion
* add_bos
* add scale_before_ffn
* fix order
* weight_before_ffn
* llm_graph_input_attn_temp
* add chunk attn mask
* build_inp_attn_scale()
* add comment about ggml_repeat
* clarify comments
* fix build
2025-04-07 23:06:44 +02:00
..
2025-02-26 08:04:48 -05:00
2024-07-18 20:40:15 +10:00
2025-04-07 23:06:44 +02:00
2025-02-26 08:04:48 -05:00
2025-04-07 23:06:44 +02:00
2023-11-11 08:04:50 +03:00
2024-08-08 13:33:09 -04:00
2025-02-28 17:44:46 +01:00
2023-08-30 11:25:50 +03:00
2024-09-05 21:48:47 -04:00
2025-03-30 22:21:03 +02:00
2025-02-15 16:40:57 +02:00
2025-03-19 08:58:13 +01:00