Akarshan Biswas
|
6c02a032fa
|
SYCL: Remove misleading ggml_sycl_op_flatten function (#12387)
* SYCL: Remove misleading ggml_sycl_op_flatten function
* remove trailing whitespace
* Fix L2 norm from rebase
* remove try catch block from element_wise.cpp
* remove comment from common.hp
* ggml-sycl.cpp: Add try catch sycl::exception block in compute_forward
* norm.cpp: remove try catch exception block
|
2025-03-31 11:25:24 +02:00 |
|
Molly Sophia
|
7dfad387e3
|
llama: Add support for RWKV v7 architecture (#12412)
* ggml: Add op l2_norm
Signed-off-by: Molly Sophia <mollysophia379@gmail.com>
* ggml: Add op rwkv_wkv7
Signed-off-by: Molly Sophia <mollysophia379@gmail.com>
* llama: Add support for RWKV7 and ARWKV7 models
Signed-off-by: Molly Sophia <mollysophia379@gmail.com>
* llama: fix inference with RWKV6Qwen2
Signed-off-by: Molly Sophia <mollysophia379@gmail.com>
* llama: add more (a)rwkv7 variants in size
Signed-off-by: Molly Sophia <mollysophia379@gmail.com>
* Apply code-format changes
Signed-off-by: Molly Sophia <mollysophia379@gmail.com>
* fix MUSA build
Signed-off-by: Molly Sophia <mollysophia379@gmail.com>
* llama: fix shape error with rwkv using llama-parallel
Signed-off-by: Molly Sophia <mollysophia379@gmail.com>
---------
Signed-off-by: Molly Sophia <mollysophia379@gmail.com>
|
2025-03-18 07:27:50 +08:00 |
|
luoyu-intel
|
d08c20edde
|
[SYCL] Fix the sub group size of Intel (#8106)
* use warp_size macro for all sycl kernels
* fix mask of permute_sub_group_by_xor
* fix rms_norm with correct warp number
* fix rms_norm_f32/group_norm_f32
* move norm to norm.cpp file
* fix quantize bug
* fix mmvq's batch size
|
2024-07-02 10:16:00 +08:00 |
|