Skip to content

Commit 3c36213

Browse files
authored
llama : remove redundant GQA check (#4796)
1 parent 72d8407 commit 3c36213

File tree

1 file changed

+0
-8
lines changed

1 file changed

+0
-8
lines changed

llama.cpp

Lines changed: 0 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -4776,7 +4776,6 @@ struct llm_build_context {
47764776
const int64_t n_embd_head = hparams.n_embd_head_v;
47774777
const int64_t n_embd_gqa = hparams.n_embd_v_gqa();
47784778
GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
4779-
GGML_ASSERT(n_embd_gqa == n_embd);
47804779

47814780
struct ggml_tensor * cur;
47824781
struct ggml_tensor * inpL;
@@ -4900,7 +4899,6 @@ struct llm_build_context {
49004899
const int64_t n_embd_head = hparams.n_embd_head_v;
49014900
const int64_t n_embd_gqa = hparams.n_embd_v_gqa();
49024901
GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
4903-
GGML_ASSERT(n_embd_gqa == n_embd);
49044902

49054903
struct ggml_tensor * cur;
49064904
struct ggml_tensor * pos;
@@ -5001,7 +4999,6 @@ struct llm_build_context {
50014999
const int64_t n_embd_head = hparams.n_embd_head_v;
50025000
const int64_t n_embd_gqa = hparams.n_embd_v_gqa();
50035001
GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
5004-
GGML_ASSERT(n_embd_gqa == n_embd);
50055002

50065003
const int64_t n_rot = n_embd_head_k / 2;
50075004

@@ -5215,7 +5212,6 @@ struct llm_build_context {
52155212
const int64_t n_embd_head = hparams.n_embd_head_v;
52165213
const int64_t n_embd_gqa = hparams.n_embd_v_gqa();
52175214
GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
5218-
GGML_ASSERT(n_embd_gqa == n_embd);
52195215

52205216
struct ggml_tensor * cur;
52215217
struct ggml_tensor * inpL;
@@ -5308,7 +5304,6 @@ struct llm_build_context {
53085304
const int64_t n_embd_head = hparams.n_embd_head_v;
53095305
const int64_t n_embd_gqa = hparams.n_embd_v_gqa();
53105306
GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
5311-
GGML_ASSERT(n_embd_gqa == n_embd);
53125307

53135308
struct ggml_tensor * cur;
53145309
struct ggml_tensor * inpL;
@@ -5404,7 +5399,6 @@ struct llm_build_context {
54045399
const int64_t n_embd_head = hparams.n_embd_head_v;
54055400
const int64_t n_embd_gqa = hparams.n_embd_v_gqa();
54065401
GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
5407-
GGML_ASSERT(n_embd_gqa == n_embd);
54085402

54095403
struct ggml_tensor * cur;
54105404
struct ggml_tensor * inpL;
@@ -5731,7 +5725,6 @@ struct llm_build_context {
57315725
const int64_t n_embd_head = hparams.n_embd_head_v;
57325726
const int64_t n_embd_gqa = hparams.n_embd_v_gqa();
57335727
GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
5734-
GGML_ASSERT(n_embd_gqa == n_embd);
57355728

57365729
struct ggml_tensor * cur;
57375730
struct ggml_tensor * attn_norm_output;
@@ -5955,7 +5948,6 @@ struct llm_build_context {
59555948
const int64_t n_embd_head = hparams.n_embd_head_v;
59565949
const int64_t n_embd_gqa = hparams.n_embd_v_gqa();
59575950
GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
5958-
GGML_ASSERT(n_embd_gqa == n_embd);
59595951

59605952
struct ggml_tensor * cur;
59615953
struct ggml_tensor * pos;

0 commit comments

Comments
 (0)