@@ -1028,7 +1028,7 @@ struct vk_op_sum_rows_push_constants
10281028 uint32_t ne0_1mp, ne0_1L;
10291029};
10301030
1031- vk_op_sum_rows_push_constants vk_op_sum_rows_push_constants_init(const ggml_tensor * src, const ggml_tensor * dst, int64_t n_cols) {
1031+ static vk_op_sum_rows_push_constants vk_op_sum_rows_push_constants_init(const ggml_tensor * src, const ggml_tensor * dst, int64_t n_cols) {
10321032 uint32_t type_size = (uint32_t)ggml_type_size(src->type);
10331033 vk_op_sum_rows_push_constants p = {};
10341034 p.n_cols = (uint32_t)n_cols;
@@ -12193,7 +12193,7 @@ static void ggml_vk_check_results_0(ggml_backend_vk_context * ctx, ggml_cgraph *
1219312193 } else if (tensor->op == GGML_OP_CONCAT) {
1219412194 tensor_clone = ggml_concat(ggml_ctx, src_clone[0], src_clone[1], *(int *)tensor->op_params);
1219512195 } else if (tensor->op == GGML_OP_UPSCALE) {
12196- tensor_clone = ggml_upscale_ext (ggml_ctx, src_clone[0], tensor->ne[0], tensor->ne[1], tensor->ne[2], tensor->ne[3], (ggml_scale_mode) tensor->op_params[0]);
12196+ tensor_clone = ggml_interpolate (ggml_ctx, src_clone[0], tensor->ne[0], tensor->ne[1], tensor->ne[2], tensor->ne[3], (ggml_scale_mode) tensor->op_params[0]);
1219712197 } else if (tensor->op == GGML_OP_SCALE) {
1219812198 const float * params = (const float *)tensor->op_params;
1219912199 tensor_clone = ggml_scale_bias(ggml_ctx, src_clone[0], params[0], params[1]);
@@ -12432,11 +12432,9 @@ static void ggml_vk_check_results_1(ggml_backend_vk_context * ctx, ggml_cgraph *
1243212432 if (tensor->op == GGML_OP_TRANSPOSE || tensor->op == GGML_OP_SET_ROWS) {
1243312433 return;
1243412434 }
12435- bool fused_rms_norm_mul = false;
1243612435 if (ctx->num_additional_fused_ops == 1 &&
1243712436 tensor->op == GGML_OP_RMS_NORM &&
1243812437 cgraph->nodes[tensor_idx + 1]->op == GGML_OP_MUL) {
12439- fused_rms_norm_mul = true;
1244012438 tensor = cgraph->nodes[tensor_idx + 1];
1244112439 }
1244212440
0 commit comments