Skip to content

Commit 60d0f9d

Browse files
committed
whisper : fix FF + remove it from README
1 parent fd54eff commit 60d0f9d

File tree

2 files changed

+3
-2
lines changed

2 files changed

+3
-2
lines changed

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ High-performance inference of [OpenAI's Whisper](https://github.com/openai/whisp
1313
- AVX intrinsics support for x86 architectures
1414
- VSX intrinsics support for POWER architectures
1515
- Mixed F16 / F32 precision
16-
- Low memory usage (Flash Attention + Flash Forward)
16+
- Low memory usage (Flash Attention)
1717
- Zero memory allocations at runtime
1818
- Runs on the CPU
1919
- [C-style API](https://github.com/ggerganov/whisper.cpp/blob/master/whisper.h)

whisper.cpp

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -619,6 +619,7 @@ struct whisper_context {
619619
buf_last = i;
620620
#else
621621
(void) i;
622+
(void) ctx;
622623
#endif
623624
}
624625

@@ -1631,7 +1632,7 @@ static bool whisper_encode(
16311632
wctx.use_buf(ctx0, 0);
16321633

16331634
cur = ggml_flash_ff(ctx0,
1634-
ggml_cpy(ctx0, cur, ggml_new_tensor_2d(ctx0, wctx.wtype, n_state, N)),
1635+
ggml_cpy(ctx0, cur, ggml_new_tensor_2d(ctx0, wctx.wtype, n_state, n_ctx)),
16351636
layer.mlp_0_w, layer.mlp_0_b, layer.mlp_1_w, layer.mlp_1_b);
16361637
#else
16371638
wctx.use_buf(ctx0, 0);

0 commit comments

Comments
 (0)