Skip to content

Commit f24378f

Browse files
author
Bodhi Hu
committed
update
1 parent 8a68656 commit f24378f

File tree

3 files changed

+4
-8
lines changed

3 files changed

+4
-8
lines changed

docs/build.md

Lines changed: 2 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -202,17 +202,14 @@ This provides GPU acceleration using the MUSA cores of your Moore Threads MTT GP
202202
- Using `CMake`:
203203

204204
```bash
205-
# build with MUSA and using the compilers from MUSA SDK:
206-
cmake -B build -DGGML_MUSA=ON \
207-
-DCMAKE_C_COMPILER=clang -DCMAKE_CXX_COMPILER=clang++
205+
cmake -B build -DGGML_MUSA=ON
208206
cmake --build build --config Release
209207
```
210208
- For static build:
211209

212210
```bash
213211
cmake -B build -DGGML_MUSA=ON \
214-
-DBUILD_SHARED_LIBS=OFF -DCMAKE_POSITION_INDEPENDENT_CODE=ON \
215-
-DCMAKE_C_COMPILER=clang -DCMAKE_CXX_COMPILER=clang++
212+
-DBUILD_SHARED_LIBS=OFF -DCMAKE_POSITION_INDEPENDENT_CODE=ON
216213
cmake --build build --config Release
217214
```
218215

@@ -222,7 +219,6 @@ The environment variable `GGML_CUDA_ENABLE_UNIFIED_MEMORY=1` can be used to enab
222219

223220
Most of the compilation options available for CUDA should also be available for MUSA, though they haven't been thoroughly tested yet.
224221

225-
226222
## HIP
227223

228224
This provides GPU acceleration on HIP-supported AMD GPUs.

ggml/src/ggml-cuda/common.cuh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -404,7 +404,7 @@ static __device__ __forceinline__ int ggml_cuda_dp4a(const int a, const int b, i
404404

405405
#if __CUDA_ARCH__ >= GGML_CUDA_CC_DP4A || defined(GGML_USE_MUSA)
406406
return __dp4a(a, b, c);
407-
#else
407+
#else // __CUDA_ARCH__ >= GGML_CUDA_CC_DP4A || defined(GGML_USE_MUSA)
408408
const int8_t * a8 = (const int8_t *) &a;
409409
const int8_t * b8 = (const int8_t *) &b;
410410
return c + a8[0]*b8[0] + a8[1]*b8[1] + a8[2]*b8[2] + a8[3]*b8[3];

src/llama-model.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3703,7 +3703,7 @@ void llama_model::print_info() const {
37033703
}
37043704

37053705
if (arch == LLM_ARCH_LLAMA) {
3706-
LLAMA_LOG_INFO("%s: expert_weights_scale = %.1f\n", __func__, hparams.expert_weights_scale);
3706+
LLAMA_LOG_INFO("%s: expert_weights_scale = %.1f\n", __func__, hparams.expert_weights_scale);
37073707
}
37083708

37093709
vocab.print_info();

0 commit comments

Comments
 (0)