@@ -131,6 +131,7 @@ pub fn build(b: *std.build.Builder) !void {
131
131
const sampling = make .obj ("sampling" , "common/sampling.cpp" );
132
132
const grammar_parser = make .obj ("grammar-parser" , "common/grammar-parser.cpp" );
133
133
const train = make .obj ("train" , "common/train.cpp" );
134
+ const clip = make .obj ("clip" , "examples/llava/clip.cpp" );
134
135
135
136
_ = make .exe ("main" , "examples/main/main.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common , sampling , console , grammar_parser });
136
137
_ = make .exe ("quantize" , "examples/quantize/quantize.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common });
@@ -139,7 +140,7 @@ pub fn build(b: *std.build.Builder) !void {
139
140
_ = make .exe ("finetune" , "examples/finetune/finetune.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common , train });
140
141
_ = make .exe ("train-text-from-scratch" , "examples/train-text-from-scratch/train-text-from-scratch.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common , train });
141
142
142
- const server = make .exe ("server" , "examples/server/server.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common , sampling , grammar_parser });
143
+ const server = make .exe ("server" , "examples/server/server.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common , sampling , grammar_parser , clip });
143
144
if (server .target .isWindows ()) {
144
145
server .linkSystemLibrary ("ws2_32" );
145
146
}
0 commit comments