1
1
#include " llama.h"
2
- #include < assert.h >
3
- #include < math.h >
2
+ #include < cassert >
3
+ #include < cmath >
4
4
#include < numeric>
5
5
#include < cassert>
6
6
#include < iostream>
@@ -20,15 +20,19 @@ void dump(const llama_token_data_array * candidates) {
20
20
void test_top_k (const std::vector<float > & probs,
21
21
const std::vector<float > & expected_probs,
22
22
int k) {
23
+ printf (" %s:%d (%s)\n " , __FILE__, __LINE__, __func__);
23
24
size_t n_vocab = probs.size ();
24
25
std::vector<llama_token_data> candidates;
25
26
candidates.reserve (n_vocab);
26
27
for (llama_token token_id = 0 ; token_id < (llama_token)n_vocab; token_id++) {
27
- float logit = log (probs[token_id]);
28
+ printf (" %s:%d (%s) token_id: %d\n " , __FILE__, __LINE__, __func__, token_id);
29
+ float logit = std::log (probs[token_id]);
28
30
candidates.emplace_back (llama_token_data{token_id, logit, 0 .0f });
29
31
}
32
+ printf (" %s:%d (%s)\n " , __FILE__, __LINE__, __func__);
30
33
31
34
llama_token_data_array candidates_p = { candidates.data (), candidates.size (), false };
35
+ printf (" %s:%d (%s)\n " , __FILE__, __LINE__, __func__);
32
36
llama_sample_softmax (nullptr , &candidates_p);
33
37
DUMP (&candidates_p);
34
38
llama_sample_top_k (nullptr , &candidates_p, k);
0 commit comments