cui-llama.rn 1.0.1 → 1.0.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/cpp/rn-llama.hpp CHANGED
@@ -253,9 +253,17 @@ struct llama_rn_context
253
253
 
254
254
  void loadPrompt()
255
255
  {
256
- std::vector<llama_token> prompt_tokens = ::llama_tokenize(ctx, params.prompt, true);
256
+ std::vector<llama_token> prompt_tokens = ::llama_tokenize(ctx, params.prompt, true, true);
257
257
  num_prompt_tokens = prompt_tokens.size();
258
258
 
259
+ // LOG tokens
260
+ std::stringstream ss;
261
+ ss << "\n" << __func__ << ": prompt_tokens = ";
262
+ for (auto& token : prompt_tokens) {
263
+ ss << token << " ";
264
+ }
265
+ LOG_INFO("%s\n", ss.str().c_str());
266
+
259
267
  if (params.n_keep < 0)
260
268
  {
261
269
  params.n_keep = (int)num_prompt_tokens;