cui-llama.rn 1.1.0 → 1.1.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/cpp/llama.h CHANGED
@@ -512,6 +512,9 @@ extern "C" {
512
512
  // to the decoder to start generating output sequence. For other models, it returns -1.
513
513
  LLAMA_API llama_token llama_model_decoder_start_token(const struct llama_model * model);
514
514
 
515
+ // Returns true if the model is recurrent (like Mamba, RWKV, etc.)
516
+ LLAMA_API bool llama_model_is_recurrent(const struct llama_model * model);
517
+
515
518
  // Returns 0 on success
516
519
  LLAMA_API uint32_t llama_model_quantize(
517
520
  const char * fname_inp,
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "cui-llama.rn",
3
- "version": "1.1.0",
3
+ "version": "1.1.2",
4
4
  "description": "Fork of llama.rn for ChatterUI",
5
5
  "main": "lib/commonjs/index",
6
6
  "module": "lib/module/index",