@elizaos/plugin-local-ai 1.0.0-beta.7 → 1.0.0-beta.8

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/LICENSE CHANGED
@@ -1,6 +1,6 @@
1
1
  MIT License
2
2
 
3
- Copyright (c) 2025 Shaw Walters, aka Moon aka @lalalune
3
+ Copyright (c) 2025 Shaw Walters and elizaOS Contributors
4
4
 
5
5
  Permission is hereby granted, free of charge, to any person obtaining a copy
6
6
  of this software and associated documentation files (the "Software"), to deal
package/README.md CHANGED
@@ -20,13 +20,13 @@ The plugin requires these environment variables (can be set in .env file or char
20
20
  "USE_STUDIOLM_TEXT_MODELS": false,
21
21
  "USE_OLLAMA_TEXT_MODELS": false,
22
22
 
23
- "OLLAMA_SERVER_URL": "http://localhost:11434",
23
+ "OLLAMA_API_ENDPOINT": "http://localhost:11434",
24
24
  "OLLAMA_MODEL": "deepseek-r1-distill-qwen-7b",
25
25
  "USE_OLLAMA_EMBEDDING": false,
26
26
  "OLLAMA_EMBEDDING_MODEL": "",
27
- "SMALL_OLLAMA_MODEL": "deepseek-r1:1.5b",
28
- "MEDIUM_OLLAMA_MODEL": "deepseek-r1:7b",
29
- "LARGE_OLLAMA_MODEL": "deepseek-r1:7b",
27
+ "OLLAMA_SMALL_MODEL": "deepseek-r1:1.5b",
28
+ "OLLAMA_MEDIUM_MODEL": "deepseek-r1:7b",
29
+ "OLLAMA_LARGE_MODEL": "deepseek-r1:7b",
30
30
 
31
31
  "STUDIOLM_SERVER_URL": "http://localhost:1234",
32
32
  "STUDIOLM_SMALL_MODEL": "lmstudio-community/deepseek-r1-distill-qwen-1.5b",
@@ -44,13 +44,13 @@ USE_STUDIOLM_TEXT_MODELS=false
44
44
  USE_OLLAMA_TEXT_MODELS=false
45
45
 
46
46
  # Ollama Configuration
47
- OLLAMA_SERVER_URL=http://localhost:11434
47
+ OLLAMA_API_ENDPOINT=http://localhost:11434
48
48
  OLLAMA_MODEL=deepseek-r1-distill-qwen-7b
49
49
  USE_OLLAMA_EMBEDDING=false
50
50
  OLLAMA_EMBEDDING_MODEL=
51
- SMALL_OLLAMA_MODEL=deepseek-r1:1.5b
52
- MEDIUM_OLLAMA_MODEL=deepseek-r1:7b
53
- LARGE_OLLAMA_MODEL=deepseek-r1:7b
51
+ OLLAMA_SMALL_MODEL=deepseek-r1:1.5b
52
+ OLLAMA_MEDIUM_MODEL=deepseek-r1:7b
53
+ OLLAMA_LARGE_MODEL=deepseek-r1:7b
54
54
 
55
55
  # StudioLM Configuration
56
56
  STUDIOLM_SERVER_URL=http://localhost:1234
@@ -69,13 +69,13 @@ STUDIOLM_EMBEDDING_MODEL=false
69
69
 
70
70
  #### Ollama Settings
71
71
 
72
- - `OLLAMA_SERVER_URL`: Ollama API endpoint (default: http://localhost:11434)
72
+ - `OLLAMA_API_ENDPOINT`: Ollama API endpoint (default: http://localhost:11434)
73
73
  - `OLLAMA_MODEL`: Default model for general use
74
74
  - `USE_OLLAMA_EMBEDDING`: Enable Ollama for embeddings
75
75
  - `OLLAMA_EMBEDDING_MODEL`: Model for embeddings when enabled
76
- - `SMALL_OLLAMA_MODEL`: Model for lighter tasks
77
- - `MEDIUM_OLLAMA_MODEL`: Model for standard tasks
78
- - `LARGE_OLLAMA_MODEL`: Model for complex tasks
76
+ - `OLLAMA_SMALL_MODEL`: Model for lighter tasks
77
+ - `OLLAMA_MEDIUM_MODEL`: Model for standard tasks
78
+ - `OLLAMA_LARGE_MODEL`: Model for complex tasks
79
79
 
80
80
  #### StudioLM Settings
81
81