@reverbia/sdk 1.0.0-next.20251125084053 → 1.0.0-next.20251125205812
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/react/index.cjs +31 -10
- package/dist/react/index.d.mts +6 -0
- package/dist/react/index.d.ts +6 -0
- package/dist/react/index.mjs +31 -10
- package/package.json +1 -1
package/dist/react/index.cjs
CHANGED
|
@@ -861,7 +861,13 @@ var client = createClient(createClientConfig(createConfig()));
|
|
|
861
861
|
|
|
862
862
|
// src/react/useChat.ts
|
|
863
863
|
function useChat(options) {
|
|
864
|
-
const {
|
|
864
|
+
const {
|
|
865
|
+
getToken,
|
|
866
|
+
baseUrl,
|
|
867
|
+
onData: globalOnData,
|
|
868
|
+
onFinish,
|
|
869
|
+
onError
|
|
870
|
+
} = options || {};
|
|
865
871
|
const [isLoading, setIsLoading] = (0, import_react.useState)(false);
|
|
866
872
|
const abortControllerRef = (0, import_react.useRef)(null);
|
|
867
873
|
const stop = (0, import_react.useCallback)(() => {
|
|
@@ -1002,7 +1008,7 @@ function useChat(options) {
|
|
|
1002
1008
|
}
|
|
1003
1009
|
}
|
|
1004
1010
|
},
|
|
1005
|
-
[getToken, globalOnData, onFinish, onError]
|
|
1011
|
+
[getToken, baseUrl, globalOnData, onFinish, onError]
|
|
1006
1012
|
);
|
|
1007
1013
|
return {
|
|
1008
1014
|
isLoading,
|
|
@@ -1401,7 +1407,11 @@ var postApiV1Embeddings = (options) => {
|
|
|
1401
1407
|
|
|
1402
1408
|
// src/lib/memory/embeddings.ts
|
|
1403
1409
|
var generateEmbeddingForText = async (text, options = {}) => {
|
|
1404
|
-
const {
|
|
1410
|
+
const {
|
|
1411
|
+
model = "openai/text-embedding-3-small",
|
|
1412
|
+
getToken,
|
|
1413
|
+
baseUrl
|
|
1414
|
+
} = options;
|
|
1405
1415
|
try {
|
|
1406
1416
|
const token = getToken ? await getToken() : null;
|
|
1407
1417
|
const headers = {};
|
|
@@ -1409,6 +1419,7 @@ var generateEmbeddingForText = async (text, options = {}) => {
|
|
|
1409
1419
|
headers.Authorization = `Bearer ${token}`;
|
|
1410
1420
|
}
|
|
1411
1421
|
const response = await postApiV1Embeddings({
|
|
1422
|
+
baseUrl,
|
|
1412
1423
|
body: {
|
|
1413
1424
|
input: text,
|
|
1414
1425
|
model
|
|
@@ -1441,14 +1452,19 @@ var generateEmbeddingForMemory = async (memory, options = {}) => {
|
|
|
1441
1452
|
return generateEmbeddingForText(text, options);
|
|
1442
1453
|
};
|
|
1443
1454
|
var generateEmbeddingsForMemories = async (memories, options = {}) => {
|
|
1444
|
-
const {
|
|
1455
|
+
const {
|
|
1456
|
+
model = "openai/text-embedding-3-small",
|
|
1457
|
+
getToken,
|
|
1458
|
+
baseUrl
|
|
1459
|
+
} = options;
|
|
1445
1460
|
const embeddings = /* @__PURE__ */ new Map();
|
|
1446
1461
|
for (const memory of memories) {
|
|
1447
1462
|
const uniqueKey = `${memory.namespace}:${memory.key}:${memory.value}`;
|
|
1448
1463
|
try {
|
|
1449
1464
|
const embedding = await generateEmbeddingForMemory(memory, {
|
|
1450
1465
|
model,
|
|
1451
|
-
getToken
|
|
1466
|
+
getToken,
|
|
1467
|
+
baseUrl
|
|
1452
1468
|
});
|
|
1453
1469
|
embeddings.set(uniqueKey, embedding);
|
|
1454
1470
|
} catch (error) {
|
|
@@ -1501,7 +1517,8 @@ function useMemory(options = {}) {
|
|
|
1501
1517
|
embeddingModel = "openai/text-embedding-3-small",
|
|
1502
1518
|
generateEmbeddings = true,
|
|
1503
1519
|
onFactsExtracted,
|
|
1504
|
-
getToken
|
|
1520
|
+
getToken,
|
|
1521
|
+
baseUrl
|
|
1505
1522
|
} = options;
|
|
1506
1523
|
const extractionInProgressRef = (0, import_react3.useRef)(false);
|
|
1507
1524
|
const extractMemoriesFromMessage = (0, import_react3.useCallback)(
|
|
@@ -1518,6 +1535,7 @@ function useMemory(options = {}) {
|
|
|
1518
1535
|
return null;
|
|
1519
1536
|
}
|
|
1520
1537
|
const completion = await (0, import_client6.postApiV1ChatCompletions)({
|
|
1538
|
+
baseUrl,
|
|
1521
1539
|
body: {
|
|
1522
1540
|
messages: [
|
|
1523
1541
|
{
|
|
@@ -1641,7 +1659,8 @@ function useMemory(options = {}) {
|
|
|
1641
1659
|
try {
|
|
1642
1660
|
await generateAndStoreEmbeddings(result.items, {
|
|
1643
1661
|
model: embeddingModel,
|
|
1644
|
-
getToken: getToken || void 0
|
|
1662
|
+
getToken: getToken || void 0,
|
|
1663
|
+
baseUrl
|
|
1645
1664
|
});
|
|
1646
1665
|
console.log(
|
|
1647
1666
|
`Generated embeddings for ${result.items.length} memories`
|
|
@@ -1670,7 +1689,8 @@ function useMemory(options = {}) {
|
|
|
1670
1689
|
embeddingModel,
|
|
1671
1690
|
generateEmbeddings,
|
|
1672
1691
|
getToken,
|
|
1673
|
-
onFactsExtracted
|
|
1692
|
+
onFactsExtracted,
|
|
1693
|
+
baseUrl
|
|
1674
1694
|
]
|
|
1675
1695
|
);
|
|
1676
1696
|
const searchMemories = (0, import_react3.useCallback)(
|
|
@@ -1685,7 +1705,8 @@ function useMemory(options = {}) {
|
|
|
1685
1705
|
console.log(`[Memory Search] Searching for: "${query}"`);
|
|
1686
1706
|
const queryEmbedding = await generateQueryEmbedding(query, {
|
|
1687
1707
|
model: embeddingModel,
|
|
1688
|
-
getToken
|
|
1708
|
+
getToken,
|
|
1709
|
+
baseUrl
|
|
1689
1710
|
});
|
|
1690
1711
|
console.log(
|
|
1691
1712
|
`[Memory Search] Generated query embedding (${queryEmbedding.length} dimensions)`
|
|
@@ -1710,7 +1731,7 @@ function useMemory(options = {}) {
|
|
|
1710
1731
|
return [];
|
|
1711
1732
|
}
|
|
1712
1733
|
},
|
|
1713
|
-
[embeddingModel, getToken]
|
|
1734
|
+
[embeddingModel, getToken, baseUrl]
|
|
1714
1735
|
);
|
|
1715
1736
|
return {
|
|
1716
1737
|
extractMemoriesFromMessage,
|
package/dist/react/index.d.mts
CHANGED
|
@@ -98,6 +98,7 @@ type SendMessageResult = {
|
|
|
98
98
|
};
|
|
99
99
|
type UseChatOptions = {
|
|
100
100
|
getToken?: () => Promise<string | null>;
|
|
101
|
+
baseUrl?: string;
|
|
101
102
|
/**
|
|
102
103
|
* Callback function to be called when a new data chunk is received.
|
|
103
104
|
*/
|
|
@@ -141,6 +142,7 @@ type UseChatResult = {
|
|
|
141
142
|
* @param options.getToken - An async function that returns an authentication token.
|
|
142
143
|
* This token will be used as a Bearer token in the Authorization header.
|
|
143
144
|
* If not provided, `sendMessage` will return an error.
|
|
145
|
+
* @param options.baseUrl - Optional base URL for the API requests.
|
|
144
146
|
* @param options.onData - Callback function to be called when a new data chunk is received.
|
|
145
147
|
* @param options.onFinish - Callback function to be called when the chat completion finishes successfully.
|
|
146
148
|
* @param options.onError - Callback function to be called when an unexpected error
|
|
@@ -258,6 +260,10 @@ type UseMemoryOptions = {
|
|
|
258
260
|
* Custom function to get auth token for API calls
|
|
259
261
|
*/
|
|
260
262
|
getToken?: () => Promise<string | null>;
|
|
263
|
+
/**
|
|
264
|
+
* Optional base URL for the API requests.
|
|
265
|
+
*/
|
|
266
|
+
baseUrl?: string;
|
|
261
267
|
};
|
|
262
268
|
type UseMemoryResult = {
|
|
263
269
|
extractMemoriesFromMessage: (options: {
|
package/dist/react/index.d.ts
CHANGED
|
@@ -98,6 +98,7 @@ type SendMessageResult = {
|
|
|
98
98
|
};
|
|
99
99
|
type UseChatOptions = {
|
|
100
100
|
getToken?: () => Promise<string | null>;
|
|
101
|
+
baseUrl?: string;
|
|
101
102
|
/**
|
|
102
103
|
* Callback function to be called when a new data chunk is received.
|
|
103
104
|
*/
|
|
@@ -141,6 +142,7 @@ type UseChatResult = {
|
|
|
141
142
|
* @param options.getToken - An async function that returns an authentication token.
|
|
142
143
|
* This token will be used as a Bearer token in the Authorization header.
|
|
143
144
|
* If not provided, `sendMessage` will return an error.
|
|
145
|
+
* @param options.baseUrl - Optional base URL for the API requests.
|
|
144
146
|
* @param options.onData - Callback function to be called when a new data chunk is received.
|
|
145
147
|
* @param options.onFinish - Callback function to be called when the chat completion finishes successfully.
|
|
146
148
|
* @param options.onError - Callback function to be called when an unexpected error
|
|
@@ -258,6 +260,10 @@ type UseMemoryOptions = {
|
|
|
258
260
|
* Custom function to get auth token for API calls
|
|
259
261
|
*/
|
|
260
262
|
getToken?: () => Promise<string | null>;
|
|
263
|
+
/**
|
|
264
|
+
* Optional base URL for the API requests.
|
|
265
|
+
*/
|
|
266
|
+
baseUrl?: string;
|
|
261
267
|
};
|
|
262
268
|
type UseMemoryResult = {
|
|
263
269
|
extractMemoriesFromMessage: (options: {
|
package/dist/react/index.mjs
CHANGED
|
@@ -817,7 +817,13 @@ var client = createClient(createClientConfig(createConfig()));
|
|
|
817
817
|
|
|
818
818
|
// src/react/useChat.ts
|
|
819
819
|
function useChat(options) {
|
|
820
|
-
const {
|
|
820
|
+
const {
|
|
821
|
+
getToken,
|
|
822
|
+
baseUrl,
|
|
823
|
+
onData: globalOnData,
|
|
824
|
+
onFinish,
|
|
825
|
+
onError
|
|
826
|
+
} = options || {};
|
|
821
827
|
const [isLoading, setIsLoading] = useState(false);
|
|
822
828
|
const abortControllerRef = useRef(null);
|
|
823
829
|
const stop = useCallback(() => {
|
|
@@ -958,7 +964,7 @@ function useChat(options) {
|
|
|
958
964
|
}
|
|
959
965
|
}
|
|
960
966
|
},
|
|
961
|
-
[getToken, globalOnData, onFinish, onError]
|
|
967
|
+
[getToken, baseUrl, globalOnData, onFinish, onError]
|
|
962
968
|
);
|
|
963
969
|
return {
|
|
964
970
|
isLoading,
|
|
@@ -1357,7 +1363,11 @@ var postApiV1Embeddings = (options) => {
|
|
|
1357
1363
|
|
|
1358
1364
|
// src/lib/memory/embeddings.ts
|
|
1359
1365
|
var generateEmbeddingForText = async (text, options = {}) => {
|
|
1360
|
-
const {
|
|
1366
|
+
const {
|
|
1367
|
+
model = "openai/text-embedding-3-small",
|
|
1368
|
+
getToken,
|
|
1369
|
+
baseUrl
|
|
1370
|
+
} = options;
|
|
1361
1371
|
try {
|
|
1362
1372
|
const token = getToken ? await getToken() : null;
|
|
1363
1373
|
const headers = {};
|
|
@@ -1365,6 +1375,7 @@ var generateEmbeddingForText = async (text, options = {}) => {
|
|
|
1365
1375
|
headers.Authorization = `Bearer ${token}`;
|
|
1366
1376
|
}
|
|
1367
1377
|
const response = await postApiV1Embeddings({
|
|
1378
|
+
baseUrl,
|
|
1368
1379
|
body: {
|
|
1369
1380
|
input: text,
|
|
1370
1381
|
model
|
|
@@ -1397,14 +1408,19 @@ var generateEmbeddingForMemory = async (memory, options = {}) => {
|
|
|
1397
1408
|
return generateEmbeddingForText(text, options);
|
|
1398
1409
|
};
|
|
1399
1410
|
var generateEmbeddingsForMemories = async (memories, options = {}) => {
|
|
1400
|
-
const {
|
|
1411
|
+
const {
|
|
1412
|
+
model = "openai/text-embedding-3-small",
|
|
1413
|
+
getToken,
|
|
1414
|
+
baseUrl
|
|
1415
|
+
} = options;
|
|
1401
1416
|
const embeddings = /* @__PURE__ */ new Map();
|
|
1402
1417
|
for (const memory of memories) {
|
|
1403
1418
|
const uniqueKey = `${memory.namespace}:${memory.key}:${memory.value}`;
|
|
1404
1419
|
try {
|
|
1405
1420
|
const embedding = await generateEmbeddingForMemory(memory, {
|
|
1406
1421
|
model,
|
|
1407
|
-
getToken
|
|
1422
|
+
getToken,
|
|
1423
|
+
baseUrl
|
|
1408
1424
|
});
|
|
1409
1425
|
embeddings.set(uniqueKey, embedding);
|
|
1410
1426
|
} catch (error) {
|
|
@@ -1457,7 +1473,8 @@ function useMemory(options = {}) {
|
|
|
1457
1473
|
embeddingModel = "openai/text-embedding-3-small",
|
|
1458
1474
|
generateEmbeddings = true,
|
|
1459
1475
|
onFactsExtracted,
|
|
1460
|
-
getToken
|
|
1476
|
+
getToken,
|
|
1477
|
+
baseUrl
|
|
1461
1478
|
} = options;
|
|
1462
1479
|
const extractionInProgressRef = useRef3(false);
|
|
1463
1480
|
const extractMemoriesFromMessage = useCallback2(
|
|
@@ -1474,6 +1491,7 @@ function useMemory(options = {}) {
|
|
|
1474
1491
|
return null;
|
|
1475
1492
|
}
|
|
1476
1493
|
const completion = await postApiV1ChatCompletions({
|
|
1494
|
+
baseUrl,
|
|
1477
1495
|
body: {
|
|
1478
1496
|
messages: [
|
|
1479
1497
|
{
|
|
@@ -1597,7 +1615,8 @@ function useMemory(options = {}) {
|
|
|
1597
1615
|
try {
|
|
1598
1616
|
await generateAndStoreEmbeddings(result.items, {
|
|
1599
1617
|
model: embeddingModel,
|
|
1600
|
-
getToken: getToken || void 0
|
|
1618
|
+
getToken: getToken || void 0,
|
|
1619
|
+
baseUrl
|
|
1601
1620
|
});
|
|
1602
1621
|
console.log(
|
|
1603
1622
|
`Generated embeddings for ${result.items.length} memories`
|
|
@@ -1626,7 +1645,8 @@ function useMemory(options = {}) {
|
|
|
1626
1645
|
embeddingModel,
|
|
1627
1646
|
generateEmbeddings,
|
|
1628
1647
|
getToken,
|
|
1629
|
-
onFactsExtracted
|
|
1648
|
+
onFactsExtracted,
|
|
1649
|
+
baseUrl
|
|
1630
1650
|
]
|
|
1631
1651
|
);
|
|
1632
1652
|
const searchMemories = useCallback2(
|
|
@@ -1641,7 +1661,8 @@ function useMemory(options = {}) {
|
|
|
1641
1661
|
console.log(`[Memory Search] Searching for: "${query}"`);
|
|
1642
1662
|
const queryEmbedding = await generateQueryEmbedding(query, {
|
|
1643
1663
|
model: embeddingModel,
|
|
1644
|
-
getToken
|
|
1664
|
+
getToken,
|
|
1665
|
+
baseUrl
|
|
1645
1666
|
});
|
|
1646
1667
|
console.log(
|
|
1647
1668
|
`[Memory Search] Generated query embedding (${queryEmbedding.length} dimensions)`
|
|
@@ -1666,7 +1687,7 @@ function useMemory(options = {}) {
|
|
|
1666
1687
|
return [];
|
|
1667
1688
|
}
|
|
1668
1689
|
},
|
|
1669
|
-
[embeddingModel, getToken]
|
|
1690
|
+
[embeddingModel, getToken, baseUrl]
|
|
1670
1691
|
);
|
|
1671
1692
|
return {
|
|
1672
1693
|
extractMemoriesFromMessage,
|