@reverbia/sdk 1.0.0-next.20251125084053 → 1.0.0-next.20251125212314

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -861,7 +861,13 @@ var client = createClient(createClientConfig(createConfig()));
861
861
 
862
862
  // src/react/useChat.ts
863
863
  function useChat(options) {
864
- const { getToken, onData: globalOnData, onFinish, onError } = options || {};
864
+ const {
865
+ getToken,
866
+ baseUrl,
867
+ onData: globalOnData,
868
+ onFinish,
869
+ onError
870
+ } = options || {};
865
871
  const [isLoading, setIsLoading] = (0, import_react.useState)(false);
866
872
  const abortControllerRef = (0, import_react.useRef)(null);
867
873
  const stop = (0, import_react.useCallback)(() => {
@@ -1002,7 +1008,7 @@ function useChat(options) {
1002
1008
  }
1003
1009
  }
1004
1010
  },
1005
- [getToken, globalOnData, onFinish, onError]
1011
+ [getToken, baseUrl, globalOnData, onFinish, onError]
1006
1012
  );
1007
1013
  return {
1008
1014
  isLoading,
@@ -1401,7 +1407,11 @@ var postApiV1Embeddings = (options) => {
1401
1407
 
1402
1408
  // src/lib/memory/embeddings.ts
1403
1409
  var generateEmbeddingForText = async (text, options = {}) => {
1404
- const { model = "openai/text-embedding-3-small", getToken } = options;
1410
+ const {
1411
+ model = "openai/text-embedding-3-small",
1412
+ getToken,
1413
+ baseUrl
1414
+ } = options;
1405
1415
  try {
1406
1416
  const token = getToken ? await getToken() : null;
1407
1417
  const headers = {};
@@ -1409,6 +1419,7 @@ var generateEmbeddingForText = async (text, options = {}) => {
1409
1419
  headers.Authorization = `Bearer ${token}`;
1410
1420
  }
1411
1421
  const response = await postApiV1Embeddings({
1422
+ baseUrl,
1412
1423
  body: {
1413
1424
  input: text,
1414
1425
  model
@@ -1441,14 +1452,19 @@ var generateEmbeddingForMemory = async (memory, options = {}) => {
1441
1452
  return generateEmbeddingForText(text, options);
1442
1453
  };
1443
1454
  var generateEmbeddingsForMemories = async (memories, options = {}) => {
1444
- const { model = "openai/text-embedding-3-small", getToken } = options;
1455
+ const {
1456
+ model = "openai/text-embedding-3-small",
1457
+ getToken,
1458
+ baseUrl
1459
+ } = options;
1445
1460
  const embeddings = /* @__PURE__ */ new Map();
1446
1461
  for (const memory of memories) {
1447
1462
  const uniqueKey = `${memory.namespace}:${memory.key}:${memory.value}`;
1448
1463
  try {
1449
1464
  const embedding = await generateEmbeddingForMemory(memory, {
1450
1465
  model,
1451
- getToken
1466
+ getToken,
1467
+ baseUrl
1452
1468
  });
1453
1469
  embeddings.set(uniqueKey, embedding);
1454
1470
  } catch (error) {
@@ -1501,7 +1517,8 @@ function useMemory(options = {}) {
1501
1517
  embeddingModel = "openai/text-embedding-3-small",
1502
1518
  generateEmbeddings = true,
1503
1519
  onFactsExtracted,
1504
- getToken
1520
+ getToken,
1521
+ baseUrl
1505
1522
  } = options;
1506
1523
  const extractionInProgressRef = (0, import_react3.useRef)(false);
1507
1524
  const extractMemoriesFromMessage = (0, import_react3.useCallback)(
@@ -1518,6 +1535,7 @@ function useMemory(options = {}) {
1518
1535
  return null;
1519
1536
  }
1520
1537
  const completion = await (0, import_client6.postApiV1ChatCompletions)({
1538
+ baseUrl,
1521
1539
  body: {
1522
1540
  messages: [
1523
1541
  {
@@ -1641,7 +1659,8 @@ function useMemory(options = {}) {
1641
1659
  try {
1642
1660
  await generateAndStoreEmbeddings(result.items, {
1643
1661
  model: embeddingModel,
1644
- getToken: getToken || void 0
1662
+ getToken: getToken || void 0,
1663
+ baseUrl
1645
1664
  });
1646
1665
  console.log(
1647
1666
  `Generated embeddings for ${result.items.length} memories`
@@ -1670,7 +1689,8 @@ function useMemory(options = {}) {
1670
1689
  embeddingModel,
1671
1690
  generateEmbeddings,
1672
1691
  getToken,
1673
- onFactsExtracted
1692
+ onFactsExtracted,
1693
+ baseUrl
1674
1694
  ]
1675
1695
  );
1676
1696
  const searchMemories = (0, import_react3.useCallback)(
@@ -1685,7 +1705,8 @@ function useMemory(options = {}) {
1685
1705
  console.log(`[Memory Search] Searching for: "${query}"`);
1686
1706
  const queryEmbedding = await generateQueryEmbedding(query, {
1687
1707
  model: embeddingModel,
1688
- getToken
1708
+ getToken,
1709
+ baseUrl
1689
1710
  });
1690
1711
  console.log(
1691
1712
  `[Memory Search] Generated query embedding (${queryEmbedding.length} dimensions)`
@@ -1710,7 +1731,7 @@ function useMemory(options = {}) {
1710
1731
  return [];
1711
1732
  }
1712
1733
  },
1713
- [embeddingModel, getToken]
1734
+ [embeddingModel, getToken, baseUrl]
1714
1735
  );
1715
1736
  return {
1716
1737
  extractMemoriesFromMessage,
@@ -98,6 +98,7 @@ type SendMessageResult = {
98
98
  };
99
99
  type UseChatOptions = {
100
100
  getToken?: () => Promise<string | null>;
101
+ baseUrl?: string;
101
102
  /**
102
103
  * Callback function to be called when a new data chunk is received.
103
104
  */
@@ -141,6 +142,7 @@ type UseChatResult = {
141
142
  * @param options.getToken - An async function that returns an authentication token.
142
143
  * This token will be used as a Bearer token in the Authorization header.
143
144
  * If not provided, `sendMessage` will return an error.
145
+ * @param options.baseUrl - Optional base URL for the API requests.
144
146
  * @param options.onData - Callback function to be called when a new data chunk is received.
145
147
  * @param options.onFinish - Callback function to be called when the chat completion finishes successfully.
146
148
  * @param options.onError - Callback function to be called when an unexpected error
@@ -258,6 +260,10 @@ type UseMemoryOptions = {
258
260
  * Custom function to get auth token for API calls
259
261
  */
260
262
  getToken?: () => Promise<string | null>;
263
+ /**
264
+ * Optional base URL for the API requests.
265
+ */
266
+ baseUrl?: string;
261
267
  };
262
268
  type UseMemoryResult = {
263
269
  extractMemoriesFromMessage: (options: {
@@ -98,6 +98,7 @@ type SendMessageResult = {
98
98
  };
99
99
  type UseChatOptions = {
100
100
  getToken?: () => Promise<string | null>;
101
+ baseUrl?: string;
101
102
  /**
102
103
  * Callback function to be called when a new data chunk is received.
103
104
  */
@@ -141,6 +142,7 @@ type UseChatResult = {
141
142
  * @param options.getToken - An async function that returns an authentication token.
142
143
  * This token will be used as a Bearer token in the Authorization header.
143
144
  * If not provided, `sendMessage` will return an error.
145
+ * @param options.baseUrl - Optional base URL for the API requests.
144
146
  * @param options.onData - Callback function to be called when a new data chunk is received.
145
147
  * @param options.onFinish - Callback function to be called when the chat completion finishes successfully.
146
148
  * @param options.onError - Callback function to be called when an unexpected error
@@ -258,6 +260,10 @@ type UseMemoryOptions = {
258
260
  * Custom function to get auth token for API calls
259
261
  */
260
262
  getToken?: () => Promise<string | null>;
263
+ /**
264
+ * Optional base URL for the API requests.
265
+ */
266
+ baseUrl?: string;
261
267
  };
262
268
  type UseMemoryResult = {
263
269
  extractMemoriesFromMessage: (options: {
@@ -817,7 +817,13 @@ var client = createClient(createClientConfig(createConfig()));
817
817
 
818
818
  // src/react/useChat.ts
819
819
  function useChat(options) {
820
- const { getToken, onData: globalOnData, onFinish, onError } = options || {};
820
+ const {
821
+ getToken,
822
+ baseUrl,
823
+ onData: globalOnData,
824
+ onFinish,
825
+ onError
826
+ } = options || {};
821
827
  const [isLoading, setIsLoading] = useState(false);
822
828
  const abortControllerRef = useRef(null);
823
829
  const stop = useCallback(() => {
@@ -958,7 +964,7 @@ function useChat(options) {
958
964
  }
959
965
  }
960
966
  },
961
- [getToken, globalOnData, onFinish, onError]
967
+ [getToken, baseUrl, globalOnData, onFinish, onError]
962
968
  );
963
969
  return {
964
970
  isLoading,
@@ -1357,7 +1363,11 @@ var postApiV1Embeddings = (options) => {
1357
1363
 
1358
1364
  // src/lib/memory/embeddings.ts
1359
1365
  var generateEmbeddingForText = async (text, options = {}) => {
1360
- const { model = "openai/text-embedding-3-small", getToken } = options;
1366
+ const {
1367
+ model = "openai/text-embedding-3-small",
1368
+ getToken,
1369
+ baseUrl
1370
+ } = options;
1361
1371
  try {
1362
1372
  const token = getToken ? await getToken() : null;
1363
1373
  const headers = {};
@@ -1365,6 +1375,7 @@ var generateEmbeddingForText = async (text, options = {}) => {
1365
1375
  headers.Authorization = `Bearer ${token}`;
1366
1376
  }
1367
1377
  const response = await postApiV1Embeddings({
1378
+ baseUrl,
1368
1379
  body: {
1369
1380
  input: text,
1370
1381
  model
@@ -1397,14 +1408,19 @@ var generateEmbeddingForMemory = async (memory, options = {}) => {
1397
1408
  return generateEmbeddingForText(text, options);
1398
1409
  };
1399
1410
  var generateEmbeddingsForMemories = async (memories, options = {}) => {
1400
- const { model = "openai/text-embedding-3-small", getToken } = options;
1411
+ const {
1412
+ model = "openai/text-embedding-3-small",
1413
+ getToken,
1414
+ baseUrl
1415
+ } = options;
1401
1416
  const embeddings = /* @__PURE__ */ new Map();
1402
1417
  for (const memory of memories) {
1403
1418
  const uniqueKey = `${memory.namespace}:${memory.key}:${memory.value}`;
1404
1419
  try {
1405
1420
  const embedding = await generateEmbeddingForMemory(memory, {
1406
1421
  model,
1407
- getToken
1422
+ getToken,
1423
+ baseUrl
1408
1424
  });
1409
1425
  embeddings.set(uniqueKey, embedding);
1410
1426
  } catch (error) {
@@ -1457,7 +1473,8 @@ function useMemory(options = {}) {
1457
1473
  embeddingModel = "openai/text-embedding-3-small",
1458
1474
  generateEmbeddings = true,
1459
1475
  onFactsExtracted,
1460
- getToken
1476
+ getToken,
1477
+ baseUrl
1461
1478
  } = options;
1462
1479
  const extractionInProgressRef = useRef3(false);
1463
1480
  const extractMemoriesFromMessage = useCallback2(
@@ -1474,6 +1491,7 @@ function useMemory(options = {}) {
1474
1491
  return null;
1475
1492
  }
1476
1493
  const completion = await postApiV1ChatCompletions({
1494
+ baseUrl,
1477
1495
  body: {
1478
1496
  messages: [
1479
1497
  {
@@ -1597,7 +1615,8 @@ function useMemory(options = {}) {
1597
1615
  try {
1598
1616
  await generateAndStoreEmbeddings(result.items, {
1599
1617
  model: embeddingModel,
1600
- getToken: getToken || void 0
1618
+ getToken: getToken || void 0,
1619
+ baseUrl
1601
1620
  });
1602
1621
  console.log(
1603
1622
  `Generated embeddings for ${result.items.length} memories`
@@ -1626,7 +1645,8 @@ function useMemory(options = {}) {
1626
1645
  embeddingModel,
1627
1646
  generateEmbeddings,
1628
1647
  getToken,
1629
- onFactsExtracted
1648
+ onFactsExtracted,
1649
+ baseUrl
1630
1650
  ]
1631
1651
  );
1632
1652
  const searchMemories = useCallback2(
@@ -1641,7 +1661,8 @@ function useMemory(options = {}) {
1641
1661
  console.log(`[Memory Search] Searching for: "${query}"`);
1642
1662
  const queryEmbedding = await generateQueryEmbedding(query, {
1643
1663
  model: embeddingModel,
1644
- getToken
1664
+ getToken,
1665
+ baseUrl
1645
1666
  });
1646
1667
  console.log(
1647
1668
  `[Memory Search] Generated query embedding (${queryEmbedding.length} dimensions)`
@@ -1666,7 +1687,7 @@ function useMemory(options = {}) {
1666
1687
  return [];
1667
1688
  }
1668
1689
  },
1669
- [embeddingModel, getToken]
1690
+ [embeddingModel, getToken, baseUrl]
1670
1691
  );
1671
1692
  return {
1672
1693
  extractMemoriesFromMessage,
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@reverbia/sdk",
3
- "version": "1.0.0-next.20251125084053",
3
+ "version": "1.0.0-next.20251125212314",
4
4
  "description": "",
5
5
  "main": "./dist/index.cjs",
6
6
  "module": "./dist/index.mjs",