@firebase/ai 2.2.1-canary.9b8ab02c5 → 2.2.1-canary.c1237662e

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -16,13 +16,13 @@
16
16
  */
17
17
  import { CountTokensRequest, GenerateContentRequest } from './requests';
18
18
  /**
19
- * <b>(EXPERIMENTAL)</b> Defines an inference "backend" that uses Chrome's on-device model,
19
+ * Defines an inference "backend" that uses Chrome's on-device model,
20
20
  * and encapsulates logic for detecting when on-device inference is
21
21
  * possible.
22
22
  *
23
23
  * These methods should not be called directly by the user.
24
24
  *
25
- * @public
25
+ * @beta
26
26
  */
27
27
  export interface ChromeAdapter {
28
28
  /**
@@ -317,7 +317,6 @@ export declare const ResponseModality: {
317
317
  */
318
318
  export type ResponseModality = (typeof ResponseModality)[keyof typeof ResponseModality];
319
319
  /**
320
- * <b>(EXPERIMENTAL)</b>
321
320
  * Determines whether inference happens on-device or in-cloud.
322
321
  *
323
322
  * @remarks
@@ -336,7 +335,7 @@ export type ResponseModality = (typeof ResponseModality)[keyof typeof ResponseMo
336
335
  * cloud-hosted model. If not available, the SDK will fall back to an
337
336
  * on-device model.
338
337
  *
339
- * @public
338
+ * @beta
340
339
  */
341
340
  export declare const InferenceMode: {
342
341
  readonly PREFER_ON_DEVICE: "prefer_on_device";
@@ -345,10 +344,9 @@ export declare const InferenceMode: {
345
344
  readonly PREFER_IN_CLOUD: "prefer_in_cloud";
346
345
  };
347
346
  /**
348
- * <b>(EXPERIMENTAL)</b>
349
347
  * Determines whether inference happens on-device or in-cloud.
350
348
  *
351
- * @public
349
+ * @beta
352
350
  */
353
351
  export type InferenceMode = (typeof InferenceMode)[keyof typeof InferenceMode];
354
352
  /**
@@ -39,9 +39,8 @@ export declare enum Availability {
39
39
  'AVAILABLE' = "available"
40
40
  }
41
41
  /**
42
- * <b>(EXPERIMENTAL)</b>
43
42
  * Configures the creation of an on-device language model session.
44
- * @public
43
+ * @beta
45
44
  */
46
45
  export interface LanguageModelCreateCoreOptions {
47
46
  topK?: number;
@@ -49,69 +48,60 @@ export interface LanguageModelCreateCoreOptions {
49
48
  expectedInputs?: LanguageModelExpected[];
50
49
  }
51
50
  /**
52
- * <b>(EXPERIMENTAL)</b>
53
51
  * Configures the creation of an on-device language model session.
54
- * @public
52
+ * @beta
55
53
  */
56
54
  export interface LanguageModelCreateOptions extends LanguageModelCreateCoreOptions {
57
55
  signal?: AbortSignal;
58
56
  initialPrompts?: LanguageModelMessage[];
59
57
  }
60
58
  /**
61
- * <b>(EXPERIMENTAL)</b>
62
59
  * Options for an on-device language model prompt.
63
- * @public
60
+ * @beta
64
61
  */
65
62
  export interface LanguageModelPromptOptions {
66
63
  responseConstraint?: object;
67
64
  }
68
65
  /**
69
- * <b>(EXPERIMENTAL)</b>
70
66
  * Options for the expected inputs for an on-device language model.
71
- * @public
67
+ * @beta
72
68
  */ export interface LanguageModelExpected {
73
69
  type: LanguageModelMessageType;
74
70
  languages?: string[];
75
71
  }
76
72
  /**
77
- * <b>(EXPERIMENTAL)</b>
78
73
  * An on-device language model prompt.
79
- * @public
74
+ * @beta
80
75
  */
81
76
  export type LanguageModelPrompt = LanguageModelMessage[];
82
77
  /**
83
- * <b>(EXPERIMENTAL)</b>
84
78
  * An on-device language model message.
85
- * @public
79
+ * @beta
86
80
  */
87
81
  export interface LanguageModelMessage {
88
82
  role: LanguageModelMessageRole;
89
83
  content: LanguageModelMessageContent[];
90
84
  }
91
85
  /**
92
- * <b>(EXPERIMENTAL)</b>
93
86
  * An on-device language model content object.
94
- * @public
87
+ * @beta
95
88
  */
96
89
  export interface LanguageModelMessageContent {
97
90
  type: LanguageModelMessageType;
98
91
  value: LanguageModelMessageContentValue;
99
92
  }
100
93
  /**
101
- * <b>(EXPERIMENTAL)</b>
102
94
  * Allowable roles for on-device language model usage.
103
- * @public
95
+ * @beta
104
96
  */
105
97
  export type LanguageModelMessageRole = 'system' | 'user' | 'assistant';
106
98
  /**
107
- * <b>(EXPERIMENTAL)</b>
108
99
  * Allowable types for on-device language model messages.
109
- * @public
100
+ * @beta
110
101
  */
111
102
  export type LanguageModelMessageType = 'text' | 'image' | 'audio';
112
103
  /**
113
- * <b>(EXPERIMENTAL)</b>
114
104
  * Content formats that can be provided as on-device message content.
115
- * @public
105
+ * @beta
116
106
  */
117
107
  export type LanguageModelMessageContentValue = ImageBitmapSource | AudioBuffer | BufferSource | string;
@@ -325,19 +325,17 @@ export interface FunctionCallingConfig {
325
325
  allowedFunctionNames?: string[];
326
326
  }
327
327
  /**
328
- * <b>(EXPERIMENTAL)</b>
329
328
  * Encapsulates configuration for on-device inference.
330
329
  *
331
- * @public
330
+ * @beta
332
331
  */
333
332
  export interface OnDeviceParams {
334
333
  createOptions?: LanguageModelCreateOptions;
335
334
  promptOptions?: LanguageModelPromptOptions;
336
335
  }
337
336
  /**
338
- * <b>(EXPERIMENTAL)</b>
339
337
  * Configures hybrid inference.
340
- * @public
338
+ * @beta
341
339
  */
342
340
  export interface HybridParams {
343
341
  /**
package/dist/index.cjs.js CHANGED
@@ -8,7 +8,7 @@ var util = require('@firebase/util');
8
8
  var logger$1 = require('@firebase/logger');
9
9
 
10
10
  var name = "@firebase/ai";
11
- var version = "2.2.1-canary.9b8ab02c5";
11
+ var version = "2.2.1-canary.c1237662e";
12
12
 
13
13
  /**
14
14
  * @license
@@ -357,7 +357,6 @@ const ResponseModality = {
357
357
  AUDIO: 'AUDIO'
358
358
  };
359
359
  /**
360
- * <b>(EXPERIMENTAL)</b>
361
360
  * Determines whether inference happens on-device or in-cloud.
362
361
  *
363
362
  * @remarks
@@ -376,7 +375,7 @@ const ResponseModality = {
376
375
  * cloud-hosted model. If not available, the SDK will fall back to an
377
376
  * on-device model.
378
377
  *
379
- * @public
378
+ * @beta
380
379
  */
381
380
  const InferenceMode = {
382
381
  'PREFER_ON_DEVICE': 'prefer_on_device',