graphlit-client 1.0.20251005001 → 1.0.20251007001

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -212,6 +212,14 @@ export const GetAlert = gql `
212
212
  id
213
213
  }
214
214
  }
215
+ openAIVideo {
216
+ model
217
+ seconds
218
+ size
219
+ seed {
220
+ id
221
+ }
222
+ }
215
223
  }
216
224
  summarySpecification {
217
225
  id
@@ -382,6 +390,14 @@ export const QueryAlerts = gql `
382
390
  id
383
391
  }
384
392
  }
393
+ openAIVideo {
394
+ model
395
+ seconds
396
+ size
397
+ seed {
398
+ id
399
+ }
400
+ }
385
401
  }
386
402
  summarySpecification {
387
403
  id
@@ -11154,6 +11170,7 @@ export const CreateWorkflow = gql `
11154
11170
  }
11155
11171
  enableEmailCollections
11156
11172
  enableFolderCollections
11173
+ enableMessageCollections
11157
11174
  }
11158
11175
  indexing {
11159
11176
  jobs {
@@ -11318,6 +11335,7 @@ export const CreateWorkflow = gql `
11318
11335
  policy {
11319
11336
  type
11320
11337
  allowDuplicates
11338
+ embeddingTypes
11321
11339
  }
11322
11340
  }
11323
11341
  actions {
@@ -11408,6 +11426,7 @@ export const GetWorkflow = gql `
11408
11426
  }
11409
11427
  enableEmailCollections
11410
11428
  enableFolderCollections
11429
+ enableMessageCollections
11411
11430
  }
11412
11431
  indexing {
11413
11432
  jobs {
@@ -11572,6 +11591,7 @@ export const GetWorkflow = gql `
11572
11591
  policy {
11573
11592
  type
11574
11593
  allowDuplicates
11594
+ embeddingTypes
11575
11595
  }
11576
11596
  }
11577
11597
  actions {
@@ -11635,6 +11655,7 @@ export const QueryWorkflows = gql `
11635
11655
  }
11636
11656
  enableEmailCollections
11637
11657
  enableFolderCollections
11658
+ enableMessageCollections
11638
11659
  }
11639
11660
  indexing {
11640
11661
  jobs {
@@ -11799,6 +11820,7 @@ export const QueryWorkflows = gql `
11799
11820
  policy {
11800
11821
  type
11801
11822
  allowDuplicates
11823
+ embeddingTypes
11802
11824
  }
11803
11825
  }
11804
11826
  actions {
@@ -11857,6 +11879,7 @@ export const UpdateWorkflow = gql `
11857
11879
  }
11858
11880
  enableEmailCollections
11859
11881
  enableFolderCollections
11882
+ enableMessageCollections
11860
11883
  }
11861
11884
  indexing {
11862
11885
  jobs {
@@ -12021,6 +12044,7 @@ export const UpdateWorkflow = gql `
12021
12044
  policy {
12022
12045
  type
12023
12046
  allowDuplicates
12047
+ embeddingTypes
12024
12048
  }
12025
12049
  }
12026
12050
  actions {
@@ -12078,6 +12102,7 @@ export const UpsertWorkflow = gql `
12078
12102
  }
12079
12103
  enableEmailCollections
12080
12104
  enableFolderCollections
12105
+ enableMessageCollections
12081
12106
  }
12082
12107
  indexing {
12083
12108
  jobs {
@@ -12242,6 +12267,7 @@ export const UpsertWorkflow = gql `
12242
12267
  policy {
12243
12268
  type
12244
12269
  allowDuplicates
12270
+ embeddingTypes
12245
12271
  }
12246
12272
  }
12247
12273
  actions {
@@ -1192,12 +1192,18 @@ export type CalendarAttendee = {
1192
1192
  };
1193
1193
  /** Represents a calendar event attendee input. */
1194
1194
  export type CalendarAttendeeInput = {
1195
+ /** The attendee display name. */
1196
+ displayName?: InputMaybe<Scalars['String']['input']>;
1195
1197
  /** The attendee email. */
1196
1198
  email?: InputMaybe<Scalars['String']['input']>;
1197
1199
  /** Whether the attendee is optional. */
1198
1200
  isOptional?: InputMaybe<Scalars['Boolean']['input']>;
1199
1201
  /** Whether the attendee is the organizer. */
1200
1202
  isOrganizer?: InputMaybe<Scalars['Boolean']['input']>;
1203
+ /** Whether the attendee is required. */
1204
+ isRequired?: InputMaybe<Scalars['Boolean']['input']>;
1205
+ /** Whether the attendee is a resource. */
1206
+ isResource?: InputMaybe<Scalars['Boolean']['input']>;
1201
1207
  /** The attendee name. */
1202
1208
  name?: InputMaybe<Scalars['String']['input']>;
1203
1209
  /** The attendee response status. */
@@ -2388,6 +2394,8 @@ export type ContentPublishingConnector = {
2388
2394
  googleImage?: Maybe<GoogleImagePublishingProperties>;
2389
2395
  /** The specific properties for OpenAI Image publishing. */
2390
2396
  openAIImage?: Maybe<OpenAiImagePublishingProperties>;
2397
+ /** The specific properties for OpenAI Video publishing. */
2398
+ openAIVideo?: Maybe<OpenAiVideoPublishingProperties>;
2391
2399
  /** The content publishing service type. */
2392
2400
  type: ContentPublishingServiceTypes;
2393
2401
  };
@@ -2401,6 +2409,8 @@ export type ContentPublishingConnectorInput = {
2401
2409
  googleImage?: InputMaybe<GoogleImagePublishingPropertiesInput>;
2402
2410
  /** The specific properties for OpenAI Image publishing. */
2403
2411
  openAIImage?: InputMaybe<OpenAiImagePublishingPropertiesInput>;
2412
+ /** The specific properties for OpenAI Video publishing. */
2413
+ openAIVideo?: InputMaybe<OpenAiVideoPublishingPropertiesInput>;
2404
2414
  /** The content publishing service type. */
2405
2415
  type: ContentPublishingServiceTypes;
2406
2416
  };
@@ -2414,6 +2424,8 @@ export type ContentPublishingConnectorUpdateInput = {
2414
2424
  googleImage?: InputMaybe<GoogleImagePublishingPropertiesInput>;
2415
2425
  /** The specific properties for OpenAI Image publishing. */
2416
2426
  openAIImage?: InputMaybe<OpenAiImagePublishingPropertiesInput>;
2427
+ /** The specific properties for OpenAI Video publishing. */
2428
+ openAIVideo?: InputMaybe<OpenAiVideoPublishingPropertiesInput>;
2417
2429
  /** The content publishing service type. */
2418
2430
  type?: InputMaybe<ContentPublishingServiceTypes>;
2419
2431
  };
@@ -2426,6 +2438,8 @@ export declare enum ContentPublishingFormats {
2426
2438
  Markdown = "MARKDOWN",
2427
2439
  /** MP3 */
2428
2440
  Mp3 = "MP3",
2441
+ /** MP4 */
2442
+ Mp4 = "MP4",
2429
2443
  /** PNG */
2430
2444
  Png = "PNG",
2431
2445
  /** Plain Text */
@@ -2441,6 +2455,8 @@ export declare enum ContentPublishingServiceTypes {
2441
2455
  GoogleImage = "GOOGLE_IMAGE",
2442
2456
  /** OpenAI Image publishing */
2443
2457
  OpenAiImage = "OPEN_AI_IMAGE",
2458
+ /** OpenAI Video publishing */
2459
+ OpenAiVideo = "OPEN_AI_VIDEO",
2444
2460
  /** Text publishing */
2445
2461
  Text = "TEXT"
2446
2462
  }
@@ -3609,6 +3625,17 @@ export type EmailPreparationPropertiesInput = {
3609
3625
  /** Whether to extract attachments from emails as ingested content. */
3610
3626
  includeAttachments?: InputMaybe<Scalars['Boolean']['input']>;
3611
3627
  };
3628
+ /** Embedding type */
3629
+ export declare enum EmbeddingTypes {
3630
+ /** Audio embeddings */
3631
+ Audio = "AUDIO",
3632
+ /** Image embeddings */
3633
+ Image = "IMAGE",
3634
+ /** Text embeddings */
3635
+ Text = "TEXT",
3636
+ /** Video embeddings */
3637
+ Video = "VIDEO"
3638
+ }
3612
3639
  /** Represents the embeddings strategy. */
3613
3640
  export type EmbeddingsStrategy = {
3614
3641
  __typename?: 'EmbeddingsStrategy';
@@ -5233,7 +5260,7 @@ export type GoogleImagePublishingProperties = {
5233
5260
  count?: Maybe<Scalars['Int']['output']>;
5234
5261
  /** The Google Image model. */
5235
5262
  model?: Maybe<GoogleImageModels>;
5236
- /** The seed image reference to use when generating image(s). */
5263
+ /** The seed image reference to use when generating image(s), optional. */
5237
5264
  seed?: Maybe<EntityReference>;
5238
5265
  };
5239
5266
  /** Represents the Google Image publishing properties. */
@@ -5242,7 +5269,7 @@ export type GoogleImagePublishingPropertiesInput = {
5242
5269
  count?: InputMaybe<Scalars['Int']['input']>;
5243
5270
  /** The Google Image model. */
5244
5271
  model?: InputMaybe<GoogleImageModels>;
5245
- /** The seed image reference to use when generating image(s). */
5272
+ /** The seed image reference to use when generating image(s), optional. */
5246
5273
  seed?: InputMaybe<EntityReferenceInput>;
5247
5274
  };
5248
5275
  /** Represents Google model properties. */
@@ -5854,6 +5881,8 @@ export type IngestionWorkflowStage = {
5854
5881
  enableEmailCollections?: Maybe<Scalars['Boolean']['output']>;
5855
5882
  /** Whether to create collections for every site folder (i.e. '<parent>/<child>'). Disabled by default. */
5856
5883
  enableFolderCollections?: Maybe<Scalars['Boolean']['output']>;
5884
+ /** Whether to create collections for every message conversation (i.e. Slack, Teams, Discord). Disabled by default. */
5885
+ enableMessageCollections?: Maybe<Scalars['Boolean']['output']>;
5857
5886
  /** The ingestion filter. */
5858
5887
  if?: Maybe<IngestionContentFilter>;
5859
5888
  /** The observations to be assigned to ingested content. */
@@ -5867,6 +5896,8 @@ export type IngestionWorkflowStageInput = {
5867
5896
  enableEmailCollections?: InputMaybe<Scalars['Boolean']['input']>;
5868
5897
  /** Whether to create collections for every site folder (i.e. '<parent>/<child>'). Disabled by default. */
5869
5898
  enableFolderCollections?: InputMaybe<Scalars['Boolean']['input']>;
5899
+ /** Whether to create collections for every message conversation (i.e. Slack, Teams, Discord). Disabled by default. */
5900
+ enableMessageCollections?: InputMaybe<Scalars['Boolean']['input']>;
5870
5901
  /** The ingestion filter. */
5871
5902
  if?: InputMaybe<IngestionContentFilterInput>;
5872
5903
  /** The observations to be assigned to ingested content. */
@@ -10503,7 +10534,9 @@ export declare enum OpenAiImageModels {
10503
10534
  /** Developer-specified model */
10504
10535
  Custom = "CUSTOM",
10505
10536
  /** GPT Image-1 */
10506
- GptImage_1 = "GPT_IMAGE_1"
10537
+ GptImage_1 = "GPT_IMAGE_1",
10538
+ /** GPT Image-1 Mini */
10539
+ GptImage_1Mini = "GPT_IMAGE_1_MINI"
10507
10540
  }
10508
10541
  /** Represents the OpenAI Image publishing properties. */
10509
10542
  export type OpenAiImagePublishingProperties = {
@@ -10512,7 +10545,7 @@ export type OpenAiImagePublishingProperties = {
10512
10545
  count?: Maybe<Scalars['Int']['output']>;
10513
10546
  /** The OpenAI Image model. */
10514
10547
  model?: Maybe<OpenAiImageModels>;
10515
- /** The seed image reference to use when generating image(s). */
10548
+ /** The seed image reference to use when generating image(s), optional. */
10516
10549
  seed?: Maybe<EntityReference>;
10517
10550
  };
10518
10551
  /** Represents the OpenAI Image publishing properties. */
@@ -10521,7 +10554,7 @@ export type OpenAiImagePublishingPropertiesInput = {
10521
10554
  count?: InputMaybe<Scalars['Int']['input']>;
10522
10555
  /** The OpenAI Image model. */
10523
10556
  model?: InputMaybe<OpenAiImageModels>;
10524
- /** The seed image reference to use when generating image(s). */
10557
+ /** The seed image reference to use when generating image(s), optional. */
10525
10558
  seed?: InputMaybe<EntityReferenceInput>;
10526
10559
  };
10527
10560
  /** Represents OpenAI model properties. */
@@ -10768,6 +10801,38 @@ export declare enum OpenAiReasoningEffortLevels {
10768
10801
  /** Medium effort */
10769
10802
  Medium = "MEDIUM"
10770
10803
  }
10804
+ /** OpenAI Video model type */
10805
+ export declare enum OpenAiVideoModels {
10806
+ /** Developer-specified model */
10807
+ Custom = "CUSTOM",
10808
+ /** Sora 2 (Latest) - Fast generation */
10809
+ Sora_2 = "SORA_2",
10810
+ /** Sora 2 Pro - High quality production */
10811
+ Sora_2Pro = "SORA_2_PRO"
10812
+ }
10813
+ /** Represents the OpenAI Video publishing properties. */
10814
+ export type OpenAiVideoPublishingProperties = {
10815
+ __typename?: 'OpenAIVideoPublishingProperties';
10816
+ /** The OpenAI Video model. */
10817
+ model?: Maybe<OpenAiVideoModels>;
10818
+ /** The video duration in seconds, optional. Must be 4, 8, or 12. Defaults to 4. */
10819
+ seconds?: Maybe<Scalars['Int']['output']>;
10820
+ /** The seed image reference to use when generating video, optional. */
10821
+ seed?: Maybe<EntityReference>;
10822
+ /** The video resolution size, optional. */
10823
+ size?: Maybe<VideoSizeTypes>;
10824
+ };
10825
+ /** Represents the OpenAI Video publishing properties. */
10826
+ export type OpenAiVideoPublishingPropertiesInput = {
10827
+ /** The OpenAI Video model. */
10828
+ model?: InputMaybe<OpenAiVideoModels>;
10829
+ /** The video duration in seconds, optional. Must be 4, 8, or 12. Defaults to 4. */
10830
+ seconds?: InputMaybe<Scalars['Int']['input']>;
10831
+ /** The seed image reference to use when generating video, optional. */
10832
+ seed?: InputMaybe<EntityReferenceInput>;
10833
+ /** The video resolution size, optional. */
10834
+ size?: InputMaybe<VideoSizeTypes>;
10835
+ };
10771
10836
  /** OpenAI vision model detail levels */
10772
10837
  export declare enum OpenAiVisionDetailLevels {
10773
10838
  /** High */
@@ -13730,10 +13795,14 @@ export type SharePointLibraryResult = {
13730
13795
  libraryId?: Maybe<Scalars['ID']['output']>;
13731
13796
  /** The SharePoint library name. */
13732
13797
  libraryName?: Maybe<Scalars['String']['output']>;
13798
+ /** The SharePoint library web URL. */
13799
+ libraryWebUrl?: Maybe<Scalars['String']['output']>;
13733
13800
  /** The SharePoint site identifier. */
13734
13801
  siteId?: Maybe<Scalars['ID']['output']>;
13735
13802
  /** The SharePoint site name. */
13736
13803
  siteName?: Maybe<Scalars['String']['output']>;
13804
+ /** The SharePoint site web URL. */
13805
+ siteWebUrl?: Maybe<Scalars['String']['output']>;
13737
13806
  };
13738
13807
  /** Represents SharePoint libraries. */
13739
13808
  export type SharePointLibraryResults = {
@@ -14306,6 +14375,8 @@ export type StoragePolicy = {
14306
14375
  __typename?: 'StoragePolicy';
14307
14376
  /** Whether duplicate content (by URI, eTag, etc.) will be allowed upon ingestion, defaults to False. When disabled, content will be reingested in-place. */
14308
14377
  allowDuplicates?: Maybe<Scalars['Boolean']['output']>;
14378
+ /** The types of embeddings to generate during enrichment. Defaults to all types if not specified. Specify an empty array to disable all embeddings. */
14379
+ embeddingTypes?: Maybe<Array<Maybe<EmbeddingTypes>>>;
14309
14380
  /** The storage policy type. */
14310
14381
  type?: Maybe<StoragePolicyTypes>;
14311
14382
  };
@@ -14313,6 +14384,8 @@ export type StoragePolicy = {
14313
14384
  export type StoragePolicyInput = {
14314
14385
  /** Whether duplicate content (by URI, eTag, etc.) will be allowed upon ingestion, defaults to False. When disabled, content will be reingested in-place. */
14315
14386
  allowDuplicates?: InputMaybe<Scalars['Boolean']['input']>;
14387
+ /** The types of embeddings to generate during enrichment. Defaults to all types if not specified. Specify an empty array to disable all embeddings. */
14388
+ embeddingTypes?: InputMaybe<Array<InputMaybe<EmbeddingTypes>>>;
14316
14389
  /** The storage policy type. */
14317
14390
  type?: InputMaybe<StoragePolicyTypes>;
14318
14391
  };
@@ -14910,6 +14983,17 @@ export type VideoMetadataInput = {
14910
14983
  /** The video width. */
14911
14984
  width?: InputMaybe<Scalars['Int']['input']>;
14912
14985
  };
14986
+ /** Video size (resolution) */
14987
+ export declare enum VideoSizeTypes {
14988
+ /** Full HD landscape */
14989
+ FullHdLandscape = "FULL_HD_LANDSCAPE",
14990
+ /** Full HD portrait */
14991
+ FullHdPortrait = "FULL_HD_PORTRAIT",
14992
+ /** HD landscape */
14993
+ HdLandscape = "HD_LANDSCAPE",
14994
+ /** HD portrait */
14995
+ HdPortrait = "HD_PORTRAIT"
14996
+ }
14913
14997
  /** Represents a view. */
14914
14998
  export type View = {
14915
14999
  __typename?: 'View';
@@ -15680,6 +15764,16 @@ export type GetAlertQuery = {
15680
15764
  id: string;
15681
15765
  } | null;
15682
15766
  } | null;
15767
+ openAIVideo?: {
15768
+ __typename?: 'OpenAIVideoPublishingProperties';
15769
+ model?: OpenAiVideoModels | null;
15770
+ seconds?: number | null;
15771
+ size?: VideoSizeTypes | null;
15772
+ seed?: {
15773
+ __typename?: 'EntityReference';
15774
+ id: string;
15775
+ } | null;
15776
+ } | null;
15683
15777
  };
15684
15778
  summarySpecification?: {
15685
15779
  __typename?: 'EntityReference';
@@ -15897,6 +15991,16 @@ export type QueryAlertsQuery = {
15897
15991
  id: string;
15898
15992
  } | null;
15899
15993
  } | null;
15994
+ openAIVideo?: {
15995
+ __typename?: 'OpenAIVideoPublishingProperties';
15996
+ model?: OpenAiVideoModels | null;
15997
+ seconds?: number | null;
15998
+ size?: VideoSizeTypes | null;
15999
+ seed?: {
16000
+ __typename?: 'EntityReference';
16001
+ id: string;
16002
+ } | null;
16003
+ } | null;
15900
16004
  };
15901
16005
  summarySpecification?: {
15902
16006
  __typename?: 'EntityReference';
@@ -28643,6 +28747,7 @@ export type CreateWorkflowMutation = {
28643
28747
  __typename?: 'IngestionWorkflowStage';
28644
28748
  enableEmailCollections?: boolean | null;
28645
28749
  enableFolderCollections?: boolean | null;
28750
+ enableMessageCollections?: boolean | null;
28646
28751
  if?: {
28647
28752
  __typename?: 'IngestionContentFilter';
28648
28753
  types?: Array<ContentTypes> | null;
@@ -28872,6 +28977,7 @@ export type CreateWorkflowMutation = {
28872
28977
  __typename?: 'StoragePolicy';
28873
28978
  type?: StoragePolicyTypes | null;
28874
28979
  allowDuplicates?: boolean | null;
28980
+ embeddingTypes?: Array<EmbeddingTypes | null> | null;
28875
28981
  } | null;
28876
28982
  } | null;
28877
28983
  actions?: Array<{
@@ -28964,6 +29070,7 @@ export type GetWorkflowQuery = {
28964
29070
  __typename?: 'IngestionWorkflowStage';
28965
29071
  enableEmailCollections?: boolean | null;
28966
29072
  enableFolderCollections?: boolean | null;
29073
+ enableMessageCollections?: boolean | null;
28967
29074
  if?: {
28968
29075
  __typename?: 'IngestionContentFilter';
28969
29076
  types?: Array<ContentTypes> | null;
@@ -29193,6 +29300,7 @@ export type GetWorkflowQuery = {
29193
29300
  __typename?: 'StoragePolicy';
29194
29301
  type?: StoragePolicyTypes | null;
29195
29302
  allowDuplicates?: boolean | null;
29303
+ embeddingTypes?: Array<EmbeddingTypes | null> | null;
29196
29304
  } | null;
29197
29305
  } | null;
29198
29306
  actions?: Array<{
@@ -29251,6 +29359,7 @@ export type QueryWorkflowsQuery = {
29251
29359
  __typename?: 'IngestionWorkflowStage';
29252
29360
  enableEmailCollections?: boolean | null;
29253
29361
  enableFolderCollections?: boolean | null;
29362
+ enableMessageCollections?: boolean | null;
29254
29363
  if?: {
29255
29364
  __typename?: 'IngestionContentFilter';
29256
29365
  types?: Array<ContentTypes> | null;
@@ -29480,6 +29589,7 @@ export type QueryWorkflowsQuery = {
29480
29589
  __typename?: 'StoragePolicy';
29481
29590
  type?: StoragePolicyTypes | null;
29482
29591
  allowDuplicates?: boolean | null;
29592
+ embeddingTypes?: Array<EmbeddingTypes | null> | null;
29483
29593
  } | null;
29484
29594
  } | null;
29485
29595
  actions?: Array<{
@@ -29530,6 +29640,7 @@ export type UpdateWorkflowMutation = {
29530
29640
  __typename?: 'IngestionWorkflowStage';
29531
29641
  enableEmailCollections?: boolean | null;
29532
29642
  enableFolderCollections?: boolean | null;
29643
+ enableMessageCollections?: boolean | null;
29533
29644
  if?: {
29534
29645
  __typename?: 'IngestionContentFilter';
29535
29646
  types?: Array<ContentTypes> | null;
@@ -29759,6 +29870,7 @@ export type UpdateWorkflowMutation = {
29759
29870
  __typename?: 'StoragePolicy';
29760
29871
  type?: StoragePolicyTypes | null;
29761
29872
  allowDuplicates?: boolean | null;
29873
+ embeddingTypes?: Array<EmbeddingTypes | null> | null;
29762
29874
  } | null;
29763
29875
  } | null;
29764
29876
  actions?: Array<{
@@ -29808,6 +29920,7 @@ export type UpsertWorkflowMutation = {
29808
29920
  __typename?: 'IngestionWorkflowStage';
29809
29921
  enableEmailCollections?: boolean | null;
29810
29922
  enableFolderCollections?: boolean | null;
29923
+ enableMessageCollections?: boolean | null;
29811
29924
  if?: {
29812
29925
  __typename?: 'IngestionContentFilter';
29813
29926
  types?: Array<ContentTypes> | null;
@@ -30037,6 +30150,7 @@ export type UpsertWorkflowMutation = {
30037
30150
  __typename?: 'StoragePolicy';
30038
30151
  type?: StoragePolicyTypes | null;
30039
30152
  allowDuplicates?: boolean | null;
30153
+ embeddingTypes?: Array<EmbeddingTypes | null> | null;
30040
30154
  } | null;
30041
30155
  } | null;
30042
30156
  actions?: Array<{
@@ -438,6 +438,8 @@ export var ContentPublishingFormats;
438
438
  ContentPublishingFormats["Markdown"] = "MARKDOWN";
439
439
  /** MP3 */
440
440
  ContentPublishingFormats["Mp3"] = "MP3";
441
+ /** MP4 */
442
+ ContentPublishingFormats["Mp4"] = "MP4";
441
443
  /** PNG */
442
444
  ContentPublishingFormats["Png"] = "PNG";
443
445
  /** Plain Text */
@@ -454,6 +456,8 @@ export var ContentPublishingServiceTypes;
454
456
  ContentPublishingServiceTypes["GoogleImage"] = "GOOGLE_IMAGE";
455
457
  /** OpenAI Image publishing */
456
458
  ContentPublishingServiceTypes["OpenAiImage"] = "OPEN_AI_IMAGE";
459
+ /** OpenAI Video publishing */
460
+ ContentPublishingServiceTypes["OpenAiVideo"] = "OPEN_AI_VIDEO";
457
461
  /** Text publishing */
458
462
  ContentPublishingServiceTypes["Text"] = "TEXT";
459
463
  })(ContentPublishingServiceTypes || (ContentPublishingServiceTypes = {}));
@@ -633,6 +637,18 @@ export var EmailListingTypes;
633
637
  /** Read past emails */
634
638
  EmailListingTypes["Past"] = "PAST";
635
639
  })(EmailListingTypes || (EmailListingTypes = {}));
640
+ /** Embedding type */
641
+ export var EmbeddingTypes;
642
+ (function (EmbeddingTypes) {
643
+ /** Audio embeddings */
644
+ EmbeddingTypes["Audio"] = "AUDIO";
645
+ /** Image embeddings */
646
+ EmbeddingTypes["Image"] = "IMAGE";
647
+ /** Text embeddings */
648
+ EmbeddingTypes["Text"] = "TEXT";
649
+ /** Video embeddings */
650
+ EmbeddingTypes["Video"] = "VIDEO";
651
+ })(EmbeddingTypes || (EmbeddingTypes = {}));
636
652
  /** Entity enrichment service types */
637
653
  export var EntityEnrichmentServiceTypes;
638
654
  (function (EntityEnrichmentServiceTypes) {
@@ -1651,6 +1667,8 @@ export var OpenAiImageModels;
1651
1667
  OpenAiImageModels["Custom"] = "CUSTOM";
1652
1668
  /** GPT Image-1 */
1653
1669
  OpenAiImageModels["GptImage_1"] = "GPT_IMAGE_1";
1670
+ /** GPT Image-1 Mini */
1671
+ OpenAiImageModels["GptImage_1Mini"] = "GPT_IMAGE_1_MINI";
1654
1672
  })(OpenAiImageModels || (OpenAiImageModels = {}));
1655
1673
  /** OpenAI model type */
1656
1674
  export var OpenAiModels;
@@ -1822,6 +1840,16 @@ export var OpenAiReasoningEffortLevels;
1822
1840
  /** Medium effort */
1823
1841
  OpenAiReasoningEffortLevels["Medium"] = "MEDIUM";
1824
1842
  })(OpenAiReasoningEffortLevels || (OpenAiReasoningEffortLevels = {}));
1843
+ /** OpenAI Video model type */
1844
+ export var OpenAiVideoModels;
1845
+ (function (OpenAiVideoModels) {
1846
+ /** Developer-specified model */
1847
+ OpenAiVideoModels["Custom"] = "CUSTOM";
1848
+ /** Sora 2 (Latest) - Fast generation */
1849
+ OpenAiVideoModels["Sora_2"] = "SORA_2";
1850
+ /** Sora 2 Pro - High quality production */
1851
+ OpenAiVideoModels["Sora_2Pro"] = "SORA_2_PRO";
1852
+ })(OpenAiVideoModels || (OpenAiVideoModels = {}));
1825
1853
  /** OpenAI vision model detail levels */
1826
1854
  export var OpenAiVisionDetailLevels;
1827
1855
  (function (OpenAiVisionDetailLevels) {
@@ -2324,6 +2352,18 @@ export var UserTypes;
2324
2352
  /** Human user */
2325
2353
  UserTypes["Human"] = "HUMAN";
2326
2354
  })(UserTypes || (UserTypes = {}));
2355
+ /** Video size (resolution) */
2356
+ export var VideoSizeTypes;
2357
+ (function (VideoSizeTypes) {
2358
+ /** Full HD landscape */
2359
+ VideoSizeTypes["FullHdLandscape"] = "FULL_HD_LANDSCAPE";
2360
+ /** Full HD portrait */
2361
+ VideoSizeTypes["FullHdPortrait"] = "FULL_HD_PORTRAIT";
2362
+ /** HD landscape */
2363
+ VideoSizeTypes["HdLandscape"] = "HD_LANDSCAPE";
2364
+ /** HD portrait */
2365
+ VideoSizeTypes["HdPortrait"] = "HD_PORTRAIT";
2366
+ })(VideoSizeTypes || (VideoSizeTypes = {}));
2327
2367
  /** View type */
2328
2368
  export var ViewTypes;
2329
2369
  (function (ViewTypes) {
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "graphlit-client",
3
- "version": "1.0.20251005001",
3
+ "version": "1.0.20251007001",
4
4
  "description": "Graphlit API Client for TypeScript",
5
5
  "type": "module",
6
6
  "main": "./dist/client.js",