graphlit-client 1.0.20241009001 → 1.0.20241021001

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -304,6 +304,8 @@ export type AmazonFeedPropertiesUpdateInput = {
304
304
  /** Represents Anthropic model properties. */
305
305
  export type AnthropicModelProperties = {
306
306
  __typename?: 'AnthropicModelProperties';
307
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
308
+ chunkTokenLimit?: Maybe<Scalars['Int']['output']>;
307
309
  /** The limit of tokens generated by prompt completion. */
308
310
  completionTokenLimit?: Maybe<Scalars['Int']['output']>;
309
311
  /** The Anthropic API key, if using developer's own account. */
@@ -499,6 +501,8 @@ export type AudioMetadataInput = {
499
501
  /** Represents Azure AI model properties. */
500
502
  export type AzureAiModelProperties = {
501
503
  __typename?: 'AzureAIModelProperties';
504
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
505
+ chunkTokenLimit?: Maybe<Scalars['Int']['output']>;
502
506
  /** The limit of tokens generated by prompt completion. */
503
507
  completionTokenLimit?: Maybe<Scalars['Int']['output']>;
504
508
  /** The Azure AI API endpoint. */
@@ -514,6 +518,8 @@ export type AzureAiModelProperties = {
514
518
  };
515
519
  /** Represents Azure AI model properties. */
516
520
  export type AzureAiModelPropertiesInput = {
521
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
522
+ chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
517
523
  /** The limit of tokens generated by prompt completion. */
518
524
  completionTokenLimit?: InputMaybe<Scalars['Int']['input']>;
519
525
  /** The Azure AI API endpoint. */
@@ -529,6 +535,8 @@ export type AzureAiModelPropertiesInput = {
529
535
  };
530
536
  /** Represents Azure AI model properties. */
531
537
  export type AzureAiModelPropertiesUpdateInput = {
538
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
539
+ chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
532
540
  /** The limit of tokens generated by prompt completion. */
533
541
  completionTokenLimit?: InputMaybe<Scalars['Int']['input']>;
534
542
  /** The Azure AI API endpoint. */
@@ -692,6 +700,8 @@ export type AzureImageExtractionPropertiesInput = {
692
700
  /** Represents Azure OpenAI model properties. */
693
701
  export type AzureOpenAiModelProperties = {
694
702
  __typename?: 'AzureOpenAIModelProperties';
703
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
704
+ chunkTokenLimit?: Maybe<Scalars['Int']['output']>;
695
705
  /** The limit of tokens generated by prompt completion. */
696
706
  completionTokenLimit?: Maybe<Scalars['Int']['output']>;
697
707
  /** The Azure OpenAI deployment name, if using developer's own account. */
@@ -711,6 +721,8 @@ export type AzureOpenAiModelProperties = {
711
721
  };
712
722
  /** Represents Azure OpenAI model properties. */
713
723
  export type AzureOpenAiModelPropertiesInput = {
724
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
725
+ chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
714
726
  /** The limit of tokens generated by prompt completion. */
715
727
  completionTokenLimit?: InputMaybe<Scalars['Int']['input']>;
716
728
  /** The Azure OpenAI deployment name, if using developer's own account. */
@@ -730,6 +742,8 @@ export type AzureOpenAiModelPropertiesInput = {
730
742
  };
731
743
  /** Represents Azure OpenAI model properties. */
732
744
  export type AzureOpenAiModelPropertiesUpdateInput = {
745
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
746
+ chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
733
747
  /** The limit of tokens generated by prompt completion. */
734
748
  completionTokenLimit?: InputMaybe<Scalars['Int']['input']>;
735
749
  /** The Azure OpenAI deployment name, if using developer's own account. */
@@ -751,11 +765,11 @@ export type AzureOpenAiModelPropertiesUpdateInput = {
751
765
  export declare enum AzureOpenAiModels {
752
766
  /** Developer-specified deployment */
753
767
  Custom = "CUSTOM",
754
- /** GPT-4 (Latest version) */
768
+ /** GPT-4 (Latest) */
755
769
  Gpt4 = "GPT4",
756
- /** GPT-4 Turbo 128k (Latest version) */
770
+ /** GPT-4 Turbo 128k (Latest) */
757
771
  Gpt4Turbo_128K = "GPT4_TURBO_128K",
758
- /** GPT-3.5 Turbo 16k (Latest version) */
772
+ /** GPT-3.5 Turbo 16k (Latest) */
759
773
  Gpt35Turbo_16K = "GPT35_TURBO_16K"
760
774
  }
761
775
  /** Represents an Azure Cognitive Services text entity extraction connector. */
@@ -973,6 +987,8 @@ export declare enum CerebrasModels {
973
987
  /** Represents Cohere model properties. */
974
988
  export type CohereModelProperties = {
975
989
  __typename?: 'CohereModelProperties';
990
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
991
+ chunkTokenLimit?: Maybe<Scalars['Int']['output']>;
976
992
  /** The limit of tokens generated by prompt completion. */
977
993
  completionTokenLimit?: Maybe<Scalars['Int']['output']>;
978
994
  /** The Cohere API key, if using developer's own account. */
@@ -990,6 +1006,8 @@ export type CohereModelProperties = {
990
1006
  };
991
1007
  /** Represents Cohere model properties. */
992
1008
  export type CohereModelPropertiesInput = {
1009
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
1010
+ chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
993
1011
  /** The limit of tokens generated by prompt completion. */
994
1012
  completionTokenLimit?: InputMaybe<Scalars['Int']['input']>;
995
1013
  /** The Cohere API key, if using developer's own account. */
@@ -1007,6 +1025,8 @@ export type CohereModelPropertiesInput = {
1007
1025
  };
1008
1026
  /** Represents Cohere model properties. */
1009
1027
  export type CohereModelPropertiesUpdateInput = {
1028
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
1029
+ chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
1010
1030
  /** The limit of tokens generated by prompt completion. */
1011
1031
  completionTokenLimit?: InputMaybe<Scalars['Int']['input']>;
1012
1032
  /** The Cohere API key, if using developer's own account. */
@@ -1037,7 +1057,11 @@ export declare enum CohereModels {
1037
1057
  /** Command R+ (2024-08 version) */
1038
1058
  CommandRPlus_202408 = "COMMAND_R_PLUS_202408",
1039
1059
  /** Developer-specified model */
1040
- Custom = "CUSTOM"
1060
+ Custom = "CUSTOM",
1061
+ /** Embed English 3.0 */
1062
+ EmbedEnglish_3_0 = "EMBED_ENGLISH_3_0",
1063
+ /** Embed Multilingual 3.0 */
1064
+ EmbedMultilingual_3_0 = "EMBED_MULTILINGUAL_3_0"
1041
1065
  }
1042
1066
  /** Represents a collection. */
1043
1067
  export type Collection = {
@@ -1837,14 +1861,10 @@ export declare enum ConversationSearchTypes {
1837
1861
  /** Represents a conversation strategy. */
1838
1862
  export type ConversationStrategy = {
1839
1863
  __typename?: 'ConversationStrategy';
1840
- /** @deprecated Moved to retrieval strategy. */
1841
- contentLimit?: Maybe<Scalars['Int']['output']>;
1842
1864
  /** The weight of contents within prompt context, in range [0.0 - 1.0]. */
1843
1865
  contentsWeight?: Maybe<Scalars['Float']['output']>;
1844
1866
  /** Embed content citations into completed converation messages. */
1845
1867
  embedCitations?: Maybe<Scalars['Boolean']['output']>;
1846
- /** @deprecated Moved to retrieval strategy. */
1847
- enableExpandedRetrieval?: Maybe<Scalars['Boolean']['output']>;
1848
1868
  /** Provide content facets with completed conversation. */
1849
1869
  enableFacets?: Maybe<Scalars['Boolean']['output']>;
1850
1870
  /** The maximum number of retrieval user messages to provide with prompt context. Defaults to 5. */
@@ -2437,13 +2457,19 @@ export type EmailPreparationPropertiesInput = {
2437
2457
  /** Represents the embeddings strategy. */
2438
2458
  export type EmbeddingsStrategy = {
2439
2459
  __typename?: 'EmbeddingsStrategy';
2440
- /** The limit of tokens per embedded text chunk, defaults to 600. */
2460
+ /** @deprecated The limit of tokens per embedded text chunk has been removed from embeddings strategy. Assign in text embeddings specification instead. */
2441
2461
  chunkTokenLimit?: Maybe<Scalars['Int']['output']>;
2462
+ /** The LLM specification used for image embeddings. */
2463
+ imageSpecification?: Maybe<EntityReference>;
2464
+ /** The LLM specification used for text embeddings. */
2465
+ textSpecification?: Maybe<EntityReference>;
2442
2466
  };
2443
2467
  /** Represents the embeddings strategy. */
2444
2468
  export type EmbeddingsStrategyInput = {
2445
- /** The limit of tokens per embedded text chunk, defaults to 600. */
2446
- chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
2469
+ /** The LLM specification used for image embeddings. */
2470
+ imageSpecification?: InputMaybe<EntityReferenceInput>;
2471
+ /** The LLM specification used for text embeddings. */
2472
+ textSpecification?: InputMaybe<EntityReferenceInput>;
2447
2473
  };
2448
2474
  /** Represents an enrichment workflow job. */
2449
2475
  export type EnrichmentWorkflowJob = {
@@ -2520,10 +2546,7 @@ export type EntityExtractionConnector = {
2520
2546
  modelImage?: Maybe<ModelImageExtractionProperties>;
2521
2547
  /** The specific properties for LLM text entity extraction. */
2522
2548
  modelText?: Maybe<ModelTextExtractionProperties>;
2523
- /**
2524
- * The specific properties for OpenAI image entity extraction.
2525
- * @deprecated Use MODEL_IMAGE instead.
2526
- */
2549
+ /** @deprecated The specific properties for OpenAI image entity extraction have been removed. Use LLM image entity extraction instead. */
2527
2550
  openAIImage?: Maybe<OpenAiImageExtractionProperties>;
2528
2551
  /** The entity extraction connector service type. */
2529
2552
  type: EntityExtractionServiceTypes;
@@ -3555,6 +3578,8 @@ export type GoogleFeedPropertiesUpdateInput = {
3555
3578
  /** Represents Google model properties. */
3556
3579
  export type GoogleModelProperties = {
3557
3580
  __typename?: 'GoogleModelProperties';
3581
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
3582
+ chunkTokenLimit?: Maybe<Scalars['Int']['output']>;
3558
3583
  /** The limit of tokens generated by prompt completion. */
3559
3584
  completionTokenLimit?: Maybe<Scalars['Int']['output']>;
3560
3585
  /** The Google API key, if using developer's own account. */
@@ -3572,6 +3597,8 @@ export type GoogleModelProperties = {
3572
3597
  };
3573
3598
  /** Represents Google model properties. */
3574
3599
  export type GoogleModelPropertiesInput = {
3600
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
3601
+ chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
3575
3602
  /** The limit of tokens generated by prompt completion. */
3576
3603
  completionTokenLimit?: InputMaybe<Scalars['Int']['input']>;
3577
3604
  /** The Google API key, if using developer's own account. */
@@ -3589,6 +3616,8 @@ export type GoogleModelPropertiesInput = {
3589
3616
  };
3590
3617
  /** Represents Google model properties. */
3591
3618
  export type GoogleModelPropertiesUpdateInput = {
3619
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
3620
+ chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
3592
3621
  /** The limit of tokens generated by prompt completion. */
3593
3622
  completionTokenLimit?: InputMaybe<Scalars['Int']['input']>;
3594
3623
  /** The Google API key, if using developer's own account. */
@@ -3608,6 +3637,8 @@ export type GoogleModelPropertiesUpdateInput = {
3608
3637
  export declare enum GoogleModels {
3609
3638
  /** Developer-specified model */
3610
3639
  Custom = "CUSTOM",
3640
+ /** Embed (004 version) */
3641
+ Embedding_004 = "EMBEDDING_004",
3611
3642
  /** Gemini 1.5 Flash (Latest) */
3612
3643
  Gemini_1_5Flash = "GEMINI_1_5_FLASH",
3613
3644
  /** Gemini 1.5 Flash (001 version) */
@@ -4193,6 +4224,51 @@ export type IssueMetadataInput = {
4193
4224
  /** The issue type, i.e. epic, story, task. */
4194
4225
  type?: InputMaybe<Scalars['String']['input']>;
4195
4226
  };
4227
+ /** Represents Jina model properties. */
4228
+ export type JinaModelProperties = {
4229
+ __typename?: 'JinaModelProperties';
4230
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
4231
+ chunkTokenLimit?: Maybe<Scalars['Int']['output']>;
4232
+ /** The Jina API key, if using developer's own account. */
4233
+ key?: Maybe<Scalars['String']['output']>;
4234
+ /** The Jina model, or custom, when using developer's own account. */
4235
+ model: JinaModels;
4236
+ /** The Jina model name, if using developer's own account. */
4237
+ modelName?: Maybe<Scalars['String']['output']>;
4238
+ };
4239
+ /** Represents Jina model properties. */
4240
+ export type JinaModelPropertiesInput = {
4241
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
4242
+ chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
4243
+ /** The Jina API key, if using developer's own account. */
4244
+ key?: InputMaybe<Scalars['String']['input']>;
4245
+ /** The Jina model, or custom, when using developer's own account. */
4246
+ model: JinaModels;
4247
+ /** The Jina model name, if using developer's own account. */
4248
+ modelName?: InputMaybe<Scalars['String']['input']>;
4249
+ };
4250
+ /** Represents Jina model properties. */
4251
+ export type JinaModelPropertiesUpdateInput = {
4252
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
4253
+ chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
4254
+ /** The Jina API key, if using developer's own account. */
4255
+ key?: InputMaybe<Scalars['String']['input']>;
4256
+ /** The Jina model, or custom, when using developer's own account. */
4257
+ model?: InputMaybe<JinaModels>;
4258
+ /** The Jina model name, if using developer's own account. */
4259
+ modelName?: InputMaybe<Scalars['String']['input']>;
4260
+ };
4261
+ /** Jina model type */
4262
+ export declare enum JinaModels {
4263
+ /** CLIP Image */
4264
+ ClipImage = "CLIP_IMAGE",
4265
+ /** Developer-specified model */
4266
+ Custom = "CUSTOM",
4267
+ /** Embed (Latest) */
4268
+ Embed = "EMBED",
4269
+ /** Embed 3.0 */
4270
+ Embed_3_0 = "EMBED_3_0"
4271
+ }
4196
4272
  /** Represents a label. */
4197
4273
  export type Label = {
4198
4274
  __typename?: 'Label';
@@ -6254,6 +6330,8 @@ export type MicrosoftTeamsTeamsInput = {
6254
6330
  /** Represents Mistral model properties. */
6255
6331
  export type MistralModelProperties = {
6256
6332
  __typename?: 'MistralModelProperties';
6333
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
6334
+ chunkTokenLimit?: Maybe<Scalars['Int']['output']>;
6257
6335
  /** The limit of tokens generated by prompt completion. */
6258
6336
  completionTokenLimit?: Maybe<Scalars['Int']['output']>;
6259
6337
  /** The Mistral API endpoint, if using developer's own account. */
@@ -6273,6 +6351,8 @@ export type MistralModelProperties = {
6273
6351
  };
6274
6352
  /** Represents Mistral model properties. */
6275
6353
  export type MistralModelPropertiesInput = {
6354
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
6355
+ chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
6276
6356
  /** The limit of tokens generated by prompt completion. */
6277
6357
  completionTokenLimit?: InputMaybe<Scalars['Int']['input']>;
6278
6358
  /** The Mistral API endpoint, if using developer's own account. */
@@ -6292,6 +6372,8 @@ export type MistralModelPropertiesInput = {
6292
6372
  };
6293
6373
  /** Represents Mistral model properties. */
6294
6374
  export type MistralModelPropertiesUpdateInput = {
6375
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
6376
+ chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
6295
6377
  /** The limit of tokens generated by prompt completion. */
6296
6378
  completionTokenLimit?: InputMaybe<Scalars['Int']['input']>;
6297
6379
  /** The Mistral API endpoint, if using developer's own account. */
@@ -6313,6 +6395,8 @@ export type MistralModelPropertiesUpdateInput = {
6313
6395
  export declare enum MistralModels {
6314
6396
  /** Developer-specified model */
6315
6397
  Custom = "CUSTOM",
6398
+ /** Mistral Embed */
6399
+ MistralEmbed = "MISTRAL_EMBED",
6316
6400
  /** Mistral Large */
6317
6401
  MistralLarge = "MISTRAL_LARGE",
6318
6402
  /** Mistral Medium */
@@ -6366,12 +6450,16 @@ export declare enum ModelServiceTypes {
6366
6450
  Google = "GOOGLE",
6367
6451
  /** Groq */
6368
6452
  Groq = "GROQ",
6453
+ /** Jina */
6454
+ Jina = "JINA",
6369
6455
  /** Mistral */
6370
6456
  Mistral = "MISTRAL",
6371
6457
  /** OpenAI */
6372
6458
  OpenAi = "OPEN_AI",
6373
6459
  /** Replicate */
6374
- Replicate = "REPLICATE"
6460
+ Replicate = "REPLICATE",
6461
+ /** Voyage */
6462
+ Voyage = "VOYAGE"
6375
6463
  }
6376
6464
  /** Represents an LLM text entity extraction connector. */
6377
6465
  export type ModelTextExtractionProperties = {
@@ -6645,6 +6733,7 @@ export type Mutation = {
6645
6733
  ingestPage?: Maybe<Content>;
6646
6734
  /** Ingests text. */
6647
6735
  ingestText?: Maybe<Content>;
6736
+ ingestTextBatch?: Maybe<Array<Maybe<Content>>>;
6648
6737
  /** Ingests content by URI. Supports files and webpages. */
6649
6738
  ingestUri?: Maybe<Content>;
6650
6739
  /** Opens an existing collection. */
@@ -7234,6 +7323,13 @@ export type MutationIngestTextArgs = {
7234
7323
  uri?: InputMaybe<Scalars['URL']['input']>;
7235
7324
  workflow?: InputMaybe<EntityReferenceInput>;
7236
7325
  };
7326
+ export type MutationIngestTextBatchArgs = {
7327
+ batch: Array<TextContentInput>;
7328
+ collections?: InputMaybe<Array<EntityReferenceInput>>;
7329
+ correlationId?: InputMaybe<Scalars['String']['input']>;
7330
+ textType?: InputMaybe<TextTypes>;
7331
+ workflow?: InputMaybe<EntityReferenceInput>;
7332
+ };
7237
7333
  export type MutationIngestUriArgs = {
7238
7334
  collections?: InputMaybe<Array<EntityReferenceInput>>;
7239
7335
  correlationId?: InputMaybe<Scalars['String']['input']>;
@@ -7697,21 +7793,14 @@ export type OpenAiImageExtractionProperties = {
7697
7793
  /** The OpenAI vision detail mode. */
7698
7794
  detailLevel?: Maybe<OpenAiVisionDetailLevels>;
7699
7795
  };
7700
- /** Represents an OpenAI image entity extraction connector. */
7701
- export type OpenAiImageExtractionPropertiesInput = {
7702
- /** The confidence threshold for entity extraction. */
7703
- confidenceThreshold?: InputMaybe<Scalars['Float']['input']>;
7704
- /** Custom instructions which are injected into the LLM prompt. */
7705
- customInstructions?: InputMaybe<Scalars['String']['input']>;
7706
- /** The OpenAI vision detail mode. */
7707
- detailLevel?: InputMaybe<OpenAiVisionDetailLevels>;
7708
- };
7709
7796
  /** Represents OpenAI model properties. */
7710
7797
  export type OpenAiModelProperties = {
7711
7798
  __typename?: 'OpenAIModelProperties';
7799
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
7800
+ chunkTokenLimit?: Maybe<Scalars['Int']['output']>;
7712
7801
  /** The limit of tokens generated by prompt completion. */
7713
7802
  completionTokenLimit?: Maybe<Scalars['Int']['output']>;
7714
- /** The OpenAI vision detail mode. Only applies when using OpenAI for image analysis. */
7803
+ /** The OpenAI vision detail mode. Only applies when using OpenAI for image completion. */
7715
7804
  detailLevel?: Maybe<OpenAiVisionDetailLevels>;
7716
7805
  /** The OpenAI API key, if using developer's own account. */
7717
7806
  key?: Maybe<Scalars['String']['output']>;
@@ -7728,9 +7817,11 @@ export type OpenAiModelProperties = {
7728
7817
  };
7729
7818
  /** Represents OpenAI model properties. */
7730
7819
  export type OpenAiModelPropertiesInput = {
7820
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
7821
+ chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
7731
7822
  /** The limit of tokens generated by prompt completion. */
7732
7823
  completionTokenLimit?: InputMaybe<Scalars['Int']['input']>;
7733
- /** The OpenAI vision detail mode. Only applies when using OpenAI for image analysis. */
7824
+ /** The OpenAI vision detail mode. Only applies when using OpenAI for image completion. */
7734
7825
  detailLevel?: InputMaybe<OpenAiVisionDetailLevels>;
7735
7826
  /** The OpenAI API key, if using developer's own account. */
7736
7827
  key?: InputMaybe<Scalars['String']['input']>;
@@ -7747,9 +7838,11 @@ export type OpenAiModelPropertiesInput = {
7747
7838
  };
7748
7839
  /** Represents OpenAI model properties. */
7749
7840
  export type OpenAiModelPropertiesUpdateInput = {
7841
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
7842
+ chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
7750
7843
  /** The limit of tokens generated by prompt completion. */
7751
7844
  completionTokenLimit?: InputMaybe<Scalars['Int']['input']>;
7752
- /** The OpenAI vision detail mode. Only applies when using OpenAI for image analysis. */
7845
+ /** The OpenAI vision detail mode. Only applies when using OpenAI for image completion. */
7753
7846
  detailLevel?: InputMaybe<OpenAiVisionDetailLevels>;
7754
7847
  /** The OpenAI API key, if using developer's own account. */
7755
7848
  key?: InputMaybe<Scalars['String']['input']>;
@@ -7766,8 +7859,14 @@ export type OpenAiModelPropertiesUpdateInput = {
7766
7859
  };
7767
7860
  /** OpenAI model type */
7768
7861
  export declare enum OpenAiModels {
7862
+ /** Embedding Ada-002 */
7863
+ Ada_002 = "ADA_002",
7769
7864
  /** Developer-specified model */
7770
7865
  Custom = "CUSTOM",
7866
+ /** Embedding 3 Large */
7867
+ Embedding_3Large = "EMBEDDING_3_LARGE",
7868
+ /** Embedding 3 Small */
7869
+ Embedding_3Small = "EMBEDDING_3_SMALL",
7771
7870
  /**
7772
7871
  * GPT-4 (Latest)
7773
7872
  * @deprecated OpenAI has deprecated this model. Use the GPT-4o model instead.
@@ -8791,6 +8890,8 @@ export type Project = {
8791
8890
  creationDate: Scalars['DateTime']['output'];
8792
8891
  /** The project credit usage. */
8793
8892
  credits?: Maybe<Scalars['Long']['output']>;
8893
+ /** The project vector storage embeddings strategy. */
8894
+ embeddings?: Maybe<EmbeddingsStrategy>;
8794
8895
  /** The project environment type. */
8795
8896
  environmentType?: Maybe<EnvironmentTypes>;
8796
8897
  /** The ID of the project. */
@@ -9037,6 +9138,8 @@ export type ProjectStorageFileFacet = {
9037
9138
  export type ProjectUpdateInput = {
9038
9139
  /** The project callback URI, optional. The platform will callback to this webhook upon credit charges. */
9039
9140
  callbackUri?: InputMaybe<Scalars['URL']['input']>;
9141
+ /** The project vector storage embeddings strategy. */
9142
+ embeddings?: InputMaybe<EmbeddingsStrategyInput>;
9040
9143
  /** The default LLM specification for conversations. */
9041
9144
  specification?: InputMaybe<EntityReferenceInput>;
9042
9145
  /** The default content workflow. */
@@ -10082,8 +10185,8 @@ export declare enum RerankingModelServiceTypes {
10082
10185
  Cohere = "COHERE",
10083
10186
  /** Jina */
10084
10187
  Jina = "JINA",
10085
- /** Pongo */
10086
- Pongo = "PONGO"
10188
+ /** Voyage */
10189
+ Voyage = "VOYAGE"
10087
10190
  }
10088
10191
  /** Represents a reranking strategy. */
10089
10192
  export type RerankingStrategy = {
@@ -10675,6 +10778,8 @@ export type Specification = {
10675
10778
  groq?: Maybe<GroqModelProperties>;
10676
10779
  /** The ID of the specification. */
10677
10780
  id: Scalars['ID']['output'];
10781
+ /** The Jina model properties. */
10782
+ jina?: Maybe<JinaModelProperties>;
10678
10783
  /** The Mistral model properties. */
10679
10784
  mistral?: Maybe<MistralModelProperties>;
10680
10785
  /** The modified date of the specification. */
@@ -10709,13 +10814,12 @@ export type Specification = {
10709
10814
  strategy?: Maybe<ConversationStrategy>;
10710
10815
  /** The LLM system prompt. */
10711
10816
  systemPrompt?: Maybe<Scalars['String']['output']>;
10712
- /**
10713
- * The tool definitions.
10714
- * @deprecated Tool definitions have been moved to prompt conversation mutation.
10715
- */
10817
+ /** @deprecated The tool definitions have been removed. Tools are now provided to the promptConversation or extractContents mutations. */
10716
10818
  tools?: Maybe<Array<ToolDefinition>>;
10717
10819
  /** The specification type. */
10718
10820
  type?: Maybe<SpecificationTypes>;
10821
+ /** The Voyage model properties. */
10822
+ voyage?: Maybe<VoyageModelProperties>;
10719
10823
  };
10720
10824
  /** Represents a filter for LLM specifications. */
10721
10825
  export type SpecificationFilter = {
@@ -10766,6 +10870,8 @@ export type SpecificationInput = {
10766
10870
  graphStrategy?: InputMaybe<GraphStrategyInput>;
10767
10871
  /** The Groq model properties. */
10768
10872
  groq?: InputMaybe<GroqModelPropertiesInput>;
10873
+ /** The Jina model properties. */
10874
+ jina?: InputMaybe<JinaModelPropertiesInput>;
10769
10875
  /** The Mistral model properties. */
10770
10876
  mistral?: InputMaybe<MistralModelPropertiesInput>;
10771
10877
  /** The name of the specification. */
@@ -10794,6 +10900,8 @@ export type SpecificationInput = {
10794
10900
  systemPrompt?: InputMaybe<Scalars['String']['input']>;
10795
10901
  /** The specification type. */
10796
10902
  type?: InputMaybe<SpecificationTypes>;
10903
+ /** The Voyage model properties. */
10904
+ voyage?: InputMaybe<VoyageModelPropertiesInput>;
10797
10905
  };
10798
10906
  /** Represents LLM specification query results. */
10799
10907
  export type SpecificationResults = {
@@ -10807,8 +10915,12 @@ export declare enum SpecificationTypes {
10807
10915
  Completion = "COMPLETION",
10808
10916
  /** Data extraction */
10809
10917
  Extraction = "EXTRACTION",
10918
+ /** Image embedding */
10919
+ ImageEmbedding = "IMAGE_EMBEDDING",
10810
10920
  /** Document preparation */
10811
- Preparation = "PREPARATION"
10921
+ Preparation = "PREPARATION",
10922
+ /** Text embedding */
10923
+ TextEmbedding = "TEXT_EMBEDDING"
10812
10924
  }
10813
10925
  /** Represents an LLM specification. */
10814
10926
  export type SpecificationUpdateInput = {
@@ -10836,6 +10948,8 @@ export type SpecificationUpdateInput = {
10836
10948
  groq?: InputMaybe<GroqModelPropertiesUpdateInput>;
10837
10949
  /** The ID of the specification to update. */
10838
10950
  id: Scalars['ID']['input'];
10951
+ /** The Jina model properties. */
10952
+ jina?: InputMaybe<JinaModelPropertiesUpdateInput>;
10839
10953
  /** The Mistral model properties. */
10840
10954
  mistral?: InputMaybe<MistralModelPropertiesUpdateInput>;
10841
10955
  /** The name of the specification. */
@@ -10864,18 +10978,15 @@ export type SpecificationUpdateInput = {
10864
10978
  systemPrompt?: InputMaybe<Scalars['String']['input']>;
10865
10979
  /** The specification type. */
10866
10980
  type?: InputMaybe<SpecificationTypes>;
10981
+ /** The Voyage model properties. */
10982
+ voyage?: InputMaybe<VoyageModelPropertiesUpdateInput>;
10867
10983
  };
10868
10984
  /** Represents the storage workflow stage. */
10869
10985
  export type StorageWorkflowStage = {
10870
10986
  __typename?: 'StorageWorkflowStage';
10871
- /** The vector storage embeddings strategy. */
10987
+ /** @deprecated The vector storage embeddings strategy has been removed. Embeddings can be configured through the project embeddings strategy. */
10872
10988
  embeddings?: Maybe<EmbeddingsStrategy>;
10873
10989
  };
10874
- /** Represents the storage workflow stage. */
10875
- export type StorageWorkflowStageInput = {
10876
- /** The vector storage embeddings strategy. */
10877
- embeddings?: InputMaybe<EmbeddingsStrategyInput>;
10878
- };
10879
10990
  /** Represents a range of string values. */
10880
10991
  export type StringRange = {
10881
10992
  __typename?: 'StringRange';
@@ -10966,6 +11077,13 @@ export type TextChunk = {
10966
11077
  /** The text chunk. */
10967
11078
  text?: Maybe<Scalars['String']['output']>;
10968
11079
  };
11080
+ /** Represents text content. */
11081
+ export type TextContentInput = {
11082
+ /** The content name. */
11083
+ name: Scalars['String']['input'];
11084
+ /** The content text. */
11085
+ text: Scalars['String']['input'];
11086
+ };
10969
11087
  /** Represents a frame of image or video. */
10970
11088
  export type TextFrame = {
10971
11089
  __typename?: 'TextFrame';
@@ -11117,15 +11235,6 @@ export type ToolDefinitionInput = {
11117
11235
  /** The tool schema. */
11118
11236
  schema: Scalars['String']['input'];
11119
11237
  };
11120
- /** Represents a tool definition. */
11121
- export type ToolDefinitionUpdateInput = {
11122
- /** The tool description. */
11123
- description?: InputMaybe<Scalars['String']['input']>;
11124
- /** The tool name. */
11125
- name?: InputMaybe<Scalars['String']['input']>;
11126
- /** The tool schema. */
11127
- schema?: InputMaybe<Scalars['String']['input']>;
11128
- };
11129
11238
  /** Unit types */
11130
11239
  export declare enum UnitTypes {
11131
11240
  /** Angstrom */
@@ -11224,6 +11333,59 @@ export type VideoMetadataInput = {
11224
11333
  /** The video width. */
11225
11334
  width?: InputMaybe<Scalars['Int']['input']>;
11226
11335
  };
11336
+ /** Represents Voyage model properties. */
11337
+ export type VoyageModelProperties = {
11338
+ __typename?: 'VoyageModelProperties';
11339
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
11340
+ chunkTokenLimit?: Maybe<Scalars['Int']['output']>;
11341
+ /** The Voyage API key, if using developer's own account. */
11342
+ key?: Maybe<Scalars['String']['output']>;
11343
+ /** The Voyage model, or custom, when using developer's own account. */
11344
+ model: VoyageModels;
11345
+ /** The Voyage model name, if using developer's own account. */
11346
+ modelName?: Maybe<Scalars['String']['output']>;
11347
+ };
11348
+ /** Represents Voyage model properties. */
11349
+ export type VoyageModelPropertiesInput = {
11350
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
11351
+ chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
11352
+ /** The Voyage API key, if using developer's own account. */
11353
+ key?: InputMaybe<Scalars['String']['input']>;
11354
+ /** The Voyage model, or custom, when using developer's own account. */
11355
+ model: VoyageModels;
11356
+ /** The Voyage model name, if using developer's own account. */
11357
+ modelName?: InputMaybe<Scalars['String']['input']>;
11358
+ };
11359
+ /** Represents Voyage model properties. */
11360
+ export type VoyageModelPropertiesUpdateInput = {
11361
+ /** The limit of tokens per embedded text chunk, defaults to 600. */
11362
+ chunkTokenLimit?: InputMaybe<Scalars['Int']['input']>;
11363
+ /** The Voyage API key, if using developer's own account. */
11364
+ key?: InputMaybe<Scalars['String']['input']>;
11365
+ /** The Voyage model, or custom, when using developer's own account. */
11366
+ model?: InputMaybe<VoyageModels>;
11367
+ /** The Voyage model name, if using developer's own account. */
11368
+ modelName?: InputMaybe<Scalars['String']['input']>;
11369
+ };
11370
+ /** Voyage model type */
11371
+ export declare enum VoyageModels {
11372
+ /** Developer-specified model */
11373
+ Custom = "CUSTOM",
11374
+ /** Voyage (Latest) */
11375
+ Voyage = "VOYAGE",
11376
+ /** Voyage 3.0 */
11377
+ Voyage_3_0 = "VOYAGE_3_0",
11378
+ /** Voyage Code 2.0 */
11379
+ VoyageCode_2_0 = "VOYAGE_CODE_2_0",
11380
+ /** Voyage Finance 2.0 */
11381
+ VoyageFinance_2_0 = "VOYAGE_FINANCE_2_0",
11382
+ /** Voyage Law 2.0 */
11383
+ VoyageLaw_2_0 = "VOYAGE_LAW_2_0",
11384
+ /** Voyage Lite 3.0 */
11385
+ VoyageLite_3_0 = "VOYAGE_LITE_3_0",
11386
+ /** Voyage Multilingual 2.0 */
11387
+ VoyageMultilingual_2_0 = "VOYAGE_MULTILINGUAL_2_0"
11388
+ }
11227
11389
  /** Represents web feed properties. */
11228
11390
  export type WebFeedProperties = {
11229
11391
  __typename?: 'WebFeedProperties';
@@ -11293,7 +11455,7 @@ export type Workflow = {
11293
11455
  relevance?: Maybe<Scalars['Float']['output']>;
11294
11456
  /** The state of the workflow (i.e. created, finished). */
11295
11457
  state: EntityState;
11296
- /** The storage stage of the content workflow. */
11458
+ /** @deprecated The storage stage of the content workflow has been removed. Embeddings can be configured through the project embeddings strategy. */
11297
11459
  storage?: Maybe<StorageWorkflowStage>;
11298
11460
  };
11299
11461
  /** Represents the workflow action. */
@@ -11344,8 +11506,6 @@ export type WorkflowInput = {
11344
11506
  name: Scalars['String']['input'];
11345
11507
  /** The preparation stage of the content workflow. */
11346
11508
  preparation?: InputMaybe<PreparationWorkflowStageInput>;
11347
- /** The storage stage of the content workflow. */
11348
- storage?: InputMaybe<StorageWorkflowStageInput>;
11349
11509
  };
11350
11510
  /** Represents workflow query results. */
11351
11511
  export type WorkflowResults = {
@@ -11371,8 +11531,6 @@ export type WorkflowUpdateInput = {
11371
11531
  name?: InputMaybe<Scalars['String']['input']>;
11372
11532
  /** The preparation stage of the content workflow. */
11373
11533
  preparation?: InputMaybe<PreparationWorkflowStageInput>;
11374
- /** The storage stage of the content workflow. */
11375
- storage?: InputMaybe<StorageWorkflowStageInput>;
11376
11534
  };
11377
11535
  /** Represents YouTube feed properties. */
11378
11536
  export type YouTubeFeedProperties = {
@@ -12490,6 +12648,31 @@ export type IngestTextMutation = {
12490
12648
  } | null> | null;
12491
12649
  } | null;
12492
12650
  };
12651
+ export type IngestTextBatchMutationVariables = Exact<{
12652
+ batch: Array<TextContentInput> | TextContentInput;
12653
+ textType?: InputMaybe<TextTypes>;
12654
+ workflow?: InputMaybe<EntityReferenceInput>;
12655
+ collections?: InputMaybe<Array<EntityReferenceInput> | EntityReferenceInput>;
12656
+ correlationId?: InputMaybe<Scalars['String']['input']>;
12657
+ }>;
12658
+ export type IngestTextBatchMutation = {
12659
+ __typename?: 'Mutation';
12660
+ ingestTextBatch?: Array<{
12661
+ __typename?: 'Content';
12662
+ id: string;
12663
+ name: string;
12664
+ state: EntityState;
12665
+ type?: ContentTypes | null;
12666
+ fileType?: FileTypes | null;
12667
+ mimeType?: string | null;
12668
+ uri?: any | null;
12669
+ collections?: Array<{
12670
+ __typename?: 'Collection';
12671
+ id: string;
12672
+ name: string;
12673
+ } | null> | null;
12674
+ } | null> | null;
12675
+ };
12493
12676
  export type IngestUriMutationVariables = Exact<{
12494
12677
  name?: InputMaybe<Scalars['String']['input']>;
12495
12678
  uri: Scalars['URL']['input'];
@@ -16936,6 +17119,17 @@ export type GetProjectQuery = {
16936
17119
  id: string;
16937
17120
  name: string;
16938
17121
  } | null;
17122
+ embeddings?: {
17123
+ __typename?: 'EmbeddingsStrategy';
17124
+ textSpecification?: {
17125
+ __typename?: 'EntityReference';
17126
+ id: string;
17127
+ } | null;
17128
+ imageSpecification?: {
17129
+ __typename?: 'EntityReference';
17130
+ id: string;
17131
+ } | null;
17132
+ } | null;
16939
17133
  quota?: {
16940
17134
  __typename?: 'ProjectQuota';
16941
17135
  storage?: any | null;
@@ -17421,6 +17615,7 @@ export type GetSpecificationQuery = {
17421
17615
  endpoint: any;
17422
17616
  temperature?: number | null;
17423
17617
  probability?: number | null;
17618
+ chunkTokenLimit?: number | null;
17424
17619
  } | null;
17425
17620
  openAI?: {
17426
17621
  __typename?: 'OpenAIModelProperties';
@@ -17431,6 +17626,7 @@ export type GetSpecificationQuery = {
17431
17626
  modelName?: string | null;
17432
17627
  temperature?: number | null;
17433
17628
  probability?: number | null;
17629
+ chunkTokenLimit?: number | null;
17434
17630
  } | null;
17435
17631
  azureOpenAI?: {
17436
17632
  __typename?: 'AzureOpenAIModelProperties';
@@ -17442,6 +17638,7 @@ export type GetSpecificationQuery = {
17442
17638
  deploymentName?: string | null;
17443
17639
  temperature?: number | null;
17444
17640
  probability?: number | null;
17641
+ chunkTokenLimit?: number | null;
17445
17642
  } | null;
17446
17643
  cohere?: {
17447
17644
  __typename?: 'CohereModelProperties';
@@ -17452,6 +17649,7 @@ export type GetSpecificationQuery = {
17452
17649
  modelName?: string | null;
17453
17650
  temperature?: number | null;
17454
17651
  probability?: number | null;
17652
+ chunkTokenLimit?: number | null;
17455
17653
  } | null;
17456
17654
  anthropic?: {
17457
17655
  __typename?: 'AnthropicModelProperties';
@@ -17472,6 +17670,7 @@ export type GetSpecificationQuery = {
17472
17670
  modelName?: string | null;
17473
17671
  temperature?: number | null;
17474
17672
  probability?: number | null;
17673
+ chunkTokenLimit?: number | null;
17475
17674
  } | null;
17476
17675
  replicate?: {
17477
17676
  __typename?: 'ReplicateModelProperties';
@@ -17493,6 +17692,7 @@ export type GetSpecificationQuery = {
17493
17692
  endpoint?: any | null;
17494
17693
  temperature?: number | null;
17495
17694
  probability?: number | null;
17695
+ chunkTokenLimit?: number | null;
17496
17696
  } | null;
17497
17697
  groq?: {
17498
17698
  __typename?: 'GroqModelProperties';
@@ -17526,12 +17726,20 @@ export type GetSpecificationQuery = {
17526
17726
  temperature?: number | null;
17527
17727
  probability?: number | null;
17528
17728
  } | null;
17529
- tools?: Array<{
17530
- __typename?: 'ToolDefinition';
17531
- name: string;
17532
- description?: string | null;
17533
- schema: string;
17534
- }> | null;
17729
+ jina?: {
17730
+ __typename?: 'JinaModelProperties';
17731
+ model: JinaModels;
17732
+ key?: string | null;
17733
+ modelName?: string | null;
17734
+ chunkTokenLimit?: number | null;
17735
+ } | null;
17736
+ voyage?: {
17737
+ __typename?: 'VoyageModelProperties';
17738
+ model: VoyageModels;
17739
+ key?: string | null;
17740
+ modelName?: string | null;
17741
+ chunkTokenLimit?: number | null;
17742
+ } | null;
17535
17743
  } | null;
17536
17744
  };
17537
17745
  export type PromptSpecificationsMutationVariables = Exact<{
@@ -17746,6 +17954,7 @@ export type QuerySpecificationsQuery = {
17746
17954
  endpoint: any;
17747
17955
  temperature?: number | null;
17748
17956
  probability?: number | null;
17957
+ chunkTokenLimit?: number | null;
17749
17958
  } | null;
17750
17959
  openAI?: {
17751
17960
  __typename?: 'OpenAIModelProperties';
@@ -17756,6 +17965,7 @@ export type QuerySpecificationsQuery = {
17756
17965
  modelName?: string | null;
17757
17966
  temperature?: number | null;
17758
17967
  probability?: number | null;
17968
+ chunkTokenLimit?: number | null;
17759
17969
  } | null;
17760
17970
  azureOpenAI?: {
17761
17971
  __typename?: 'AzureOpenAIModelProperties';
@@ -17767,6 +17977,7 @@ export type QuerySpecificationsQuery = {
17767
17977
  deploymentName?: string | null;
17768
17978
  temperature?: number | null;
17769
17979
  probability?: number | null;
17980
+ chunkTokenLimit?: number | null;
17770
17981
  } | null;
17771
17982
  cohere?: {
17772
17983
  __typename?: 'CohereModelProperties';
@@ -17777,6 +17988,7 @@ export type QuerySpecificationsQuery = {
17777
17988
  modelName?: string | null;
17778
17989
  temperature?: number | null;
17779
17990
  probability?: number | null;
17991
+ chunkTokenLimit?: number | null;
17780
17992
  } | null;
17781
17993
  anthropic?: {
17782
17994
  __typename?: 'AnthropicModelProperties';
@@ -17797,6 +18009,7 @@ export type QuerySpecificationsQuery = {
17797
18009
  modelName?: string | null;
17798
18010
  temperature?: number | null;
17799
18011
  probability?: number | null;
18012
+ chunkTokenLimit?: number | null;
17800
18013
  } | null;
17801
18014
  replicate?: {
17802
18015
  __typename?: 'ReplicateModelProperties';
@@ -17818,6 +18031,7 @@ export type QuerySpecificationsQuery = {
17818
18031
  endpoint?: any | null;
17819
18032
  temperature?: number | null;
17820
18033
  probability?: number | null;
18034
+ chunkTokenLimit?: number | null;
17821
18035
  } | null;
17822
18036
  groq?: {
17823
18037
  __typename?: 'GroqModelProperties';
@@ -17851,12 +18065,20 @@ export type QuerySpecificationsQuery = {
17851
18065
  temperature?: number | null;
17852
18066
  probability?: number | null;
17853
18067
  } | null;
17854
- tools?: Array<{
17855
- __typename?: 'ToolDefinition';
17856
- name: string;
17857
- description?: string | null;
17858
- schema: string;
17859
- }> | null;
18068
+ jina?: {
18069
+ __typename?: 'JinaModelProperties';
18070
+ model: JinaModels;
18071
+ key?: string | null;
18072
+ modelName?: string | null;
18073
+ chunkTokenLimit?: number | null;
18074
+ } | null;
18075
+ voyage?: {
18076
+ __typename?: 'VoyageModelProperties';
18077
+ model: VoyageModels;
18078
+ key?: string | null;
18079
+ modelName?: string | null;
18080
+ chunkTokenLimit?: number | null;
18081
+ } | null;
17860
18082
  } | null> | null;
17861
18083
  } | null;
17862
18084
  };
@@ -17992,12 +18214,6 @@ export type CreateWorkflowMutation = {
17992
18214
  __typename?: 'AzureImageExtractionProperties';
17993
18215
  confidenceThreshold?: number | null;
17994
18216
  } | null;
17995
- openAIImage?: {
17996
- __typename?: 'OpenAIImageExtractionProperties';
17997
- confidenceThreshold?: number | null;
17998
- detailLevel?: OpenAiVisionDetailLevels | null;
17999
- customInstructions?: string | null;
18000
- } | null;
18001
18217
  modelImage?: {
18002
18218
  __typename?: 'ModelImageExtractionProperties';
18003
18219
  specification?: {
@@ -18044,13 +18260,6 @@ export type CreateWorkflowMutation = {
18044
18260
  } | null;
18045
18261
  } | null> | null;
18046
18262
  } | null;
18047
- storage?: {
18048
- __typename?: 'StorageWorkflowStage';
18049
- embeddings?: {
18050
- __typename?: 'EmbeddingsStrategy';
18051
- chunkTokenLimit?: number | null;
18052
- } | null;
18053
- } | null;
18054
18263
  actions?: Array<{
18055
18264
  __typename?: 'WorkflowAction';
18056
18265
  connector?: {
@@ -18216,12 +18425,6 @@ export type GetWorkflowQuery = {
18216
18425
  __typename?: 'AzureImageExtractionProperties';
18217
18426
  confidenceThreshold?: number | null;
18218
18427
  } | null;
18219
- openAIImage?: {
18220
- __typename?: 'OpenAIImageExtractionProperties';
18221
- confidenceThreshold?: number | null;
18222
- detailLevel?: OpenAiVisionDetailLevels | null;
18223
- customInstructions?: string | null;
18224
- } | null;
18225
18428
  modelImage?: {
18226
18429
  __typename?: 'ModelImageExtractionProperties';
18227
18430
  specification?: {
@@ -18268,13 +18471,6 @@ export type GetWorkflowQuery = {
18268
18471
  } | null;
18269
18472
  } | null> | null;
18270
18473
  } | null;
18271
- storage?: {
18272
- __typename?: 'StorageWorkflowStage';
18273
- embeddings?: {
18274
- __typename?: 'EmbeddingsStrategy';
18275
- chunkTokenLimit?: number | null;
18276
- } | null;
18277
- } | null;
18278
18474
  actions?: Array<{
18279
18475
  __typename?: 'WorkflowAction';
18280
18476
  connector?: {
@@ -18406,12 +18602,6 @@ export type QueryWorkflowsQuery = {
18406
18602
  __typename?: 'AzureImageExtractionProperties';
18407
18603
  confidenceThreshold?: number | null;
18408
18604
  } | null;
18409
- openAIImage?: {
18410
- __typename?: 'OpenAIImageExtractionProperties';
18411
- confidenceThreshold?: number | null;
18412
- detailLevel?: OpenAiVisionDetailLevels | null;
18413
- customInstructions?: string | null;
18414
- } | null;
18415
18605
  modelImage?: {
18416
18606
  __typename?: 'ModelImageExtractionProperties';
18417
18607
  specification?: {
@@ -18458,13 +18648,6 @@ export type QueryWorkflowsQuery = {
18458
18648
  } | null;
18459
18649
  } | null> | null;
18460
18650
  } | null;
18461
- storage?: {
18462
- __typename?: 'StorageWorkflowStage';
18463
- embeddings?: {
18464
- __typename?: 'EmbeddingsStrategy';
18465
- chunkTokenLimit?: number | null;
18466
- } | null;
18467
- } | null;
18468
18651
  actions?: Array<{
18469
18652
  __typename?: 'WorkflowAction';
18470
18653
  connector?: {
@@ -18589,12 +18772,6 @@ export type UpdateWorkflowMutation = {
18589
18772
  __typename?: 'AzureImageExtractionProperties';
18590
18773
  confidenceThreshold?: number | null;
18591
18774
  } | null;
18592
- openAIImage?: {
18593
- __typename?: 'OpenAIImageExtractionProperties';
18594
- confidenceThreshold?: number | null;
18595
- detailLevel?: OpenAiVisionDetailLevels | null;
18596
- customInstructions?: string | null;
18597
- } | null;
18598
18775
  modelImage?: {
18599
18776
  __typename?: 'ModelImageExtractionProperties';
18600
18777
  specification?: {
@@ -18641,13 +18818,6 @@ export type UpdateWorkflowMutation = {
18641
18818
  } | null;
18642
18819
  } | null> | null;
18643
18820
  } | null;
18644
- storage?: {
18645
- __typename?: 'StorageWorkflowStage';
18646
- embeddings?: {
18647
- __typename?: 'EmbeddingsStrategy';
18648
- chunkTokenLimit?: number | null;
18649
- } | null;
18650
- } | null;
18651
18821
  actions?: Array<{
18652
18822
  __typename?: 'WorkflowAction';
18653
18823
  connector?: {