@azure/search-documents 12.0.0 → 12.1.0-beta.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +52 -32
- package/dist/index.js +13767 -11641
- package/dist/index.js.map +1 -1
- package/dist-esm/src/constants.js +2 -1
- package/dist-esm/src/constants.js.map +1 -1
- package/dist-esm/src/errorModels.js +4 -0
- package/dist-esm/src/errorModels.js.map +1 -0
- package/dist-esm/src/generated/data/models/index.js +191 -7
- package/dist-esm/src/generated/data/models/index.js.map +1 -1
- package/dist-esm/src/generated/data/models/mappers.js +534 -294
- package/dist-esm/src/generated/data/models/mappers.js.map +1 -1
- package/dist-esm/src/generated/data/models/parameters.js +254 -203
- package/dist-esm/src/generated/data/models/parameters.js.map +1 -1
- package/dist-esm/src/generated/data/operations/documents.js +50 -45
- package/dist-esm/src/generated/data/operations/documents.js.map +1 -1
- package/dist-esm/src/generated/data/operationsInterfaces/documents.js.map +1 -1
- package/dist-esm/src/generated/data/searchClient.js +4 -4
- package/dist-esm/src/generated/data/searchClient.js.map +1 -1
- package/dist-esm/src/generated/service/models/index.js +176 -84
- package/dist-esm/src/generated/service/models/index.js.map +1 -1
- package/dist-esm/src/generated/service/models/mappers.js +2370 -1437
- package/dist-esm/src/generated/service/models/mappers.js.map +1 -1
- package/dist-esm/src/generated/service/models/parameters.js +99 -49
- package/dist-esm/src/generated/service/models/parameters.js.map +1 -1
- package/dist-esm/src/generated/service/operations/aliases.js +160 -0
- package/dist-esm/src/generated/service/operations/aliases.js.map +1 -0
- package/dist-esm/src/generated/service/operations/dataSources.js +26 -23
- package/dist-esm/src/generated/service/operations/dataSources.js.map +1 -1
- package/dist-esm/src/generated/service/operations/index.js +1 -0
- package/dist-esm/src/generated/service/operations/index.js.map +1 -1
- package/dist-esm/src/generated/service/operations/indexers.js +61 -33
- package/dist-esm/src/generated/service/operations/indexers.js.map +1 -1
- package/dist-esm/src/generated/service/operations/indexes.js +30 -30
- package/dist-esm/src/generated/service/operations/indexes.js.map +1 -1
- package/dist-esm/src/generated/service/operations/skillsets.js +52 -23
- package/dist-esm/src/generated/service/operations/skillsets.js.map +1 -1
- package/dist-esm/src/generated/service/operations/synonymMaps.js +22 -22
- package/dist-esm/src/generated/service/operations/synonymMaps.js.map +1 -1
- package/dist-esm/src/generated/service/operationsInterfaces/aliases.js +9 -0
- package/dist-esm/src/generated/service/operationsInterfaces/aliases.js.map +1 -0
- package/dist-esm/src/generated/service/operationsInterfaces/dataSources.js.map +1 -1
- package/dist-esm/src/generated/service/operationsInterfaces/index.js +1 -0
- package/dist-esm/src/generated/service/operationsInterfaces/index.js.map +1 -1
- package/dist-esm/src/generated/service/operationsInterfaces/indexers.js.map +1 -1
- package/dist-esm/src/generated/service/operationsInterfaces/indexes.js.map +1 -1
- package/dist-esm/src/generated/service/operationsInterfaces/skillsets.js.map +1 -1
- package/dist-esm/src/generated/service/operationsInterfaces/synonymMaps.js.map +1 -1
- package/dist-esm/src/generated/service/searchServiceClient.js +10 -9
- package/dist-esm/src/generated/service/searchServiceClient.js.map +1 -1
- package/dist-esm/src/generatedStringLiteralUnions.js +4 -0
- package/dist-esm/src/generatedStringLiteralUnions.js.map +1 -0
- package/dist-esm/src/index.js +9 -9
- package/dist-esm/src/index.js.map +1 -1
- package/dist-esm/src/indexDocumentsBatch.js.map +1 -1
- package/dist-esm/src/indexModels.js.map +1 -1
- package/dist-esm/src/odata.js.map +1 -1
- package/dist-esm/src/odataMetadataPolicy.js.map +1 -1
- package/dist-esm/src/searchApiKeyCredentialPolicy.js.map +1 -1
- package/dist-esm/src/searchClient.js +23 -17
- package/dist-esm/src/searchClient.js.map +1 -1
- package/dist-esm/src/searchIndexClient.js +155 -6
- package/dist-esm/src/searchIndexClient.js.map +1 -1
- package/dist-esm/src/searchIndexerClient.js +53 -4
- package/dist-esm/src/searchIndexerClient.js.map +1 -1
- package/dist-esm/src/searchIndexingBufferedSender.js +3 -8
- package/dist-esm/src/searchIndexingBufferedSender.js.map +1 -1
- package/dist-esm/src/serialization.js.map +1 -1
- package/dist-esm/src/serviceModels.js.map +1 -1
- package/dist-esm/src/serviceUtils.js +71 -26
- package/dist-esm/src/serviceUtils.js.map +1 -1
- package/dist-esm/src/synonymMapHelper.browser.js.map +1 -1
- package/dist-esm/src/synonymMapHelper.js +1 -1
- package/dist-esm/src/synonymMapHelper.js.map +1 -1
- package/dist-esm/src/tracing.js +1 -1
- package/dist-esm/src/tracing.js.map +1 -1
- package/package.json +39 -40
- package/types/search-documents.d.ts +1655 -152
|
@@ -5,12 +5,12 @@
|
|
|
5
5
|
* Code generated by Microsoft (R) AutoRest Code Generator.
|
|
6
6
|
* Changes may cause incorrect behavior and will be lost if the code is regenerated.
|
|
7
7
|
*/
|
|
8
|
-
/** Known values of {@link
|
|
9
|
-
export var
|
|
10
|
-
(function (
|
|
11
|
-
/** Api Version '
|
|
12
|
-
|
|
13
|
-
})(
|
|
8
|
+
/** Known values of {@link ApiVersion20240301Preview} that the service accepts. */
|
|
9
|
+
export var KnownApiVersion20240301Preview;
|
|
10
|
+
(function (KnownApiVersion20240301Preview) {
|
|
11
|
+
/** Api Version '2024-03-01-Preview' */
|
|
12
|
+
KnownApiVersion20240301Preview["TwoThousandTwentyFour0301Preview"] = "2024-03-01-Preview";
|
|
13
|
+
})(KnownApiVersion20240301Preview || (KnownApiVersion20240301Preview = {}));
|
|
14
14
|
/** Known values of {@link SearchIndexerDataSourceType} that the service accepts. */
|
|
15
15
|
export var KnownSearchIndexerDataSourceType;
|
|
16
16
|
(function (KnownSearchIndexerDataSourceType) {
|
|
@@ -79,6 +79,28 @@ export var KnownIndexerExecutionEnvironment;
|
|
|
79
79
|
/** Indicates that the indexer should run with the environment provisioned specifically for the search service. This should only be specified as the execution environment if the indexer needs to access resources securely over shared private link resources. */
|
|
80
80
|
KnownIndexerExecutionEnvironment["Private"] = "private";
|
|
81
81
|
})(KnownIndexerExecutionEnvironment || (KnownIndexerExecutionEnvironment = {}));
|
|
82
|
+
/** Known values of {@link IndexerExecutionStatusDetail} that the service accepts. */
|
|
83
|
+
export var KnownIndexerExecutionStatusDetail;
|
|
84
|
+
(function (KnownIndexerExecutionStatusDetail) {
|
|
85
|
+
/** Indicates that the reset that occurred was for a call to ResetDocs. */
|
|
86
|
+
KnownIndexerExecutionStatusDetail["ResetDocs"] = "resetDocs";
|
|
87
|
+
})(KnownIndexerExecutionStatusDetail || (KnownIndexerExecutionStatusDetail = {}));
|
|
88
|
+
/** Known values of {@link IndexingMode} that the service accepts. */
|
|
89
|
+
export var KnownIndexingMode;
|
|
90
|
+
(function (KnownIndexingMode) {
|
|
91
|
+
/** The indexer is indexing all documents in the datasource. */
|
|
92
|
+
KnownIndexingMode["IndexingAllDocs"] = "indexingAllDocs";
|
|
93
|
+
/** The indexer is indexing selective, reset documents in the datasource. The documents being indexed are defined on indexer status. */
|
|
94
|
+
KnownIndexingMode["IndexingResetDocs"] = "indexingResetDocs";
|
|
95
|
+
})(KnownIndexingMode || (KnownIndexingMode = {}));
|
|
96
|
+
/** Known values of {@link IndexProjectionMode} that the service accepts. */
|
|
97
|
+
export var KnownIndexProjectionMode;
|
|
98
|
+
(function (KnownIndexProjectionMode) {
|
|
99
|
+
/** The source document will be skipped from writing into the indexer's target index. */
|
|
100
|
+
KnownIndexProjectionMode["SkipIndexingParentDocuments"] = "skipIndexingParentDocuments";
|
|
101
|
+
/** The source document will be written into the indexer's target index. This is the default pattern. */
|
|
102
|
+
KnownIndexProjectionMode["IncludeIndexingParentDocuments"] = "includeIndexingParentDocuments";
|
|
103
|
+
})(KnownIndexProjectionMode || (KnownIndexProjectionMode = {}));
|
|
82
104
|
/** Known values of {@link SearchFieldDataType} that the service accepts. */
|
|
83
105
|
export var KnownSearchFieldDataType;
|
|
84
106
|
(function (KnownSearchFieldDataType) {
|
|
@@ -100,6 +122,12 @@ export var KnownSearchFieldDataType;
|
|
|
100
122
|
KnownSearchFieldDataType["Complex"] = "Edm.ComplexType";
|
|
101
123
|
/** Indicates that a field contains a single-precision floating point number. This is only valid when used with Collection(Edm.Single). */
|
|
102
124
|
KnownSearchFieldDataType["Single"] = "Edm.Single";
|
|
125
|
+
/** Indicates that a field contains a half-precision floating point number. This is only valid when used with Collection(Edm.Half). */
|
|
126
|
+
KnownSearchFieldDataType["Half"] = "Edm.Half";
|
|
127
|
+
/** Indicates that a field contains a 16-bit signed integer. This is only valid when used with Collection(Edm.Int16). */
|
|
128
|
+
KnownSearchFieldDataType["Int16"] = "Edm.Int16";
|
|
129
|
+
/** Indicates that a field contains a 8-bit signed integer. This is only valid when used with Collection(Edm.SByte). */
|
|
130
|
+
KnownSearchFieldDataType["SByte"] = "Edm.SByte";
|
|
103
131
|
})(KnownSearchFieldDataType || (KnownSearchFieldDataType = {}));
|
|
104
132
|
/** Known values of {@link LexicalAnalyzerName} that the service accepts. */
|
|
105
133
|
export var KnownLexicalAnalyzerName;
|
|
@@ -291,6 +319,20 @@ export var KnownLexicalAnalyzerName;
|
|
|
291
319
|
/** An analyzer that uses the whitespace tokenizer. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/core\/WhitespaceAnalyzer.html */
|
|
292
320
|
KnownLexicalAnalyzerName["Whitespace"] = "whitespace";
|
|
293
321
|
})(KnownLexicalAnalyzerName || (KnownLexicalAnalyzerName = {}));
|
|
322
|
+
/** Known values of {@link LexicalNormalizerName} that the service accepts. */
|
|
323
|
+
export var KnownLexicalNormalizerName;
|
|
324
|
+
(function (KnownLexicalNormalizerName) {
|
|
325
|
+
/** Converts alphabetic, numeric, and symbolic Unicode characters which are not in the first 127 ASCII characters (the "Basic Latin" Unicode block) into their ASCII equivalents, if such equivalents exist. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/ASCIIFoldingFilter.html */
|
|
326
|
+
KnownLexicalNormalizerName["AsciiFolding"] = "asciifolding";
|
|
327
|
+
/** Removes elisions. For example, "l'avion" (the plane) will be converted to "avion" (plane). See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/util\/ElisionFilter.html */
|
|
328
|
+
KnownLexicalNormalizerName["Elision"] = "elision";
|
|
329
|
+
/** Normalizes token text to lowercase. See https:\//lucene.apache.org\/core\/6_6_1\/analyzers-common\/org\/apache\/lucene\/analysis\/core\/LowerCaseFilter.html */
|
|
330
|
+
KnownLexicalNormalizerName["Lowercase"] = "lowercase";
|
|
331
|
+
/** Standard normalizer, which consists of lowercase and asciifolding. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/reverse\/ReverseStringFilter.html */
|
|
332
|
+
KnownLexicalNormalizerName["Standard"] = "standard";
|
|
333
|
+
/** Normalizes token text to uppercase. See https:\//lucene.apache.org\/core\/6_6_1\/analyzers-common\/org\/apache\/lucene\/analysis\/core\/UpperCaseFilter.html */
|
|
334
|
+
KnownLexicalNormalizerName["Uppercase"] = "uppercase";
|
|
335
|
+
})(KnownLexicalNormalizerName || (KnownLexicalNormalizerName = {}));
|
|
294
336
|
/** Known values of {@link VectorSearchAlgorithmKind} that the service accepts. */
|
|
295
337
|
export var KnownVectorSearchAlgorithmKind;
|
|
296
338
|
(function (KnownVectorSearchAlgorithmKind) {
|
|
@@ -299,6 +341,98 @@ export var KnownVectorSearchAlgorithmKind;
|
|
|
299
341
|
/** Exhaustive KNN algorithm which will perform brute-force search. */
|
|
300
342
|
KnownVectorSearchAlgorithmKind["ExhaustiveKnn"] = "exhaustiveKnn";
|
|
301
343
|
})(KnownVectorSearchAlgorithmKind || (KnownVectorSearchAlgorithmKind = {}));
|
|
344
|
+
/** Known values of {@link VectorSearchVectorizerKind} that the service accepts. */
|
|
345
|
+
export var KnownVectorSearchVectorizerKind;
|
|
346
|
+
(function (KnownVectorSearchVectorizerKind) {
|
|
347
|
+
/** Generate embeddings using an Azure OpenAI resource at query time. */
|
|
348
|
+
KnownVectorSearchVectorizerKind["AzureOpenAI"] = "azureOpenAI";
|
|
349
|
+
/** Generate embeddings using a custom web endpoint at query time. */
|
|
350
|
+
KnownVectorSearchVectorizerKind["CustomWebApi"] = "customWebApi";
|
|
351
|
+
})(KnownVectorSearchVectorizerKind || (KnownVectorSearchVectorizerKind = {}));
|
|
352
|
+
/** Known values of {@link VectorSearchCompressionKind} that the service accepts. */
|
|
353
|
+
export var KnownVectorSearchCompressionKind;
|
|
354
|
+
(function (KnownVectorSearchCompressionKind) {
|
|
355
|
+
/** Scalar Quantization, a type of compression method. In scalar quantization, the original vectors values are compressed to a narrower type by discretizing and representing each component of a vector using a reduced set of quantized values, thereby reducing the overall data size. */
|
|
356
|
+
KnownVectorSearchCompressionKind["ScalarQuantization"] = "scalarQuantization";
|
|
357
|
+
})(KnownVectorSearchCompressionKind || (KnownVectorSearchCompressionKind = {}));
|
|
358
|
+
/** Known values of {@link TokenFilterName} that the service accepts. */
|
|
359
|
+
export var KnownTokenFilterName;
|
|
360
|
+
(function (KnownTokenFilterName) {
|
|
361
|
+
/** A token filter that applies the Arabic normalizer to normalize the orthography. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/ar\/ArabicNormalizationFilter.html */
|
|
362
|
+
KnownTokenFilterName["ArabicNormalization"] = "arabic_normalization";
|
|
363
|
+
/** Strips all characters after an apostrophe (including the apostrophe itself). See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/tr\/ApostropheFilter.html */
|
|
364
|
+
KnownTokenFilterName["Apostrophe"] = "apostrophe";
|
|
365
|
+
/** Converts alphabetic, numeric, and symbolic Unicode characters which are not in the first 127 ASCII characters (the "Basic Latin" Unicode block) into their ASCII equivalents, if such equivalents exist. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/ASCIIFoldingFilter.html */
|
|
366
|
+
KnownTokenFilterName["AsciiFolding"] = "asciifolding";
|
|
367
|
+
/** Forms bigrams of CJK terms that are generated from the standard tokenizer. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/cjk\/CJKBigramFilter.html */
|
|
368
|
+
KnownTokenFilterName["CjkBigram"] = "cjk_bigram";
|
|
369
|
+
/** Normalizes CJK width differences. Folds fullwidth ASCII variants into the equivalent basic Latin, and half-width Katakana variants into the equivalent Kana. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/cjk\/CJKWidthFilter.html */
|
|
370
|
+
KnownTokenFilterName["CjkWidth"] = "cjk_width";
|
|
371
|
+
/** Removes English possessives, and dots from acronyms. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/standard\/ClassicFilter.html */
|
|
372
|
+
KnownTokenFilterName["Classic"] = "classic";
|
|
373
|
+
/** Construct bigrams for frequently occurring terms while indexing. Single terms are still indexed too, with bigrams overlaid. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/commongrams\/CommonGramsFilter.html */
|
|
374
|
+
KnownTokenFilterName["CommonGram"] = "common_grams";
|
|
375
|
+
/** Generates n-grams of the given size(s) starting from the front or the back of an input token. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/ngram\/EdgeNGramTokenFilter.html */
|
|
376
|
+
KnownTokenFilterName["EdgeNGram"] = "edgeNGram_v2";
|
|
377
|
+
/** Removes elisions. For example, "l'avion" (the plane) will be converted to "avion" (plane). See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/util\/ElisionFilter.html */
|
|
378
|
+
KnownTokenFilterName["Elision"] = "elision";
|
|
379
|
+
/** Normalizes German characters according to the heuristics of the German2 snowball algorithm. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/de\/GermanNormalizationFilter.html */
|
|
380
|
+
KnownTokenFilterName["GermanNormalization"] = "german_normalization";
|
|
381
|
+
/** Normalizes text in Hindi to remove some differences in spelling variations. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/hi\/HindiNormalizationFilter.html */
|
|
382
|
+
KnownTokenFilterName["HindiNormalization"] = "hindi_normalization";
|
|
383
|
+
/** Normalizes the Unicode representation of text in Indian languages. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/in\/IndicNormalizationFilter.html */
|
|
384
|
+
KnownTokenFilterName["IndicNormalization"] = "indic_normalization";
|
|
385
|
+
/** Emits each incoming token twice, once as keyword and once as non-keyword. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/KeywordRepeatFilter.html */
|
|
386
|
+
KnownTokenFilterName["KeywordRepeat"] = "keyword_repeat";
|
|
387
|
+
/** A high-performance kstem filter for English. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/en\/KStemFilter.html */
|
|
388
|
+
KnownTokenFilterName["KStem"] = "kstem";
|
|
389
|
+
/** Removes words that are too long or too short. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/LengthFilter.html */
|
|
390
|
+
KnownTokenFilterName["Length"] = "length";
|
|
391
|
+
/** Limits the number of tokens while indexing. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/LimitTokenCountFilter.html */
|
|
392
|
+
KnownTokenFilterName["Limit"] = "limit";
|
|
393
|
+
/** Normalizes token text to lower case. See https:\//lucene.apache.org\/core\/6_6_1\/analyzers-common\/org\/apache\/lucene\/analysis\/core\/LowerCaseFilter.html */
|
|
394
|
+
KnownTokenFilterName["Lowercase"] = "lowercase";
|
|
395
|
+
/** Generates n-grams of the given size(s). See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/ngram\/NGramTokenFilter.html */
|
|
396
|
+
KnownTokenFilterName["NGram"] = "nGram_v2";
|
|
397
|
+
/** Applies normalization for Persian. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/fa\/PersianNormalizationFilter.html */
|
|
398
|
+
KnownTokenFilterName["PersianNormalization"] = "persian_normalization";
|
|
399
|
+
/** Create tokens for phonetic matches. See https:\//lucene.apache.org\/core\/4_10_3\/analyzers-phonetic\/org\/apache\/lucene\/analysis\/phonetic\/package-tree.html */
|
|
400
|
+
KnownTokenFilterName["Phonetic"] = "phonetic";
|
|
401
|
+
/** Uses the Porter stemming algorithm to transform the token stream. See http:\//tartarus.org\/~martin\/PorterStemmer */
|
|
402
|
+
KnownTokenFilterName["PorterStem"] = "porter_stem";
|
|
403
|
+
/** Reverses the token string. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/reverse\/ReverseStringFilter.html */
|
|
404
|
+
KnownTokenFilterName["Reverse"] = "reverse";
|
|
405
|
+
/** Normalizes use of the interchangeable Scandinavian characters. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/ScandinavianNormalizationFilter.html */
|
|
406
|
+
KnownTokenFilterName["ScandinavianNormalization"] = "scandinavian_normalization";
|
|
407
|
+
/** Folds Scandinavian characters åÅäæÄÆ->a and öÖøØ->o. It also discriminates against use of double vowels aa, ae, ao, oe and oo, leaving just the first one. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/ScandinavianFoldingFilter.html */
|
|
408
|
+
KnownTokenFilterName["ScandinavianFoldingNormalization"] = "scandinavian_folding";
|
|
409
|
+
/** Creates combinations of tokens as a single token. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/shingle\/ShingleFilter.html */
|
|
410
|
+
KnownTokenFilterName["Shingle"] = "shingle";
|
|
411
|
+
/** A filter that stems words using a Snowball-generated stemmer. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/snowball\/SnowballFilter.html */
|
|
412
|
+
KnownTokenFilterName["Snowball"] = "snowball";
|
|
413
|
+
/** Normalizes the Unicode representation of Sorani text. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/ckb\/SoraniNormalizationFilter.html */
|
|
414
|
+
KnownTokenFilterName["SoraniNormalization"] = "sorani_normalization";
|
|
415
|
+
/** Language specific stemming filter. See https:\//docs.microsoft.com\/rest\/api\/searchservice\/Custom-analyzers-in-Azure-Search#TokenFilters */
|
|
416
|
+
KnownTokenFilterName["Stemmer"] = "stemmer";
|
|
417
|
+
/** Removes stop words from a token stream. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/core\/StopFilter.html */
|
|
418
|
+
KnownTokenFilterName["Stopwords"] = "stopwords";
|
|
419
|
+
/** Trims leading and trailing whitespace from tokens. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/TrimFilter.html */
|
|
420
|
+
KnownTokenFilterName["Trim"] = "trim";
|
|
421
|
+
/** Truncates the terms to a specific length. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/TruncateTokenFilter.html */
|
|
422
|
+
KnownTokenFilterName["Truncate"] = "truncate";
|
|
423
|
+
/** Filters out tokens with same text as the previous token. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/RemoveDuplicatesTokenFilter.html */
|
|
424
|
+
KnownTokenFilterName["Unique"] = "unique";
|
|
425
|
+
/** Normalizes token text to upper case. See https:\//lucene.apache.org\/core\/6_6_1\/analyzers-common\/org\/apache\/lucene\/analysis\/core\/UpperCaseFilter.html */
|
|
426
|
+
KnownTokenFilterName["Uppercase"] = "uppercase";
|
|
427
|
+
/** Splits words into subwords and performs optional transformations on subword groups. */
|
|
428
|
+
KnownTokenFilterName["WordDelimiter"] = "word_delimiter";
|
|
429
|
+
})(KnownTokenFilterName || (KnownTokenFilterName = {}));
|
|
430
|
+
/** Known values of {@link CharFilterName} that the service accepts. */
|
|
431
|
+
export var KnownCharFilterName;
|
|
432
|
+
(function (KnownCharFilterName) {
|
|
433
|
+
/** A character filter that attempts to strip out HTML constructs. See https:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/charfilter\/HTMLStripCharFilter.html */
|
|
434
|
+
KnownCharFilterName["HtmlStrip"] = "html_strip";
|
|
435
|
+
})(KnownCharFilterName || (KnownCharFilterName = {}));
|
|
302
436
|
/** Known values of {@link VectorSearchAlgorithmMetric} that the service accepts. */
|
|
303
437
|
export var KnownVectorSearchAlgorithmMetric;
|
|
304
438
|
(function (KnownVectorSearchAlgorithmMetric) {
|
|
@@ -309,6 +443,12 @@ export var KnownVectorSearchAlgorithmMetric;
|
|
|
309
443
|
/** DotProduct */
|
|
310
444
|
KnownVectorSearchAlgorithmMetric["DotProduct"] = "dotProduct";
|
|
311
445
|
})(KnownVectorSearchAlgorithmMetric || (KnownVectorSearchAlgorithmMetric = {}));
|
|
446
|
+
/** Known values of {@link VectorSearchCompressionTargetDataType} that the service accepts. */
|
|
447
|
+
export var KnownVectorSearchCompressionTargetDataType;
|
|
448
|
+
(function (KnownVectorSearchCompressionTargetDataType) {
|
|
449
|
+
/** Int8 */
|
|
450
|
+
KnownVectorSearchCompressionTargetDataType["Int8"] = "int8";
|
|
451
|
+
})(KnownVectorSearchCompressionTargetDataType || (KnownVectorSearchCompressionTargetDataType = {}));
|
|
312
452
|
/** Known values of {@link KeyPhraseExtractionSkillLanguage} that the service accepts. */
|
|
313
453
|
export var KnownKeyPhraseExtractionSkillLanguage;
|
|
314
454
|
(function (KnownKeyPhraseExtractionSkillLanguage) {
|
|
@@ -689,6 +829,18 @@ export var KnownOcrSkillLanguage;
|
|
|
689
829
|
/** Unknown (All) */
|
|
690
830
|
KnownOcrSkillLanguage["Unk"] = "unk";
|
|
691
831
|
})(KnownOcrSkillLanguage || (KnownOcrSkillLanguage = {}));
|
|
832
|
+
/** Known values of {@link LineEnding} that the service accepts. */
|
|
833
|
+
export var KnownLineEnding;
|
|
834
|
+
(function (KnownLineEnding) {
|
|
835
|
+
/** Lines are separated by a single space character. */
|
|
836
|
+
KnownLineEnding["Space"] = "space";
|
|
837
|
+
/** Lines are separated by a carriage return ('\r') character. */
|
|
838
|
+
KnownLineEnding["CarriageReturn"] = "carriageReturn";
|
|
839
|
+
/** Lines are separated by a single line feed ('\n') character. */
|
|
840
|
+
KnownLineEnding["LineFeed"] = "lineFeed";
|
|
841
|
+
/** Lines are separated by a carriage return and a line feed ('\r\n') character. */
|
|
842
|
+
KnownLineEnding["CarriageReturnLineFeed"] = "carriageReturnLineFeed";
|
|
843
|
+
})(KnownLineEnding || (KnownLineEnding = {}));
|
|
692
844
|
/** Known values of {@link ImageAnalysisSkillLanguage} that the service accepts. */
|
|
693
845
|
export var KnownImageAnalysisSkillLanguage;
|
|
694
846
|
(function (KnownImageAnalysisSkillLanguage) {
|
|
@@ -1100,6 +1252,10 @@ export var KnownTextTranslationSkillLanguage;
|
|
|
1100
1252
|
KnownTextTranslationSkillLanguage["Sw"] = "sw";
|
|
1101
1253
|
/** Klingon */
|
|
1102
1254
|
KnownTextTranslationSkillLanguage["Tlh"] = "tlh";
|
|
1255
|
+
/** Klingon (Latin script) */
|
|
1256
|
+
KnownTextTranslationSkillLanguage["TlhLatn"] = "tlh-Latn";
|
|
1257
|
+
/** Klingon (Klingon script) */
|
|
1258
|
+
KnownTextTranslationSkillLanguage["TlhPiqd"] = "tlh-Piqd";
|
|
1103
1259
|
/** Korean */
|
|
1104
1260
|
KnownTextTranslationSkillLanguage["Ko"] = "ko";
|
|
1105
1261
|
/** Latvian */
|
|
@@ -1120,6 +1276,10 @@ export var KnownTextTranslationSkillLanguage;
|
|
|
1120
1276
|
KnownTextTranslationSkillLanguage["Pl"] = "pl";
|
|
1121
1277
|
/** Portuguese */
|
|
1122
1278
|
KnownTextTranslationSkillLanguage["Pt"] = "pt";
|
|
1279
|
+
/** Portuguese (Brazil) */
|
|
1280
|
+
KnownTextTranslationSkillLanguage["PtBr"] = "pt-br";
|
|
1281
|
+
/** Portuguese (Portugal) */
|
|
1282
|
+
KnownTextTranslationSkillLanguage["PtPT"] = "pt-PT";
|
|
1123
1283
|
/** Queretaro Otomi */
|
|
1124
1284
|
KnownTextTranslationSkillLanguage["Otq"] = "otq";
|
|
1125
1285
|
/** Romanian */
|
|
@@ -1162,6 +1322,16 @@ export var KnownTextTranslationSkillLanguage;
|
|
|
1162
1322
|
KnownTextTranslationSkillLanguage["Cy"] = "cy";
|
|
1163
1323
|
/** Yucatec Maya */
|
|
1164
1324
|
KnownTextTranslationSkillLanguage["Yua"] = "yua";
|
|
1325
|
+
/** Irish */
|
|
1326
|
+
KnownTextTranslationSkillLanguage["Ga"] = "ga";
|
|
1327
|
+
/** Kannada */
|
|
1328
|
+
KnownTextTranslationSkillLanguage["Kn"] = "kn";
|
|
1329
|
+
/** Maori */
|
|
1330
|
+
KnownTextTranslationSkillLanguage["Mi"] = "mi";
|
|
1331
|
+
/** Malayalam */
|
|
1332
|
+
KnownTextTranslationSkillLanguage["Ml"] = "ml";
|
|
1333
|
+
/** Punjabi */
|
|
1334
|
+
KnownTextTranslationSkillLanguage["Pa"] = "pa";
|
|
1165
1335
|
})(KnownTextTranslationSkillLanguage || (KnownTextTranslationSkillLanguage = {}));
|
|
1166
1336
|
/** Known values of {@link LexicalTokenizerName} that the service accepts. */
|
|
1167
1337
|
export var KnownLexicalTokenizerName;
|
|
@@ -1193,84 +1363,6 @@ export var KnownLexicalTokenizerName;
|
|
|
1193
1363
|
/** Divides text at whitespace. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/core\/WhitespaceTokenizer.html */
|
|
1194
1364
|
KnownLexicalTokenizerName["Whitespace"] = "whitespace";
|
|
1195
1365
|
})(KnownLexicalTokenizerName || (KnownLexicalTokenizerName = {}));
|
|
1196
|
-
/** Known values of {@link TokenFilterName} that the service accepts. */
|
|
1197
|
-
export var KnownTokenFilterName;
|
|
1198
|
-
(function (KnownTokenFilterName) {
|
|
1199
|
-
/** A token filter that applies the Arabic normalizer to normalize the orthography. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/ar\/ArabicNormalizationFilter.html */
|
|
1200
|
-
KnownTokenFilterName["ArabicNormalization"] = "arabic_normalization";
|
|
1201
|
-
/** Strips all characters after an apostrophe (including the apostrophe itself). See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/tr\/ApostropheFilter.html */
|
|
1202
|
-
KnownTokenFilterName["Apostrophe"] = "apostrophe";
|
|
1203
|
-
/** Converts alphabetic, numeric, and symbolic Unicode characters which are not in the first 127 ASCII characters (the "Basic Latin" Unicode block) into their ASCII equivalents, if such equivalents exist. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/ASCIIFoldingFilter.html */
|
|
1204
|
-
KnownTokenFilterName["AsciiFolding"] = "asciifolding";
|
|
1205
|
-
/** Forms bigrams of CJK terms that are generated from the standard tokenizer. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/cjk\/CJKBigramFilter.html */
|
|
1206
|
-
KnownTokenFilterName["CjkBigram"] = "cjk_bigram";
|
|
1207
|
-
/** Normalizes CJK width differences. Folds fullwidth ASCII variants into the equivalent basic Latin, and half-width Katakana variants into the equivalent Kana. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/cjk\/CJKWidthFilter.html */
|
|
1208
|
-
KnownTokenFilterName["CjkWidth"] = "cjk_width";
|
|
1209
|
-
/** Removes English possessives, and dots from acronyms. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/standard\/ClassicFilter.html */
|
|
1210
|
-
KnownTokenFilterName["Classic"] = "classic";
|
|
1211
|
-
/** Construct bigrams for frequently occurring terms while indexing. Single terms are still indexed too, with bigrams overlaid. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/commongrams\/CommonGramsFilter.html */
|
|
1212
|
-
KnownTokenFilterName["CommonGram"] = "common_grams";
|
|
1213
|
-
/** Generates n-grams of the given size(s) starting from the front or the back of an input token. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/ngram\/EdgeNGramTokenFilter.html */
|
|
1214
|
-
KnownTokenFilterName["EdgeNGram"] = "edgeNGram_v2";
|
|
1215
|
-
/** Removes elisions. For example, "l'avion" (the plane) will be converted to "avion" (plane). See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/util\/ElisionFilter.html */
|
|
1216
|
-
KnownTokenFilterName["Elision"] = "elision";
|
|
1217
|
-
/** Normalizes German characters according to the heuristics of the German2 snowball algorithm. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/de\/GermanNormalizationFilter.html */
|
|
1218
|
-
KnownTokenFilterName["GermanNormalization"] = "german_normalization";
|
|
1219
|
-
/** Normalizes text in Hindi to remove some differences in spelling variations. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/hi\/HindiNormalizationFilter.html */
|
|
1220
|
-
KnownTokenFilterName["HindiNormalization"] = "hindi_normalization";
|
|
1221
|
-
/** Normalizes the Unicode representation of text in Indian languages. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/in\/IndicNormalizationFilter.html */
|
|
1222
|
-
KnownTokenFilterName["IndicNormalization"] = "indic_normalization";
|
|
1223
|
-
/** Emits each incoming token twice, once as keyword and once as non-keyword. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/KeywordRepeatFilter.html */
|
|
1224
|
-
KnownTokenFilterName["KeywordRepeat"] = "keyword_repeat";
|
|
1225
|
-
/** A high-performance kstem filter for English. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/en\/KStemFilter.html */
|
|
1226
|
-
KnownTokenFilterName["KStem"] = "kstem";
|
|
1227
|
-
/** Removes words that are too long or too short. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/LengthFilter.html */
|
|
1228
|
-
KnownTokenFilterName["Length"] = "length";
|
|
1229
|
-
/** Limits the number of tokens while indexing. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/LimitTokenCountFilter.html */
|
|
1230
|
-
KnownTokenFilterName["Limit"] = "limit";
|
|
1231
|
-
/** Normalizes token text to lower case. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/core\/LowerCaseFilter.htm */
|
|
1232
|
-
KnownTokenFilterName["Lowercase"] = "lowercase";
|
|
1233
|
-
/** Generates n-grams of the given size(s). See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/ngram\/NGramTokenFilter.html */
|
|
1234
|
-
KnownTokenFilterName["NGram"] = "nGram_v2";
|
|
1235
|
-
/** Applies normalization for Persian. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/fa\/PersianNormalizationFilter.html */
|
|
1236
|
-
KnownTokenFilterName["PersianNormalization"] = "persian_normalization";
|
|
1237
|
-
/** Create tokens for phonetic matches. See https:\//lucene.apache.org\/core\/4_10_3\/analyzers-phonetic\/org\/apache\/lucene\/analysis\/phonetic\/package-tree.html */
|
|
1238
|
-
KnownTokenFilterName["Phonetic"] = "phonetic";
|
|
1239
|
-
/** Uses the Porter stemming algorithm to transform the token stream. See http:\//tartarus.org\/~martin\/PorterStemmer */
|
|
1240
|
-
KnownTokenFilterName["PorterStem"] = "porter_stem";
|
|
1241
|
-
/** Reverses the token string. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/reverse\/ReverseStringFilter.html */
|
|
1242
|
-
KnownTokenFilterName["Reverse"] = "reverse";
|
|
1243
|
-
/** Normalizes use of the interchangeable Scandinavian characters. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/ScandinavianNormalizationFilter.html */
|
|
1244
|
-
KnownTokenFilterName["ScandinavianNormalization"] = "scandinavian_normalization";
|
|
1245
|
-
/** Folds Scandinavian characters åÅäæÄÆ->a and öÖøØ->o. It also discriminates against use of double vowels aa, ae, ao, oe and oo, leaving just the first one. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/ScandinavianFoldingFilter.html */
|
|
1246
|
-
KnownTokenFilterName["ScandinavianFoldingNormalization"] = "scandinavian_folding";
|
|
1247
|
-
/** Creates combinations of tokens as a single token. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/shingle\/ShingleFilter.html */
|
|
1248
|
-
KnownTokenFilterName["Shingle"] = "shingle";
|
|
1249
|
-
/** A filter that stems words using a Snowball-generated stemmer. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/snowball\/SnowballFilter.html */
|
|
1250
|
-
KnownTokenFilterName["Snowball"] = "snowball";
|
|
1251
|
-
/** Normalizes the Unicode representation of Sorani text. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/ckb\/SoraniNormalizationFilter.html */
|
|
1252
|
-
KnownTokenFilterName["SoraniNormalization"] = "sorani_normalization";
|
|
1253
|
-
/** Language specific stemming filter. See https:\//docs.microsoft.com\/rest\/api\/searchservice\/Custom-analyzers-in-Azure-Search#TokenFilters */
|
|
1254
|
-
KnownTokenFilterName["Stemmer"] = "stemmer";
|
|
1255
|
-
/** Removes stop words from a token stream. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/core\/StopFilter.html */
|
|
1256
|
-
KnownTokenFilterName["Stopwords"] = "stopwords";
|
|
1257
|
-
/** Trims leading and trailing whitespace from tokens. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/TrimFilter.html */
|
|
1258
|
-
KnownTokenFilterName["Trim"] = "trim";
|
|
1259
|
-
/** Truncates the terms to a specific length. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/TruncateTokenFilter.html */
|
|
1260
|
-
KnownTokenFilterName["Truncate"] = "truncate";
|
|
1261
|
-
/** Filters out tokens with same text as the previous token. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/miscellaneous\/RemoveDuplicatesTokenFilter.html */
|
|
1262
|
-
KnownTokenFilterName["Unique"] = "unique";
|
|
1263
|
-
/** Normalizes token text to upper case. See http:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/core\/UpperCaseFilter.html */
|
|
1264
|
-
KnownTokenFilterName["Uppercase"] = "uppercase";
|
|
1265
|
-
/** Splits words into subwords and performs optional transformations on subword groups. */
|
|
1266
|
-
KnownTokenFilterName["WordDelimiter"] = "word_delimiter";
|
|
1267
|
-
})(KnownTokenFilterName || (KnownTokenFilterName = {}));
|
|
1268
|
-
/** Known values of {@link CharFilterName} that the service accepts. */
|
|
1269
|
-
export var KnownCharFilterName;
|
|
1270
|
-
(function (KnownCharFilterName) {
|
|
1271
|
-
/** A character filter that attempts to strip out HTML constructs. See https:\//lucene.apache.org\/core\/4_10_3\/analyzers-common\/org\/apache\/lucene\/analysis\/charfilter\/HTMLStripCharFilter.html */
|
|
1272
|
-
KnownCharFilterName["HtmlStrip"] = "html_strip";
|
|
1273
|
-
})(KnownCharFilterName || (KnownCharFilterName = {}));
|
|
1274
1366
|
/** Known values of {@link RegexFlags} that the service accepts. */
|
|
1275
1367
|
export var KnownRegexFlags;
|
|
1276
1368
|
(function (KnownRegexFlags) {
|