@memberjunction/server 2.51.0 → 2.52.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/generated/generated.d.ts +63 -0
- package/dist/generated/generated.d.ts.map +1 -1
- package/dist/generated/generated.js +317 -0
- package/dist/generated/generated.js.map +1 -1
- package/dist/resolvers/RunAIPromptResolver.d.ts +1 -1
- package/dist/resolvers/RunAIPromptResolver.d.ts.map +1 -1
- package/dist/resolvers/RunAIPromptResolver.js +53 -3
- package/dist/resolvers/RunAIPromptResolver.js.map +1 -1
- package/package.json +25 -25
- package/src/generated/generated.ts +191 -0
- package/src/resolvers/RunAIPromptResolver.ts +52 -2
|
@@ -5829,6 +5829,16 @@ let AIPrompt_ = class AIPrompt_ {
|
|
|
5829
5829
|
CacheMustMatchConfig;
|
|
5830
5830
|
PromptRole;
|
|
5831
5831
|
PromptPosition;
|
|
5832
|
+
Temperature;
|
|
5833
|
+
TopP;
|
|
5834
|
+
TopK;
|
|
5835
|
+
MinP;
|
|
5836
|
+
FrequencyPenalty;
|
|
5837
|
+
PresencePenalty;
|
|
5838
|
+
Seed;
|
|
5839
|
+
StopSequences;
|
|
5840
|
+
IncludeLogProbs;
|
|
5841
|
+
TopLogProbs;
|
|
5832
5842
|
Template;
|
|
5833
5843
|
Category;
|
|
5834
5844
|
Type;
|
|
@@ -6005,6 +6015,47 @@ __decorate([
|
|
|
6005
6015
|
MaxLength(40),
|
|
6006
6016
|
__metadata("design:type", String)
|
|
6007
6017
|
], AIPrompt_.prototype, "PromptPosition", void 0);
|
|
6018
|
+
__decorate([
|
|
6019
|
+
Field(() => Float, { nullable: true, description: `Default temperature setting for this prompt. Controls randomness in the output. 0 = more focused and deterministic, 2 = more random and creative. Can be overridden at runtime.` }),
|
|
6020
|
+
__metadata("design:type", Number)
|
|
6021
|
+
], AIPrompt_.prototype, "Temperature", void 0);
|
|
6022
|
+
__decorate([
|
|
6023
|
+
Field(() => Float, { nullable: true, description: `Default TopP (nucleus sampling) for this prompt. Only consider tokens with cumulative probability up to this value. 1 = consider all tokens. Can be overridden at runtime.` }),
|
|
6024
|
+
__metadata("design:type", Number)
|
|
6025
|
+
], AIPrompt_.prototype, "TopP", void 0);
|
|
6026
|
+
__decorate([
|
|
6027
|
+
Field(() => Int, { nullable: true, description: `Default TopK sampling for this prompt. Only sample from the top K tokens. Lower values reduce randomness. Can be overridden at runtime.` }),
|
|
6028
|
+
__metadata("design:type", Number)
|
|
6029
|
+
], AIPrompt_.prototype, "TopK", void 0);
|
|
6030
|
+
__decorate([
|
|
6031
|
+
Field(() => Float, { nullable: true, description: `Default MinP (minimum probability) for this prompt. Tokens with probability below this threshold are filtered out. Can be overridden at runtime.` }),
|
|
6032
|
+
__metadata("design:type", Number)
|
|
6033
|
+
], AIPrompt_.prototype, "MinP", void 0);
|
|
6034
|
+
__decorate([
|
|
6035
|
+
Field(() => Float, { nullable: true, description: `Default frequency penalty for this prompt. Penalizes tokens based on their frequency in the text. Positive values decrease likelihood of repetition. Can be overridden at runtime.` }),
|
|
6036
|
+
__metadata("design:type", Number)
|
|
6037
|
+
], AIPrompt_.prototype, "FrequencyPenalty", void 0);
|
|
6038
|
+
__decorate([
|
|
6039
|
+
Field(() => Float, { nullable: true, description: `Default presence penalty for this prompt. Penalizes tokens that have appeared in the text. Positive values increase topic diversity. Can be overridden at runtime.` }),
|
|
6040
|
+
__metadata("design:type", Number)
|
|
6041
|
+
], AIPrompt_.prototype, "PresencePenalty", void 0);
|
|
6042
|
+
__decorate([
|
|
6043
|
+
Field(() => Int, { nullable: true, description: `Default random seed for this prompt. Used for deterministic generation. Same seed produces same output. Can be overridden at runtime.` }),
|
|
6044
|
+
__metadata("design:type", Number)
|
|
6045
|
+
], AIPrompt_.prototype, "Seed", void 0);
|
|
6046
|
+
__decorate([
|
|
6047
|
+
Field({ nullable: true, description: `Default stop sequences for this prompt. Comma-delimited list of sequences that will stop generation when encountered. Can be overridden at runtime.` }),
|
|
6048
|
+
MaxLength(2000),
|
|
6049
|
+
__metadata("design:type", String)
|
|
6050
|
+
], AIPrompt_.prototype, "StopSequences", void 0);
|
|
6051
|
+
__decorate([
|
|
6052
|
+
Field(() => Boolean, { nullable: true, description: `Default setting for including log probabilities in the response. Can be overridden at runtime.` }),
|
|
6053
|
+
__metadata("design:type", Boolean)
|
|
6054
|
+
], AIPrompt_.prototype, "IncludeLogProbs", void 0);
|
|
6055
|
+
__decorate([
|
|
6056
|
+
Field(() => Int, { nullable: true, description: `Default number of top log probabilities to include when IncludeLogProbs is true. Can be overridden at runtime.` }),
|
|
6057
|
+
__metadata("design:type", Number)
|
|
6058
|
+
], AIPrompt_.prototype, "TopLogProbs", void 0);
|
|
6008
6059
|
__decorate([
|
|
6009
6060
|
Field(),
|
|
6010
6061
|
MaxLength(510),
|
|
@@ -6104,6 +6155,16 @@ let CreateAIPromptInput = class CreateAIPromptInput {
|
|
|
6104
6155
|
CacheMustMatchConfig;
|
|
6105
6156
|
PromptRole;
|
|
6106
6157
|
PromptPosition;
|
|
6158
|
+
Temperature;
|
|
6159
|
+
TopP;
|
|
6160
|
+
TopK;
|
|
6161
|
+
MinP;
|
|
6162
|
+
FrequencyPenalty;
|
|
6163
|
+
PresencePenalty;
|
|
6164
|
+
Seed;
|
|
6165
|
+
StopSequences;
|
|
6166
|
+
IncludeLogProbs;
|
|
6167
|
+
TopLogProbs;
|
|
6107
6168
|
};
|
|
6108
6169
|
__decorate([
|
|
6109
6170
|
Field({ nullable: true }),
|
|
@@ -6237,6 +6298,46 @@ __decorate([
|
|
|
6237
6298
|
Field({ nullable: true }),
|
|
6238
6299
|
__metadata("design:type", String)
|
|
6239
6300
|
], CreateAIPromptInput.prototype, "PromptPosition", void 0);
|
|
6301
|
+
__decorate([
|
|
6302
|
+
Field(() => Float, { nullable: true }),
|
|
6303
|
+
__metadata("design:type", Number)
|
|
6304
|
+
], CreateAIPromptInput.prototype, "Temperature", void 0);
|
|
6305
|
+
__decorate([
|
|
6306
|
+
Field(() => Float, { nullable: true }),
|
|
6307
|
+
__metadata("design:type", Number)
|
|
6308
|
+
], CreateAIPromptInput.prototype, "TopP", void 0);
|
|
6309
|
+
__decorate([
|
|
6310
|
+
Field(() => Int, { nullable: true }),
|
|
6311
|
+
__metadata("design:type", Number)
|
|
6312
|
+
], CreateAIPromptInput.prototype, "TopK", void 0);
|
|
6313
|
+
__decorate([
|
|
6314
|
+
Field(() => Float, { nullable: true }),
|
|
6315
|
+
__metadata("design:type", Number)
|
|
6316
|
+
], CreateAIPromptInput.prototype, "MinP", void 0);
|
|
6317
|
+
__decorate([
|
|
6318
|
+
Field(() => Float, { nullable: true }),
|
|
6319
|
+
__metadata("design:type", Number)
|
|
6320
|
+
], CreateAIPromptInput.prototype, "FrequencyPenalty", void 0);
|
|
6321
|
+
__decorate([
|
|
6322
|
+
Field(() => Float, { nullable: true }),
|
|
6323
|
+
__metadata("design:type", Number)
|
|
6324
|
+
], CreateAIPromptInput.prototype, "PresencePenalty", void 0);
|
|
6325
|
+
__decorate([
|
|
6326
|
+
Field(() => Int, { nullable: true }),
|
|
6327
|
+
__metadata("design:type", Number)
|
|
6328
|
+
], CreateAIPromptInput.prototype, "Seed", void 0);
|
|
6329
|
+
__decorate([
|
|
6330
|
+
Field({ nullable: true }),
|
|
6331
|
+
__metadata("design:type", String)
|
|
6332
|
+
], CreateAIPromptInput.prototype, "StopSequences", void 0);
|
|
6333
|
+
__decorate([
|
|
6334
|
+
Field(() => Boolean, { nullable: true }),
|
|
6335
|
+
__metadata("design:type", Boolean)
|
|
6336
|
+
], CreateAIPromptInput.prototype, "IncludeLogProbs", void 0);
|
|
6337
|
+
__decorate([
|
|
6338
|
+
Field(() => Int, { nullable: true }),
|
|
6339
|
+
__metadata("design:type", Number)
|
|
6340
|
+
], CreateAIPromptInput.prototype, "TopLogProbs", void 0);
|
|
6240
6341
|
CreateAIPromptInput = __decorate([
|
|
6241
6342
|
InputType()
|
|
6242
6343
|
], CreateAIPromptInput);
|
|
@@ -6275,6 +6376,16 @@ let UpdateAIPromptInput = class UpdateAIPromptInput {
|
|
|
6275
6376
|
CacheMustMatchConfig;
|
|
6276
6377
|
PromptRole;
|
|
6277
6378
|
PromptPosition;
|
|
6379
|
+
Temperature;
|
|
6380
|
+
TopP;
|
|
6381
|
+
TopK;
|
|
6382
|
+
MinP;
|
|
6383
|
+
FrequencyPenalty;
|
|
6384
|
+
PresencePenalty;
|
|
6385
|
+
Seed;
|
|
6386
|
+
StopSequences;
|
|
6387
|
+
IncludeLogProbs;
|
|
6388
|
+
TopLogProbs;
|
|
6278
6389
|
OldValues___;
|
|
6279
6390
|
};
|
|
6280
6391
|
__decorate([
|
|
@@ -6409,6 +6520,46 @@ __decorate([
|
|
|
6409
6520
|
Field({ nullable: true }),
|
|
6410
6521
|
__metadata("design:type", String)
|
|
6411
6522
|
], UpdateAIPromptInput.prototype, "PromptPosition", void 0);
|
|
6523
|
+
__decorate([
|
|
6524
|
+
Field(() => Float, { nullable: true }),
|
|
6525
|
+
__metadata("design:type", Number)
|
|
6526
|
+
], UpdateAIPromptInput.prototype, "Temperature", void 0);
|
|
6527
|
+
__decorate([
|
|
6528
|
+
Field(() => Float, { nullable: true }),
|
|
6529
|
+
__metadata("design:type", Number)
|
|
6530
|
+
], UpdateAIPromptInput.prototype, "TopP", void 0);
|
|
6531
|
+
__decorate([
|
|
6532
|
+
Field(() => Int, { nullable: true }),
|
|
6533
|
+
__metadata("design:type", Number)
|
|
6534
|
+
], UpdateAIPromptInput.prototype, "TopK", void 0);
|
|
6535
|
+
__decorate([
|
|
6536
|
+
Field(() => Float, { nullable: true }),
|
|
6537
|
+
__metadata("design:type", Number)
|
|
6538
|
+
], UpdateAIPromptInput.prototype, "MinP", void 0);
|
|
6539
|
+
__decorate([
|
|
6540
|
+
Field(() => Float, { nullable: true }),
|
|
6541
|
+
__metadata("design:type", Number)
|
|
6542
|
+
], UpdateAIPromptInput.prototype, "FrequencyPenalty", void 0);
|
|
6543
|
+
__decorate([
|
|
6544
|
+
Field(() => Float, { nullable: true }),
|
|
6545
|
+
__metadata("design:type", Number)
|
|
6546
|
+
], UpdateAIPromptInput.prototype, "PresencePenalty", void 0);
|
|
6547
|
+
__decorate([
|
|
6548
|
+
Field(() => Int, { nullable: true }),
|
|
6549
|
+
__metadata("design:type", Number)
|
|
6550
|
+
], UpdateAIPromptInput.prototype, "Seed", void 0);
|
|
6551
|
+
__decorate([
|
|
6552
|
+
Field({ nullable: true }),
|
|
6553
|
+
__metadata("design:type", String)
|
|
6554
|
+
], UpdateAIPromptInput.prototype, "StopSequences", void 0);
|
|
6555
|
+
__decorate([
|
|
6556
|
+
Field(() => Boolean, { nullable: true }),
|
|
6557
|
+
__metadata("design:type", Boolean)
|
|
6558
|
+
], UpdateAIPromptInput.prototype, "IncludeLogProbs", void 0);
|
|
6559
|
+
__decorate([
|
|
6560
|
+
Field(() => Int, { nullable: true }),
|
|
6561
|
+
__metadata("design:type", Number)
|
|
6562
|
+
], UpdateAIPromptInput.prototype, "TopLogProbs", void 0);
|
|
6412
6563
|
__decorate([
|
|
6413
6564
|
Field(() => [KeyValuePairInput], { nullable: true }),
|
|
6414
6565
|
__metadata("design:type", Array)
|
|
@@ -59066,6 +59217,17 @@ let AIPromptRun_ = class AIPromptRun_ {
|
|
|
59066
59217
|
TokensUsedRollup;
|
|
59067
59218
|
TokensPromptRollup;
|
|
59068
59219
|
TokensCompletionRollup;
|
|
59220
|
+
Temperature;
|
|
59221
|
+
TopP;
|
|
59222
|
+
TopK;
|
|
59223
|
+
MinP;
|
|
59224
|
+
FrequencyPenalty;
|
|
59225
|
+
PresencePenalty;
|
|
59226
|
+
Seed;
|
|
59227
|
+
StopSequences;
|
|
59228
|
+
ResponseFormat;
|
|
59229
|
+
LogProbs;
|
|
59230
|
+
TopLogProbs;
|
|
59069
59231
|
Prompt;
|
|
59070
59232
|
Model;
|
|
59071
59233
|
Vendor;
|
|
@@ -59200,6 +59362,51 @@ __decorate([
|
|
|
59200
59362
|
Field(() => Int, { nullable: true, description: `Total completion/output tokens including this execution and all child/grandchild executions. For leaf nodes (no children), this equals TokensCompletion. For parent nodes, this includes the sum of all descendant completion tokens.` }),
|
|
59201
59363
|
__metadata("design:type", Number)
|
|
59202
59364
|
], AIPromptRun_.prototype, "TokensCompletionRollup", void 0);
|
|
59365
|
+
__decorate([
|
|
59366
|
+
Field(() => Float, { nullable: true, description: `The temperature parameter used for this prompt run, controlling randomness in the output (0.0 = deterministic, 2.0 = very random)` }),
|
|
59367
|
+
__metadata("design:type", Number)
|
|
59368
|
+
], AIPromptRun_.prototype, "Temperature", void 0);
|
|
59369
|
+
__decorate([
|
|
59370
|
+
Field(() => Float, { nullable: true, description: `Top-p (nucleus) sampling parameter used for this run. Considers tokens with cumulative probability up to this value (0-1)` }),
|
|
59371
|
+
__metadata("design:type", Number)
|
|
59372
|
+
], AIPromptRun_.prototype, "TopP", void 0);
|
|
59373
|
+
__decorate([
|
|
59374
|
+
Field(() => Int, { nullable: true, description: `Top-k sampling parameter used for this run. Limits sampling to the top K most likely tokens` }),
|
|
59375
|
+
__metadata("design:type", Number)
|
|
59376
|
+
], AIPromptRun_.prototype, "TopK", void 0);
|
|
59377
|
+
__decorate([
|
|
59378
|
+
Field(() => Float, { nullable: true, description: `Minimum probability threshold used for token sampling (0-1). Tokens below this probability are filtered out` }),
|
|
59379
|
+
__metadata("design:type", Number)
|
|
59380
|
+
], AIPromptRun_.prototype, "MinP", void 0);
|
|
59381
|
+
__decorate([
|
|
59382
|
+
Field(() => Float, { nullable: true, description: `Frequency penalty parameter used (-2.0 to 2.0). Positive values reduce repetition of tokens based on their frequency in the output` }),
|
|
59383
|
+
__metadata("design:type", Number)
|
|
59384
|
+
], AIPromptRun_.prototype, "FrequencyPenalty", void 0);
|
|
59385
|
+
__decorate([
|
|
59386
|
+
Field(() => Float, { nullable: true, description: `Presence penalty parameter used (-2.0 to 2.0). Positive values encourage the model to talk about new topics` }),
|
|
59387
|
+
__metadata("design:type", Number)
|
|
59388
|
+
], AIPromptRun_.prototype, "PresencePenalty", void 0);
|
|
59389
|
+
__decorate([
|
|
59390
|
+
Field(() => Int, { nullable: true, description: `Random seed used for reproducible outputs. When set, the same seed with identical inputs should produce the same output` }),
|
|
59391
|
+
__metadata("design:type", Number)
|
|
59392
|
+
], AIPromptRun_.prototype, "Seed", void 0);
|
|
59393
|
+
__decorate([
|
|
59394
|
+
Field({ nullable: true, description: `JSON array of stop sequences used. The model stops generating when any of these sequences are encountered` }),
|
|
59395
|
+
__metadata("design:type", String)
|
|
59396
|
+
], AIPromptRun_.prototype, "StopSequences", void 0);
|
|
59397
|
+
__decorate([
|
|
59398
|
+
Field({ nullable: true, description: `The response format requested for this run (e.g., 'JSON', 'Text', 'Markdown')` }),
|
|
59399
|
+
MaxLength(100),
|
|
59400
|
+
__metadata("design:type", String)
|
|
59401
|
+
], AIPromptRun_.prototype, "ResponseFormat", void 0);
|
|
59402
|
+
__decorate([
|
|
59403
|
+
Field(() => Boolean, { nullable: true, description: `Whether log probabilities were requested for this run` }),
|
|
59404
|
+
__metadata("design:type", Boolean)
|
|
59405
|
+
], AIPromptRun_.prototype, "LogProbs", void 0);
|
|
59406
|
+
__decorate([
|
|
59407
|
+
Field(() => Int, { nullable: true, description: `Number of top log probabilities requested per token (if LogProbs is true)` }),
|
|
59408
|
+
__metadata("design:type", Number)
|
|
59409
|
+
], AIPromptRun_.prototype, "TopLogProbs", void 0);
|
|
59203
59410
|
__decorate([
|
|
59204
59411
|
Field(),
|
|
59205
59412
|
MaxLength(510),
|
|
@@ -59264,6 +59471,17 @@ let CreateAIPromptRunInput = class CreateAIPromptRunInput {
|
|
|
59264
59471
|
TokensUsedRollup;
|
|
59265
59472
|
TokensPromptRollup;
|
|
59266
59473
|
TokensCompletionRollup;
|
|
59474
|
+
Temperature;
|
|
59475
|
+
TopP;
|
|
59476
|
+
TopK;
|
|
59477
|
+
MinP;
|
|
59478
|
+
FrequencyPenalty;
|
|
59479
|
+
PresencePenalty;
|
|
59480
|
+
Seed;
|
|
59481
|
+
StopSequences;
|
|
59482
|
+
ResponseFormat;
|
|
59483
|
+
LogProbs;
|
|
59484
|
+
TopLogProbs;
|
|
59267
59485
|
};
|
|
59268
59486
|
__decorate([
|
|
59269
59487
|
Field({ nullable: true }),
|
|
@@ -59369,6 +59587,50 @@ __decorate([
|
|
|
59369
59587
|
Field(() => Int, { nullable: true }),
|
|
59370
59588
|
__metadata("design:type", Number)
|
|
59371
59589
|
], CreateAIPromptRunInput.prototype, "TokensCompletionRollup", void 0);
|
|
59590
|
+
__decorate([
|
|
59591
|
+
Field(() => Float, { nullable: true }),
|
|
59592
|
+
__metadata("design:type", Number)
|
|
59593
|
+
], CreateAIPromptRunInput.prototype, "Temperature", void 0);
|
|
59594
|
+
__decorate([
|
|
59595
|
+
Field(() => Float, { nullable: true }),
|
|
59596
|
+
__metadata("design:type", Number)
|
|
59597
|
+
], CreateAIPromptRunInput.prototype, "TopP", void 0);
|
|
59598
|
+
__decorate([
|
|
59599
|
+
Field(() => Int, { nullable: true }),
|
|
59600
|
+
__metadata("design:type", Number)
|
|
59601
|
+
], CreateAIPromptRunInput.prototype, "TopK", void 0);
|
|
59602
|
+
__decorate([
|
|
59603
|
+
Field(() => Float, { nullable: true }),
|
|
59604
|
+
__metadata("design:type", Number)
|
|
59605
|
+
], CreateAIPromptRunInput.prototype, "MinP", void 0);
|
|
59606
|
+
__decorate([
|
|
59607
|
+
Field(() => Float, { nullable: true }),
|
|
59608
|
+
__metadata("design:type", Number)
|
|
59609
|
+
], CreateAIPromptRunInput.prototype, "FrequencyPenalty", void 0);
|
|
59610
|
+
__decorate([
|
|
59611
|
+
Field(() => Float, { nullable: true }),
|
|
59612
|
+
__metadata("design:type", Number)
|
|
59613
|
+
], CreateAIPromptRunInput.prototype, "PresencePenalty", void 0);
|
|
59614
|
+
__decorate([
|
|
59615
|
+
Field(() => Int, { nullable: true }),
|
|
59616
|
+
__metadata("design:type", Number)
|
|
59617
|
+
], CreateAIPromptRunInput.prototype, "Seed", void 0);
|
|
59618
|
+
__decorate([
|
|
59619
|
+
Field({ nullable: true }),
|
|
59620
|
+
__metadata("design:type", String)
|
|
59621
|
+
], CreateAIPromptRunInput.prototype, "StopSequences", void 0);
|
|
59622
|
+
__decorate([
|
|
59623
|
+
Field({ nullable: true }),
|
|
59624
|
+
__metadata("design:type", String)
|
|
59625
|
+
], CreateAIPromptRunInput.prototype, "ResponseFormat", void 0);
|
|
59626
|
+
__decorate([
|
|
59627
|
+
Field(() => Boolean, { nullable: true }),
|
|
59628
|
+
__metadata("design:type", Boolean)
|
|
59629
|
+
], CreateAIPromptRunInput.prototype, "LogProbs", void 0);
|
|
59630
|
+
__decorate([
|
|
59631
|
+
Field(() => Int, { nullable: true }),
|
|
59632
|
+
__metadata("design:type", Number)
|
|
59633
|
+
], CreateAIPromptRunInput.prototype, "TopLogProbs", void 0);
|
|
59372
59634
|
CreateAIPromptRunInput = __decorate([
|
|
59373
59635
|
InputType()
|
|
59374
59636
|
], CreateAIPromptRunInput);
|
|
@@ -59400,6 +59662,17 @@ let UpdateAIPromptRunInput = class UpdateAIPromptRunInput {
|
|
|
59400
59662
|
TokensUsedRollup;
|
|
59401
59663
|
TokensPromptRollup;
|
|
59402
59664
|
TokensCompletionRollup;
|
|
59665
|
+
Temperature;
|
|
59666
|
+
TopP;
|
|
59667
|
+
TopK;
|
|
59668
|
+
MinP;
|
|
59669
|
+
FrequencyPenalty;
|
|
59670
|
+
PresencePenalty;
|
|
59671
|
+
Seed;
|
|
59672
|
+
StopSequences;
|
|
59673
|
+
ResponseFormat;
|
|
59674
|
+
LogProbs;
|
|
59675
|
+
TopLogProbs;
|
|
59403
59676
|
OldValues___;
|
|
59404
59677
|
};
|
|
59405
59678
|
__decorate([
|
|
@@ -59506,6 +59779,50 @@ __decorate([
|
|
|
59506
59779
|
Field(() => Int, { nullable: true }),
|
|
59507
59780
|
__metadata("design:type", Number)
|
|
59508
59781
|
], UpdateAIPromptRunInput.prototype, "TokensCompletionRollup", void 0);
|
|
59782
|
+
__decorate([
|
|
59783
|
+
Field(() => Float, { nullable: true }),
|
|
59784
|
+
__metadata("design:type", Number)
|
|
59785
|
+
], UpdateAIPromptRunInput.prototype, "Temperature", void 0);
|
|
59786
|
+
__decorate([
|
|
59787
|
+
Field(() => Float, { nullable: true }),
|
|
59788
|
+
__metadata("design:type", Number)
|
|
59789
|
+
], UpdateAIPromptRunInput.prototype, "TopP", void 0);
|
|
59790
|
+
__decorate([
|
|
59791
|
+
Field(() => Int, { nullable: true }),
|
|
59792
|
+
__metadata("design:type", Number)
|
|
59793
|
+
], UpdateAIPromptRunInput.prototype, "TopK", void 0);
|
|
59794
|
+
__decorate([
|
|
59795
|
+
Field(() => Float, { nullable: true }),
|
|
59796
|
+
__metadata("design:type", Number)
|
|
59797
|
+
], UpdateAIPromptRunInput.prototype, "MinP", void 0);
|
|
59798
|
+
__decorate([
|
|
59799
|
+
Field(() => Float, { nullable: true }),
|
|
59800
|
+
__metadata("design:type", Number)
|
|
59801
|
+
], UpdateAIPromptRunInput.prototype, "FrequencyPenalty", void 0);
|
|
59802
|
+
__decorate([
|
|
59803
|
+
Field(() => Float, { nullable: true }),
|
|
59804
|
+
__metadata("design:type", Number)
|
|
59805
|
+
], UpdateAIPromptRunInput.prototype, "PresencePenalty", void 0);
|
|
59806
|
+
__decorate([
|
|
59807
|
+
Field(() => Int, { nullable: true }),
|
|
59808
|
+
__metadata("design:type", Number)
|
|
59809
|
+
], UpdateAIPromptRunInput.prototype, "Seed", void 0);
|
|
59810
|
+
__decorate([
|
|
59811
|
+
Field({ nullable: true }),
|
|
59812
|
+
__metadata("design:type", String)
|
|
59813
|
+
], UpdateAIPromptRunInput.prototype, "StopSequences", void 0);
|
|
59814
|
+
__decorate([
|
|
59815
|
+
Field({ nullable: true }),
|
|
59816
|
+
__metadata("design:type", String)
|
|
59817
|
+
], UpdateAIPromptRunInput.prototype, "ResponseFormat", void 0);
|
|
59818
|
+
__decorate([
|
|
59819
|
+
Field(() => Boolean, { nullable: true }),
|
|
59820
|
+
__metadata("design:type", Boolean)
|
|
59821
|
+
], UpdateAIPromptRunInput.prototype, "LogProbs", void 0);
|
|
59822
|
+
__decorate([
|
|
59823
|
+
Field(() => Int, { nullable: true }),
|
|
59824
|
+
__metadata("design:type", Number)
|
|
59825
|
+
], UpdateAIPromptRunInput.prototype, "TopLogProbs", void 0);
|
|
59509
59826
|
__decorate([
|
|
59510
59827
|
Field(() => [KeyValuePairInput], { nullable: true }),
|
|
59511
59828
|
__metadata("design:type", Array)
|