@memberjunction/server 2.51.0 → 2.52.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -5829,6 +5829,16 @@ let AIPrompt_ = class AIPrompt_ {
5829
5829
  CacheMustMatchConfig;
5830
5830
  PromptRole;
5831
5831
  PromptPosition;
5832
+ Temperature;
5833
+ TopP;
5834
+ TopK;
5835
+ MinP;
5836
+ FrequencyPenalty;
5837
+ PresencePenalty;
5838
+ Seed;
5839
+ StopSequences;
5840
+ IncludeLogProbs;
5841
+ TopLogProbs;
5832
5842
  Template;
5833
5843
  Category;
5834
5844
  Type;
@@ -6005,6 +6015,47 @@ __decorate([
6005
6015
  MaxLength(40),
6006
6016
  __metadata("design:type", String)
6007
6017
  ], AIPrompt_.prototype, "PromptPosition", void 0);
6018
+ __decorate([
6019
+ Field(() => Float, { nullable: true, description: `Default temperature setting for this prompt. Controls randomness in the output. 0 = more focused and deterministic, 2 = more random and creative. Can be overridden at runtime.` }),
6020
+ __metadata("design:type", Number)
6021
+ ], AIPrompt_.prototype, "Temperature", void 0);
6022
+ __decorate([
6023
+ Field(() => Float, { nullable: true, description: `Default TopP (nucleus sampling) for this prompt. Only consider tokens with cumulative probability up to this value. 1 = consider all tokens. Can be overridden at runtime.` }),
6024
+ __metadata("design:type", Number)
6025
+ ], AIPrompt_.prototype, "TopP", void 0);
6026
+ __decorate([
6027
+ Field(() => Int, { nullable: true, description: `Default TopK sampling for this prompt. Only sample from the top K tokens. Lower values reduce randomness. Can be overridden at runtime.` }),
6028
+ __metadata("design:type", Number)
6029
+ ], AIPrompt_.prototype, "TopK", void 0);
6030
+ __decorate([
6031
+ Field(() => Float, { nullable: true, description: `Default MinP (minimum probability) for this prompt. Tokens with probability below this threshold are filtered out. Can be overridden at runtime.` }),
6032
+ __metadata("design:type", Number)
6033
+ ], AIPrompt_.prototype, "MinP", void 0);
6034
+ __decorate([
6035
+ Field(() => Float, { nullable: true, description: `Default frequency penalty for this prompt. Penalizes tokens based on their frequency in the text. Positive values decrease likelihood of repetition. Can be overridden at runtime.` }),
6036
+ __metadata("design:type", Number)
6037
+ ], AIPrompt_.prototype, "FrequencyPenalty", void 0);
6038
+ __decorate([
6039
+ Field(() => Float, { nullable: true, description: `Default presence penalty for this prompt. Penalizes tokens that have appeared in the text. Positive values increase topic diversity. Can be overridden at runtime.` }),
6040
+ __metadata("design:type", Number)
6041
+ ], AIPrompt_.prototype, "PresencePenalty", void 0);
6042
+ __decorate([
6043
+ Field(() => Int, { nullable: true, description: `Default random seed for this prompt. Used for deterministic generation. Same seed produces same output. Can be overridden at runtime.` }),
6044
+ __metadata("design:type", Number)
6045
+ ], AIPrompt_.prototype, "Seed", void 0);
6046
+ __decorate([
6047
+ Field({ nullable: true, description: `Default stop sequences for this prompt. Comma-delimited list of sequences that will stop generation when encountered. Can be overridden at runtime.` }),
6048
+ MaxLength(2000),
6049
+ __metadata("design:type", String)
6050
+ ], AIPrompt_.prototype, "StopSequences", void 0);
6051
+ __decorate([
6052
+ Field(() => Boolean, { nullable: true, description: `Default setting for including log probabilities in the response. Can be overridden at runtime.` }),
6053
+ __metadata("design:type", Boolean)
6054
+ ], AIPrompt_.prototype, "IncludeLogProbs", void 0);
6055
+ __decorate([
6056
+ Field(() => Int, { nullable: true, description: `Default number of top log probabilities to include when IncludeLogProbs is true. Can be overridden at runtime.` }),
6057
+ __metadata("design:type", Number)
6058
+ ], AIPrompt_.prototype, "TopLogProbs", void 0);
6008
6059
  __decorate([
6009
6060
  Field(),
6010
6061
  MaxLength(510),
@@ -6104,6 +6155,16 @@ let CreateAIPromptInput = class CreateAIPromptInput {
6104
6155
  CacheMustMatchConfig;
6105
6156
  PromptRole;
6106
6157
  PromptPosition;
6158
+ Temperature;
6159
+ TopP;
6160
+ TopK;
6161
+ MinP;
6162
+ FrequencyPenalty;
6163
+ PresencePenalty;
6164
+ Seed;
6165
+ StopSequences;
6166
+ IncludeLogProbs;
6167
+ TopLogProbs;
6107
6168
  };
6108
6169
  __decorate([
6109
6170
  Field({ nullable: true }),
@@ -6237,6 +6298,46 @@ __decorate([
6237
6298
  Field({ nullable: true }),
6238
6299
  __metadata("design:type", String)
6239
6300
  ], CreateAIPromptInput.prototype, "PromptPosition", void 0);
6301
+ __decorate([
6302
+ Field(() => Float, { nullable: true }),
6303
+ __metadata("design:type", Number)
6304
+ ], CreateAIPromptInput.prototype, "Temperature", void 0);
6305
+ __decorate([
6306
+ Field(() => Float, { nullable: true }),
6307
+ __metadata("design:type", Number)
6308
+ ], CreateAIPromptInput.prototype, "TopP", void 0);
6309
+ __decorate([
6310
+ Field(() => Int, { nullable: true }),
6311
+ __metadata("design:type", Number)
6312
+ ], CreateAIPromptInput.prototype, "TopK", void 0);
6313
+ __decorate([
6314
+ Field(() => Float, { nullable: true }),
6315
+ __metadata("design:type", Number)
6316
+ ], CreateAIPromptInput.prototype, "MinP", void 0);
6317
+ __decorate([
6318
+ Field(() => Float, { nullable: true }),
6319
+ __metadata("design:type", Number)
6320
+ ], CreateAIPromptInput.prototype, "FrequencyPenalty", void 0);
6321
+ __decorate([
6322
+ Field(() => Float, { nullable: true }),
6323
+ __metadata("design:type", Number)
6324
+ ], CreateAIPromptInput.prototype, "PresencePenalty", void 0);
6325
+ __decorate([
6326
+ Field(() => Int, { nullable: true }),
6327
+ __metadata("design:type", Number)
6328
+ ], CreateAIPromptInput.prototype, "Seed", void 0);
6329
+ __decorate([
6330
+ Field({ nullable: true }),
6331
+ __metadata("design:type", String)
6332
+ ], CreateAIPromptInput.prototype, "StopSequences", void 0);
6333
+ __decorate([
6334
+ Field(() => Boolean, { nullable: true }),
6335
+ __metadata("design:type", Boolean)
6336
+ ], CreateAIPromptInput.prototype, "IncludeLogProbs", void 0);
6337
+ __decorate([
6338
+ Field(() => Int, { nullable: true }),
6339
+ __metadata("design:type", Number)
6340
+ ], CreateAIPromptInput.prototype, "TopLogProbs", void 0);
6240
6341
  CreateAIPromptInput = __decorate([
6241
6342
  InputType()
6242
6343
  ], CreateAIPromptInput);
@@ -6275,6 +6376,16 @@ let UpdateAIPromptInput = class UpdateAIPromptInput {
6275
6376
  CacheMustMatchConfig;
6276
6377
  PromptRole;
6277
6378
  PromptPosition;
6379
+ Temperature;
6380
+ TopP;
6381
+ TopK;
6382
+ MinP;
6383
+ FrequencyPenalty;
6384
+ PresencePenalty;
6385
+ Seed;
6386
+ StopSequences;
6387
+ IncludeLogProbs;
6388
+ TopLogProbs;
6278
6389
  OldValues___;
6279
6390
  };
6280
6391
  __decorate([
@@ -6409,6 +6520,46 @@ __decorate([
6409
6520
  Field({ nullable: true }),
6410
6521
  __metadata("design:type", String)
6411
6522
  ], UpdateAIPromptInput.prototype, "PromptPosition", void 0);
6523
+ __decorate([
6524
+ Field(() => Float, { nullable: true }),
6525
+ __metadata("design:type", Number)
6526
+ ], UpdateAIPromptInput.prototype, "Temperature", void 0);
6527
+ __decorate([
6528
+ Field(() => Float, { nullable: true }),
6529
+ __metadata("design:type", Number)
6530
+ ], UpdateAIPromptInput.prototype, "TopP", void 0);
6531
+ __decorate([
6532
+ Field(() => Int, { nullable: true }),
6533
+ __metadata("design:type", Number)
6534
+ ], UpdateAIPromptInput.prototype, "TopK", void 0);
6535
+ __decorate([
6536
+ Field(() => Float, { nullable: true }),
6537
+ __metadata("design:type", Number)
6538
+ ], UpdateAIPromptInput.prototype, "MinP", void 0);
6539
+ __decorate([
6540
+ Field(() => Float, { nullable: true }),
6541
+ __metadata("design:type", Number)
6542
+ ], UpdateAIPromptInput.prototype, "FrequencyPenalty", void 0);
6543
+ __decorate([
6544
+ Field(() => Float, { nullable: true }),
6545
+ __metadata("design:type", Number)
6546
+ ], UpdateAIPromptInput.prototype, "PresencePenalty", void 0);
6547
+ __decorate([
6548
+ Field(() => Int, { nullable: true }),
6549
+ __metadata("design:type", Number)
6550
+ ], UpdateAIPromptInput.prototype, "Seed", void 0);
6551
+ __decorate([
6552
+ Field({ nullable: true }),
6553
+ __metadata("design:type", String)
6554
+ ], UpdateAIPromptInput.prototype, "StopSequences", void 0);
6555
+ __decorate([
6556
+ Field(() => Boolean, { nullable: true }),
6557
+ __metadata("design:type", Boolean)
6558
+ ], UpdateAIPromptInput.prototype, "IncludeLogProbs", void 0);
6559
+ __decorate([
6560
+ Field(() => Int, { nullable: true }),
6561
+ __metadata("design:type", Number)
6562
+ ], UpdateAIPromptInput.prototype, "TopLogProbs", void 0);
6412
6563
  __decorate([
6413
6564
  Field(() => [KeyValuePairInput], { nullable: true }),
6414
6565
  __metadata("design:type", Array)
@@ -59066,6 +59217,17 @@ let AIPromptRun_ = class AIPromptRun_ {
59066
59217
  TokensUsedRollup;
59067
59218
  TokensPromptRollup;
59068
59219
  TokensCompletionRollup;
59220
+ Temperature;
59221
+ TopP;
59222
+ TopK;
59223
+ MinP;
59224
+ FrequencyPenalty;
59225
+ PresencePenalty;
59226
+ Seed;
59227
+ StopSequences;
59228
+ ResponseFormat;
59229
+ LogProbs;
59230
+ TopLogProbs;
59069
59231
  Prompt;
59070
59232
  Model;
59071
59233
  Vendor;
@@ -59200,6 +59362,51 @@ __decorate([
59200
59362
  Field(() => Int, { nullable: true, description: `Total completion/output tokens including this execution and all child/grandchild executions. For leaf nodes (no children), this equals TokensCompletion. For parent nodes, this includes the sum of all descendant completion tokens.` }),
59201
59363
  __metadata("design:type", Number)
59202
59364
  ], AIPromptRun_.prototype, "TokensCompletionRollup", void 0);
59365
+ __decorate([
59366
+ Field(() => Float, { nullable: true, description: `The temperature parameter used for this prompt run, controlling randomness in the output (0.0 = deterministic, 2.0 = very random)` }),
59367
+ __metadata("design:type", Number)
59368
+ ], AIPromptRun_.prototype, "Temperature", void 0);
59369
+ __decorate([
59370
+ Field(() => Float, { nullable: true, description: `Top-p (nucleus) sampling parameter used for this run. Considers tokens with cumulative probability up to this value (0-1)` }),
59371
+ __metadata("design:type", Number)
59372
+ ], AIPromptRun_.prototype, "TopP", void 0);
59373
+ __decorate([
59374
+ Field(() => Int, { nullable: true, description: `Top-k sampling parameter used for this run. Limits sampling to the top K most likely tokens` }),
59375
+ __metadata("design:type", Number)
59376
+ ], AIPromptRun_.prototype, "TopK", void 0);
59377
+ __decorate([
59378
+ Field(() => Float, { nullable: true, description: `Minimum probability threshold used for token sampling (0-1). Tokens below this probability are filtered out` }),
59379
+ __metadata("design:type", Number)
59380
+ ], AIPromptRun_.prototype, "MinP", void 0);
59381
+ __decorate([
59382
+ Field(() => Float, { nullable: true, description: `Frequency penalty parameter used (-2.0 to 2.0). Positive values reduce repetition of tokens based on their frequency in the output` }),
59383
+ __metadata("design:type", Number)
59384
+ ], AIPromptRun_.prototype, "FrequencyPenalty", void 0);
59385
+ __decorate([
59386
+ Field(() => Float, { nullable: true, description: `Presence penalty parameter used (-2.0 to 2.0). Positive values encourage the model to talk about new topics` }),
59387
+ __metadata("design:type", Number)
59388
+ ], AIPromptRun_.prototype, "PresencePenalty", void 0);
59389
+ __decorate([
59390
+ Field(() => Int, { nullable: true, description: `Random seed used for reproducible outputs. When set, the same seed with identical inputs should produce the same output` }),
59391
+ __metadata("design:type", Number)
59392
+ ], AIPromptRun_.prototype, "Seed", void 0);
59393
+ __decorate([
59394
+ Field({ nullable: true, description: `JSON array of stop sequences used. The model stops generating when any of these sequences are encountered` }),
59395
+ __metadata("design:type", String)
59396
+ ], AIPromptRun_.prototype, "StopSequences", void 0);
59397
+ __decorate([
59398
+ Field({ nullable: true, description: `The response format requested for this run (e.g., 'JSON', 'Text', 'Markdown')` }),
59399
+ MaxLength(100),
59400
+ __metadata("design:type", String)
59401
+ ], AIPromptRun_.prototype, "ResponseFormat", void 0);
59402
+ __decorate([
59403
+ Field(() => Boolean, { nullable: true, description: `Whether log probabilities were requested for this run` }),
59404
+ __metadata("design:type", Boolean)
59405
+ ], AIPromptRun_.prototype, "LogProbs", void 0);
59406
+ __decorate([
59407
+ Field(() => Int, { nullable: true, description: `Number of top log probabilities requested per token (if LogProbs is true)` }),
59408
+ __metadata("design:type", Number)
59409
+ ], AIPromptRun_.prototype, "TopLogProbs", void 0);
59203
59410
  __decorate([
59204
59411
  Field(),
59205
59412
  MaxLength(510),
@@ -59264,6 +59471,17 @@ let CreateAIPromptRunInput = class CreateAIPromptRunInput {
59264
59471
  TokensUsedRollup;
59265
59472
  TokensPromptRollup;
59266
59473
  TokensCompletionRollup;
59474
+ Temperature;
59475
+ TopP;
59476
+ TopK;
59477
+ MinP;
59478
+ FrequencyPenalty;
59479
+ PresencePenalty;
59480
+ Seed;
59481
+ StopSequences;
59482
+ ResponseFormat;
59483
+ LogProbs;
59484
+ TopLogProbs;
59267
59485
  };
59268
59486
  __decorate([
59269
59487
  Field({ nullable: true }),
@@ -59369,6 +59587,50 @@ __decorate([
59369
59587
  Field(() => Int, { nullable: true }),
59370
59588
  __metadata("design:type", Number)
59371
59589
  ], CreateAIPromptRunInput.prototype, "TokensCompletionRollup", void 0);
59590
+ __decorate([
59591
+ Field(() => Float, { nullable: true }),
59592
+ __metadata("design:type", Number)
59593
+ ], CreateAIPromptRunInput.prototype, "Temperature", void 0);
59594
+ __decorate([
59595
+ Field(() => Float, { nullable: true }),
59596
+ __metadata("design:type", Number)
59597
+ ], CreateAIPromptRunInput.prototype, "TopP", void 0);
59598
+ __decorate([
59599
+ Field(() => Int, { nullable: true }),
59600
+ __metadata("design:type", Number)
59601
+ ], CreateAIPromptRunInput.prototype, "TopK", void 0);
59602
+ __decorate([
59603
+ Field(() => Float, { nullable: true }),
59604
+ __metadata("design:type", Number)
59605
+ ], CreateAIPromptRunInput.prototype, "MinP", void 0);
59606
+ __decorate([
59607
+ Field(() => Float, { nullable: true }),
59608
+ __metadata("design:type", Number)
59609
+ ], CreateAIPromptRunInput.prototype, "FrequencyPenalty", void 0);
59610
+ __decorate([
59611
+ Field(() => Float, { nullable: true }),
59612
+ __metadata("design:type", Number)
59613
+ ], CreateAIPromptRunInput.prototype, "PresencePenalty", void 0);
59614
+ __decorate([
59615
+ Field(() => Int, { nullable: true }),
59616
+ __metadata("design:type", Number)
59617
+ ], CreateAIPromptRunInput.prototype, "Seed", void 0);
59618
+ __decorate([
59619
+ Field({ nullable: true }),
59620
+ __metadata("design:type", String)
59621
+ ], CreateAIPromptRunInput.prototype, "StopSequences", void 0);
59622
+ __decorate([
59623
+ Field({ nullable: true }),
59624
+ __metadata("design:type", String)
59625
+ ], CreateAIPromptRunInput.prototype, "ResponseFormat", void 0);
59626
+ __decorate([
59627
+ Field(() => Boolean, { nullable: true }),
59628
+ __metadata("design:type", Boolean)
59629
+ ], CreateAIPromptRunInput.prototype, "LogProbs", void 0);
59630
+ __decorate([
59631
+ Field(() => Int, { nullable: true }),
59632
+ __metadata("design:type", Number)
59633
+ ], CreateAIPromptRunInput.prototype, "TopLogProbs", void 0);
59372
59634
  CreateAIPromptRunInput = __decorate([
59373
59635
  InputType()
59374
59636
  ], CreateAIPromptRunInput);
@@ -59400,6 +59662,17 @@ let UpdateAIPromptRunInput = class UpdateAIPromptRunInput {
59400
59662
  TokensUsedRollup;
59401
59663
  TokensPromptRollup;
59402
59664
  TokensCompletionRollup;
59665
+ Temperature;
59666
+ TopP;
59667
+ TopK;
59668
+ MinP;
59669
+ FrequencyPenalty;
59670
+ PresencePenalty;
59671
+ Seed;
59672
+ StopSequences;
59673
+ ResponseFormat;
59674
+ LogProbs;
59675
+ TopLogProbs;
59403
59676
  OldValues___;
59404
59677
  };
59405
59678
  __decorate([
@@ -59506,6 +59779,50 @@ __decorate([
59506
59779
  Field(() => Int, { nullable: true }),
59507
59780
  __metadata("design:type", Number)
59508
59781
  ], UpdateAIPromptRunInput.prototype, "TokensCompletionRollup", void 0);
59782
+ __decorate([
59783
+ Field(() => Float, { nullable: true }),
59784
+ __metadata("design:type", Number)
59785
+ ], UpdateAIPromptRunInput.prototype, "Temperature", void 0);
59786
+ __decorate([
59787
+ Field(() => Float, { nullable: true }),
59788
+ __metadata("design:type", Number)
59789
+ ], UpdateAIPromptRunInput.prototype, "TopP", void 0);
59790
+ __decorate([
59791
+ Field(() => Int, { nullable: true }),
59792
+ __metadata("design:type", Number)
59793
+ ], UpdateAIPromptRunInput.prototype, "TopK", void 0);
59794
+ __decorate([
59795
+ Field(() => Float, { nullable: true }),
59796
+ __metadata("design:type", Number)
59797
+ ], UpdateAIPromptRunInput.prototype, "MinP", void 0);
59798
+ __decorate([
59799
+ Field(() => Float, { nullable: true }),
59800
+ __metadata("design:type", Number)
59801
+ ], UpdateAIPromptRunInput.prototype, "FrequencyPenalty", void 0);
59802
+ __decorate([
59803
+ Field(() => Float, { nullable: true }),
59804
+ __metadata("design:type", Number)
59805
+ ], UpdateAIPromptRunInput.prototype, "PresencePenalty", void 0);
59806
+ __decorate([
59807
+ Field(() => Int, { nullable: true }),
59808
+ __metadata("design:type", Number)
59809
+ ], UpdateAIPromptRunInput.prototype, "Seed", void 0);
59810
+ __decorate([
59811
+ Field({ nullable: true }),
59812
+ __metadata("design:type", String)
59813
+ ], UpdateAIPromptRunInput.prototype, "StopSequences", void 0);
59814
+ __decorate([
59815
+ Field({ nullable: true }),
59816
+ __metadata("design:type", String)
59817
+ ], UpdateAIPromptRunInput.prototype, "ResponseFormat", void 0);
59818
+ __decorate([
59819
+ Field(() => Boolean, { nullable: true }),
59820
+ __metadata("design:type", Boolean)
59821
+ ], UpdateAIPromptRunInput.prototype, "LogProbs", void 0);
59822
+ __decorate([
59823
+ Field(() => Int, { nullable: true }),
59824
+ __metadata("design:type", Number)
59825
+ ], UpdateAIPromptRunInput.prototype, "TopLogProbs", void 0);
59509
59826
  __decorate([
59510
59827
  Field(() => [KeyValuePairInput], { nullable: true }),
59511
59828
  __metadata("design:type", Array)