@maxim_mazurok/gapi.client.aiplatform-v1beta1 0.2.20250820 → 0.2.20250825
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/index.d.ts +11 -1
- package/package.json +1 -1
package/index.d.ts
CHANGED
|
@@ -9,7 +9,7 @@
|
|
|
9
9
|
// This file was generated by https://github.com/Maxim-Mazurok/google-api-typings-generator. Please do not edit it manually.
|
|
10
10
|
// In case of any problems please post issue to https://github.com/Maxim-Mazurok/google-api-typings-generator
|
|
11
11
|
// Generated from: https://aiplatform.googleapis.com/$discovery/rest?version=v1beta1
|
|
12
|
-
// Revision:
|
|
12
|
+
// Revision: 20250825
|
|
13
13
|
|
|
14
14
|
/// <reference types="gapi.client" />
|
|
15
15
|
|
|
@@ -2298,6 +2298,14 @@ declare namespace gapi.client {
|
|
|
2298
2298
|
/** Optional. The summary metrics for the evaluation run. */
|
|
2299
2299
|
summaryMetrics?: GoogleCloudAiplatformV1beta1SummaryMetrics;
|
|
2300
2300
|
}
|
|
2301
|
+
interface GoogleCloudAiplatformV1beta1EvaluationRubricConfig {
|
|
2302
|
+
/** Dynamically generate rubrics using a predefined spec. */
|
|
2303
|
+
predefinedRubricGenerationSpec?: GoogleCloudAiplatformV1beta1EvaluationRunMetricPredefinedMetricSpec;
|
|
2304
|
+
/** Dynamically generate rubrics using this specification. */
|
|
2305
|
+
rubricGenerationSpec?: GoogleCloudAiplatformV1beta1EvaluationRunMetricRubricGenerationSpec;
|
|
2306
|
+
/** Required. The key used to save the generated rubrics. If a generation spec is provided, this key will be used for the name of the generated rubric group. Otherwise, this key will be used to look up the existing rubric group on the evaluation item. Note that if a rubric group key is specified on both a rubric config and an evaluation metric, the key from the metric will be used to select the rubrics for evaluation. */
|
|
2307
|
+
rubricGroupKey?: string;
|
|
2308
|
+
}
|
|
2301
2309
|
interface GoogleCloudAiplatformV1beta1EvaluationRun {
|
|
2302
2310
|
/** Output only. Time when the evaluation run was completed. */
|
|
2303
2311
|
completionTime?: string;
|
|
@@ -2343,6 +2351,8 @@ declare namespace gapi.client {
|
|
|
2343
2351
|
outputConfig?: GoogleCloudAiplatformV1beta1EvaluationRunEvaluationConfigOutputConfig;
|
|
2344
2352
|
/** The prompt template used for inference. The values for variables in the prompt template are defined in EvaluationItem.EvaluationPrompt.PromptTemplateData.values. */
|
|
2345
2353
|
promptTemplate?: GoogleCloudAiplatformV1beta1EvaluationRunEvaluationConfigPromptTemplate;
|
|
2354
|
+
/** Optional. The rubric configs for the evaluation run. They are used to generate rubrics which can be used by rubric-based metrics. Multiple rubric configs can be specified for rubric generation but only one rubric config can be used for a rubric-based metric. If more than one rubric config is provided, the evaluation metric must specify a rubric group key. Note that if a generation spec is specified on both a rubric config and an evaluation metric, the rubrics generated for the metric will be used for evaluation. */
|
|
2355
|
+
rubricConfigs?: GoogleCloudAiplatformV1beta1EvaluationRubricConfig[];
|
|
2346
2356
|
}
|
|
2347
2357
|
interface GoogleCloudAiplatformV1beta1EvaluationRunEvaluationConfigAutoraterConfig {
|
|
2348
2358
|
/** Optional. The fully qualified name of the publisher model or tuned autorater endpoint to use. Publisher model format: `projects/{project}/locations/{location}/publishers/*/models/*` Tuned model endpoint format: `projects/{project}/locations/{location}/endpoints/{endpoint}` */
|