weave-typescript 0.13.0 → 0.14.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.d.ts +9 -16
- package/dist/index.js +13 -21
- package/dist/weaveapi/project/v1/project.pb.d.ts +67 -5
- package/dist/weaveapi/project/v1/project.pb.js +764 -65
- package/dist/weaveapi/project/v1/service.pb.d.ts +110 -69
- package/dist/weaveapi/project/v1/service.pb.js +1056 -392
- package/dist/weaveapi/task/v1/service.pb.d.ts +171 -0
- package/dist/weaveapi/task/v1/service.pb.js +1416 -0
- package/dist/weaveapi/task/v1/task.pb.d.ts +84 -0
- package/dist/weaveapi/task/v1/task.pb.js +782 -0
- package/dist/weavesql/atcdb/event_sql.d.ts +10 -0
- package/dist/weavesql/atcdb/event_sql.js +11 -1
- package/dist/weavesql/atcdb/mailbox_sql.d.ts +2 -2
- package/dist/weavesql/atcdb/mailbox_sql.js +1 -1
- package/dist/weavesql/weavedb/project_sql.d.ts +121 -46
- package/dist/weavesql/weavedb/project_sql.js +304 -112
- package/dist/weavesql/weavedb/task_sql.d.ts +163 -0
- package/dist/weavesql/weavedb/task_sql.js +328 -0
- package/package.json +1 -1
- package/dist/weaveapi/auth/v1/service.pb.d.ts +0 -709
- package/dist/weaveapi/auth/v1/service.pb.js +0 -5058
- package/dist/weaveapi/auth/v1/session.pb.d.ts +0 -43
- package/dist/weaveapi/auth/v1/session.pb.js +0 -308
- package/dist/weaveapi/auth/v1/usage.pb.d.ts +0 -76
- package/dist/weaveapi/auth/v1/usage.pb.js +0 -849
- package/dist/weaveapi/auth/v1/user.pb.d.ts +0 -141
- package/dist/weaveapi/auth/v1/user.pb.js +0 -1625
- package/dist/weaveapi/data/v1/data.pb.d.ts +0 -54
- package/dist/weaveapi/data/v1/data.pb.js +0 -411
- package/dist/weaveapi/data/v1/service.pb.d.ts +0 -296
- package/dist/weaveapi/data/v1/service.pb.js +0 -2777
- package/dist/weaveapi/generate/v1/configuration.pb.d.ts +0 -122
- package/dist/weaveapi/generate/v1/configuration.pb.js +0 -877
- package/dist/weaveapi/generate/v1/generate.pb.d.ts +0 -48
- package/dist/weaveapi/generate/v1/generate.pb.js +0 -367
- package/dist/weaveapi/generate/v1/service.pb.d.ts +0 -173
- package/dist/weaveapi/generate/v1/service.pb.js +0 -1207
- package/dist/weaveapi/input/v1/input.pb.d.ts +0 -77
- package/dist/weaveapi/input/v1/input.pb.js +0 -632
- package/dist/weaveapi/input/v1/service.pb.d.ts +0 -390
- package/dist/weaveapi/input/v1/service.pb.js +0 -3610
- package/dist/weaveapi/integration/v1/integration.pb.d.ts +0 -67
- package/dist/weaveapi/integration/v1/integration.pb.js +0 -486
- package/dist/weaveapi/integration/v1/service.pb.d.ts +0 -318
- package/dist/weaveapi/integration/v1/service.pb.js +0 -2992
- package/dist/weaveapi/llmx/v1/architecture.pb.d.ts +0 -665
- package/dist/weaveapi/llmx/v1/architecture.pb.js +0 -3028
- package/dist/weaveapi/llmx/v1/capabilities.pb.d.ts +0 -709
- package/dist/weaveapi/llmx/v1/capabilities.pb.js +0 -3458
- package/dist/weaveapi/llmx/v1/model.pb.d.ts +0 -375
- package/dist/weaveapi/llmx/v1/model.pb.js +0 -1472
- package/dist/weaveapi/llmx/v1/pricing.pb.d.ts +0 -340
- package/dist/weaveapi/llmx/v1/pricing.pb.js +0 -974
- package/dist/weaveapi/llmx/v1/provider.pb.d.ts +0 -72
- package/dist/weaveapi/llmx/v1/provider.pb.js +0 -218
- package/dist/weaveapi/llmx/v1/service.pb.d.ts +0 -683
- package/dist/weaveapi/llmx/v1/service.pb.js +0 -5311
- package/dist/weaveapi/mcpregistry/v1/server.pb.d.ts +0 -54
- package/dist/weaveapi/mcpregistry/v1/server.pb.js +0 -436
- package/dist/weaveapi/mcpregistry/v1/service.pb.d.ts +0 -171
- package/dist/weaveapi/mcpregistry/v1/service.pb.js +0 -963
- package/dist/weaveapi/payment/v1/invoice.pb.d.ts +0 -92
- package/dist/weaveapi/payment/v1/invoice.pb.js +0 -998
- package/dist/weaveapi/payment/v1/service.pb.d.ts +0 -662
- package/dist/weaveapi/payment/v1/service.pb.js +0 -5395
- package/dist/weaveapi/payment/v1/subscription.pb.d.ts +0 -165
- package/dist/weaveapi/payment/v1/subscription.pb.js +0 -1930
- package/dist/weaveapi/report/v1/report.pb.d.ts +0 -81
- package/dist/weaveapi/report/v1/report.pb.js +0 -678
- package/dist/weaveapi/report/v1/service.pb.d.ts +0 -194
- package/dist/weaveapi/report/v1/service.pb.js +0 -1923
- package/dist/weaveapi/requirement/v1/requirement.pb.d.ts +0 -70
- package/dist/weaveapi/requirement/v1/requirement.pb.js +0 -570
- package/dist/weaveapi/requirement/v1/service.pb.d.ts +0 -390
- package/dist/weaveapi/requirement/v1/service.pb.js +0 -3636
- package/dist/weaveapi/run/v1/run.pb.d.ts +0 -117
- package/dist/weaveapi/run/v1/run.pb.js +0 -1155
- package/dist/weaveapi/run/v1/service.pb.d.ts +0 -229
- package/dist/weaveapi/run/v1/service.pb.js +0 -2426
- package/dist/weaveapi/script/v1/script.pb.d.ts +0 -102
- package/dist/weaveapi/script/v1/script.pb.js +0 -1025
- package/dist/weaveapi/script/v1/service.pb.d.ts +0 -502
- package/dist/weaveapi/script/v1/service.pb.js +0 -4871
- package/dist/weaveapi/storage/v1/auth.pb.d.ts +0 -27
- package/dist/weaveapi/storage/v1/auth.pb.js +0 -110
- package/dist/weaveapi/storage/v1/nosql_database.pb.d.ts +0 -91
- package/dist/weaveapi/storage/v1/nosql_database.pb.js +0 -1056
- package/dist/weaveapi/storage/v1/object_store.pb.d.ts +0 -49
- package/dist/weaveapi/storage/v1/object_store.pb.js +0 -441
- package/dist/weaveapi/storage/v1/service.pb.d.ts +0 -245
- package/dist/weaveapi/storage/v1/service.pb.js +0 -1339
- package/dist/weaveapi/storage/v1/sql_database.pb.d.ts +0 -98
- package/dist/weaveapi/storage/v1/sql_database.pb.js +0 -1184
- package/dist/weaveapi/storage/v1/storage.pb.d.ts +0 -35
- package/dist/weaveapi/storage/v1/storage.pb.js +0 -198
- package/dist/weaveapi/storage/v1/vcs.pb.d.ts +0 -75
- package/dist/weaveapi/storage/v1/vcs.pb.js +0 -864
- package/dist/weaveapi/suite/v1/service.pb.d.ts +0 -539
- package/dist/weaveapi/suite/v1/service.pb.js +0 -5279
- package/dist/weaveapi/suite/v1/suite.pb.d.ts +0 -92
- package/dist/weaveapi/suite/v1/suite.pb.js +0 -876
- package/dist/weaveapi/synthesize/v1/dataset.pb.d.ts +0 -68
- package/dist/weaveapi/synthesize/v1/dataset.pb.js +0 -460
- package/dist/weaveapi/synthesize/v1/inline_data.pb.d.ts +0 -45
- package/dist/weaveapi/synthesize/v1/inline_data.pb.js +0 -166
- package/dist/weaveapi/synthesize/v1/relationship.pb.d.ts +0 -58
- package/dist/weaveapi/synthesize/v1/relationship.pb.js +0 -253
- package/dist/weaveapi/synthesize/v1/service.pb.d.ts +0 -144
- package/dist/weaveapi/synthesize/v1/service.pb.js +0 -937
- package/dist/weaveapi/synthesize/v1/training.pb.d.ts +0 -58
- package/dist/weaveapi/synthesize/v1/training.pb.js +0 -386
- package/dist/weaveapi/testcase/v1/service.pb.d.ts +0 -498
- package/dist/weaveapi/testcase/v1/service.pb.js +0 -4728
- package/dist/weaveapi/testcase/v1/testcase.pb.d.ts +0 -91
- package/dist/weaveapi/testcase/v1/testcase.pb.js +0 -794
- package/dist/weaveapi/workflow/v1/service.pb.d.ts +0 -261
- package/dist/weaveapi/workflow/v1/service.pb.js +0 -2120
- package/dist/weaveapi/workflow/v1/workflow.pb.d.ts +0 -77
- package/dist/weaveapi/workflow/v1/workflow.pb.js +0 -555
- package/dist/weaveapi/workflowdata/v1/service.pb.d.ts +0 -295
- package/dist/weaveapi/workflowdata/v1/service.pb.js +0 -2612
|
@@ -1,665 +0,0 @@
|
|
|
1
|
-
import { BinaryReader, BinaryWriter } from "@bufbuild/protobuf/wire";
|
|
2
|
-
import { DataFormat } from "./capabilities.pb";
|
|
3
|
-
export declare const protobufPackage = "weaveapi.llmx.v1";
|
|
4
|
-
/** Base architecture types */
|
|
5
|
-
export declare enum BaseArchitecture {
|
|
6
|
-
BASE_ARCHITECTURE_UNSPECIFIED = 0,
|
|
7
|
-
BASE_ARCHITECTURE_TRANSFORMER = 1,
|
|
8
|
-
BASE_ARCHITECTURE_MAMBA = 2,
|
|
9
|
-
BASE_ARCHITECTURE_HYBRID = 3,
|
|
10
|
-
BASE_ARCHITECTURE_RECURRENT_NEURAL_NETWORK = 4,
|
|
11
|
-
BASE_ARCHITECTURE_CONVOLUTIONAL_NEURAL_NETWORK = 5,
|
|
12
|
-
BASE_ARCHITECTURE_DIFFUSION = 6,
|
|
13
|
-
BASE_ARCHITECTURE_GENERATIVE_ADVERSARIAL_NETWORK = 7,
|
|
14
|
-
UNRECOGNIZED = -1
|
|
15
|
-
}
|
|
16
|
-
export declare function baseArchitectureFromJSON(object: any): BaseArchitecture;
|
|
17
|
-
export declare function baseArchitectureToJSON(object: BaseArchitecture): string;
|
|
18
|
-
/** Model architecture types */
|
|
19
|
-
export declare enum ModelArchitecture {
|
|
20
|
-
MODEL_ARCHITECTURE_UNSPECIFIED = 0,
|
|
21
|
-
MODEL_ARCHITECTURE_GPT = 1,
|
|
22
|
-
MODEL_ARCHITECTURE_BERT = 2,
|
|
23
|
-
MODEL_ARCHITECTURE_T5 = 3,
|
|
24
|
-
MODEL_ARCHITECTURE_LLAMA = 4,
|
|
25
|
-
MODEL_ARCHITECTURE_MISTRAL = 5,
|
|
26
|
-
MODEL_ARCHITECTURE_GEMMA = 6,
|
|
27
|
-
MODEL_ARCHITECTURE_QWEN = 7,
|
|
28
|
-
MODEL_ARCHITECTURE_PHI = 8,
|
|
29
|
-
MODEL_ARCHITECTURE_CLAUDE = 9,
|
|
30
|
-
MODEL_ARCHITECTURE_COMMAND = 10,
|
|
31
|
-
MODEL_ARCHITECTURE_PALM = 11,
|
|
32
|
-
MODEL_ARCHITECTURE_FALCON = 12,
|
|
33
|
-
MODEL_ARCHITECTURE_STARCODER = 13,
|
|
34
|
-
MODEL_ARCHITECTURE_CODEGEN = 14,
|
|
35
|
-
MODEL_ARCHITECTURE_DEEPSEEK = 15,
|
|
36
|
-
MODEL_ARCHITECTURE_YI = 16,
|
|
37
|
-
MODEL_ARCHITECTURE_MIXTRAL = 17,
|
|
38
|
-
MODEL_ARCHITECTURE_GEMINI = 18,
|
|
39
|
-
UNRECOGNIZED = -1
|
|
40
|
-
}
|
|
41
|
-
export declare function modelArchitectureFromJSON(object: any): ModelArchitecture;
|
|
42
|
-
export declare function modelArchitectureToJSON(object: ModelArchitecture): string;
|
|
43
|
-
/** Position embedding types */
|
|
44
|
-
export declare enum PositionEmbedding {
|
|
45
|
-
POSITION_EMBEDDING_UNSPECIFIED = 0,
|
|
46
|
-
POSITION_EMBEDDING_ABSOLUTE = 1,
|
|
47
|
-
POSITION_EMBEDDING_RELATIVE = 2,
|
|
48
|
-
POSITION_EMBEDDING_ROTARY = 3,
|
|
49
|
-
POSITION_EMBEDDING_ALIBI = 4,
|
|
50
|
-
POSITION_EMBEDDING_LEARNED = 5,
|
|
51
|
-
POSITION_EMBEDDING_SINUSOIDAL = 6,
|
|
52
|
-
UNRECOGNIZED = -1
|
|
53
|
-
}
|
|
54
|
-
export declare function positionEmbeddingFromJSON(object: any): PositionEmbedding;
|
|
55
|
-
export declare function positionEmbeddingToJSON(object: PositionEmbedding): string;
|
|
56
|
-
/** Activation function types */
|
|
57
|
-
export declare enum ActivationFunction {
|
|
58
|
-
ACTIVATION_FUNCTION_UNSPECIFIED = 0,
|
|
59
|
-
ACTIVATION_FUNCTION_GELU = 1,
|
|
60
|
-
ACTIVATION_FUNCTION_SWIGLU = 2,
|
|
61
|
-
ACTIVATION_FUNCTION_RELU = 3,
|
|
62
|
-
ACTIVATION_FUNCTION_SILU = 4,
|
|
63
|
-
ACTIVATION_FUNCTION_TANH = 5,
|
|
64
|
-
ACTIVATION_FUNCTION_SIGMOID = 6,
|
|
65
|
-
ACTIVATION_FUNCTION_MISH = 7,
|
|
66
|
-
ACTIVATION_FUNCTION_LEAKY_RELU = 8,
|
|
67
|
-
UNRECOGNIZED = -1
|
|
68
|
-
}
|
|
69
|
-
export declare function activationFunctionFromJSON(object: any): ActivationFunction;
|
|
70
|
-
export declare function activationFunctionToJSON(object: ActivationFunction): string;
|
|
71
|
-
/** Training technique types */
|
|
72
|
-
export declare enum TrainingTechnique {
|
|
73
|
-
TRAINING_TECHNIQUE_UNSPECIFIED = 0,
|
|
74
|
-
TRAINING_TECHNIQUE_SUPERVISED = 1,
|
|
75
|
-
TRAINING_TECHNIQUE_REINFORCEMENT_LEARNING_FROM_HUMAN_FEEDBACK = 2,
|
|
76
|
-
TRAINING_TECHNIQUE_CONSTITUTIONAL_AI = 3,
|
|
77
|
-
TRAINING_TECHNIQUE_DIRECT_PREFERENCE_OPTIMIZATION = 4,
|
|
78
|
-
TRAINING_TECHNIQUE_PROXIMAL_POLICY_OPTIMIZATION = 5,
|
|
79
|
-
TRAINING_TECHNIQUE_SUPERVISED_FINE_TUNING = 6,
|
|
80
|
-
TRAINING_TECHNIQUE_INSTRUCTION_TUNING = 7,
|
|
81
|
-
TRAINING_TECHNIQUE_FEW_SHOT = 8,
|
|
82
|
-
TRAINING_TECHNIQUE_CHAIN_OF_THOUGHT = 9,
|
|
83
|
-
UNRECOGNIZED = -1
|
|
84
|
-
}
|
|
85
|
-
export declare function trainingTechniqueFromJSON(object: any): TrainingTechnique;
|
|
86
|
-
export declare function trainingTechniqueToJSON(object: TrainingTechnique): string;
|
|
87
|
-
/** Quantization types */
|
|
88
|
-
export declare enum Quantization {
|
|
89
|
-
QUANTIZATION_UNSPECIFIED = 0,
|
|
90
|
-
QUANTIZATION_NONE = 1,
|
|
91
|
-
QUANTIZATION_INT8 = 2,
|
|
92
|
-
QUANTIZATION_INT4 = 3,
|
|
93
|
-
QUANTIZATION_FP8 = 4,
|
|
94
|
-
QUANTIZATION_GPTQ = 5,
|
|
95
|
-
QUANTIZATION_AWQ = 6,
|
|
96
|
-
QUANTIZATION_GGUF = 7,
|
|
97
|
-
QUANTIZATION_BITSANDBYTES = 8,
|
|
98
|
-
UNRECOGNIZED = -1
|
|
99
|
-
}
|
|
100
|
-
export declare function quantizationFromJSON(object: any): Quantization;
|
|
101
|
-
export declare function quantizationToJSON(object: Quantization): string;
|
|
102
|
-
/** Precision types */
|
|
103
|
-
export declare enum Precision {
|
|
104
|
-
PRECISION_UNSPECIFIED = 0,
|
|
105
|
-
PRECISION_FP32 = 1,
|
|
106
|
-
PRECISION_FP16 = 2,
|
|
107
|
-
PRECISION_BF16 = 3,
|
|
108
|
-
PRECISION_INT8 = 4,
|
|
109
|
-
PRECISION_MIXED = 5,
|
|
110
|
-
UNRECOGNIZED = -1
|
|
111
|
-
}
|
|
112
|
-
export declare function precisionFromJSON(object: any): Precision;
|
|
113
|
-
export declare function precisionToJSON(object: Precision): string;
|
|
114
|
-
/** Inference framework types */
|
|
115
|
-
export declare enum InferenceFramework {
|
|
116
|
-
INFERENCE_FRAMEWORK_UNSPECIFIED = 0,
|
|
117
|
-
INFERENCE_FRAMEWORK_VLLM = 1,
|
|
118
|
-
INFERENCE_FRAMEWORK_TGI = 2,
|
|
119
|
-
INFERENCE_FRAMEWORK_TRITON = 3,
|
|
120
|
-
INFERENCE_FRAMEWORK_TENSORRT = 4,
|
|
121
|
-
INFERENCE_FRAMEWORK_ONNX = 5,
|
|
122
|
-
INFERENCE_FRAMEWORK_TORCHSERVE = 6,
|
|
123
|
-
INFERENCE_FRAMEWORK_TENSORFLOW_SERVING = 7,
|
|
124
|
-
INFERENCE_FRAMEWORK_RAY_SERVE = 8,
|
|
125
|
-
INFERENCE_FRAMEWORK_DEEPSPEED = 9,
|
|
126
|
-
INFERENCE_FRAMEWORK_FASTERTRANSFORMER = 10,
|
|
127
|
-
UNRECOGNIZED = -1
|
|
128
|
-
}
|
|
129
|
-
export declare function inferenceFrameworkFromJSON(object: any): InferenceFramework;
|
|
130
|
-
export declare function inferenceFrameworkToJSON(object: InferenceFramework): string;
|
|
131
|
-
/** Model format types */
|
|
132
|
-
export declare enum ModelFormat {
|
|
133
|
-
MODEL_FORMAT_UNSPECIFIED = 0,
|
|
134
|
-
MODEL_FORMAT_SAFETENSORS = 1,
|
|
135
|
-
MODEL_FORMAT_GGUF = 2,
|
|
136
|
-
MODEL_FORMAT_ONNX = 3,
|
|
137
|
-
MODEL_FORMAT_PYTORCH = 4,
|
|
138
|
-
MODEL_FORMAT_TENSORFLOW = 5,
|
|
139
|
-
MODEL_FORMAT_JAX = 6,
|
|
140
|
-
MODEL_FORMAT_COREML = 7,
|
|
141
|
-
MODEL_FORMAT_OPENVINO = 8,
|
|
142
|
-
MODEL_FORMAT_TENSORRT = 9,
|
|
143
|
-
UNRECOGNIZED = -1
|
|
144
|
-
}
|
|
145
|
-
export declare function modelFormatFromJSON(object: any): ModelFormat;
|
|
146
|
-
export declare function modelFormatToJSON(object: ModelFormat): string;
|
|
147
|
-
/** Checkpoint format types */
|
|
148
|
-
export declare enum CheckpointFormat {
|
|
149
|
-
CHECKPOINT_FORMAT_UNSPECIFIED = 0,
|
|
150
|
-
CHECKPOINT_FORMAT_PYTORCH = 1,
|
|
151
|
-
CHECKPOINT_FORMAT_TENSORFLOW = 2,
|
|
152
|
-
CHECKPOINT_FORMAT_SAFETENSORS = 3,
|
|
153
|
-
CHECKPOINT_FORMAT_HF = 4,
|
|
154
|
-
CHECKPOINT_FORMAT_MEGATRON = 5,
|
|
155
|
-
CHECKPOINT_FORMAT_DEEPSPEED = 6,
|
|
156
|
-
UNRECOGNIZED = -1
|
|
157
|
-
}
|
|
158
|
-
export declare function checkpointFormatFromJSON(object: any): CheckpointFormat;
|
|
159
|
-
export declare function checkpointFormatToJSON(object: CheckpointFormat): string;
|
|
160
|
-
/** GPU types */
|
|
161
|
-
export declare enum GPUType {
|
|
162
|
-
GPU_TYPE_UNSPECIFIED = 0,
|
|
163
|
-
/** GPU_TYPE_A100 - Datacenter/Enterprise GPUs */
|
|
164
|
-
GPU_TYPE_A100 = 1,
|
|
165
|
-
GPU_TYPE_H100 = 2,
|
|
166
|
-
GPU_TYPE_V100 = 3,
|
|
167
|
-
GPU_TYPE_T4 = 4,
|
|
168
|
-
GPU_TYPE_L4 = 5,
|
|
169
|
-
GPU_TYPE_L40 = 6,
|
|
170
|
-
GPU_TYPE_A10 = 7,
|
|
171
|
-
GPU_TYPE_A40 = 8,
|
|
172
|
-
GPU_TYPE_RTX_A6000 = 9,
|
|
173
|
-
GPU_TYPE_RTX_A5000 = 10,
|
|
174
|
-
GPU_TYPE_RTX_A4000 = 11,
|
|
175
|
-
/** GPU_TYPE_RTX_4090 - Consumer GPUs (40 series) */
|
|
176
|
-
GPU_TYPE_RTX_4090 = 20,
|
|
177
|
-
GPU_TYPE_RTX_4080 = 21,
|
|
178
|
-
GPU_TYPE_RTX_4070_TI = 22,
|
|
179
|
-
GPU_TYPE_RTX_4070 = 23,
|
|
180
|
-
GPU_TYPE_RTX_4060_TI = 24,
|
|
181
|
-
GPU_TYPE_RTX_4060 = 25,
|
|
182
|
-
/** GPU_TYPE_RTX_3090_TI - Consumer GPUs (30 series) */
|
|
183
|
-
GPU_TYPE_RTX_3090_TI = 30,
|
|
184
|
-
GPU_TYPE_RTX_3090 = 31,
|
|
185
|
-
GPU_TYPE_RTX_3080_TI = 32,
|
|
186
|
-
GPU_TYPE_RTX_3080 = 33,
|
|
187
|
-
GPU_TYPE_RTX_3070_TI = 34,
|
|
188
|
-
GPU_TYPE_RTX_3070 = 35,
|
|
189
|
-
GPU_TYPE_RTX_3060_TI = 36,
|
|
190
|
-
GPU_TYPE_RTX_3060 = 37,
|
|
191
|
-
/** GPU_TYPE_MI300 - AMD GPUs */
|
|
192
|
-
GPU_TYPE_MI300 = 50,
|
|
193
|
-
GPU_TYPE_MI250 = 51,
|
|
194
|
-
GPU_TYPE_MI210 = 52,
|
|
195
|
-
GPU_TYPE_RX_7900_XTX = 53,
|
|
196
|
-
GPU_TYPE_RX_7900_XT = 54,
|
|
197
|
-
GPU_TYPE_RX_6900_XT = 55,
|
|
198
|
-
/** GPU_TYPE_M3_MAX - Apple Silicon (unified memory) */
|
|
199
|
-
GPU_TYPE_M3_MAX = 60,
|
|
200
|
-
GPU_TYPE_M3_PRO = 61,
|
|
201
|
-
GPU_TYPE_M3 = 62,
|
|
202
|
-
GPU_TYPE_M2_ULTRA = 63,
|
|
203
|
-
GPU_TYPE_M2_MAX = 64,
|
|
204
|
-
GPU_TYPE_M2_PRO = 65,
|
|
205
|
-
GPU_TYPE_M2 = 66,
|
|
206
|
-
GPU_TYPE_M1_ULTRA = 67,
|
|
207
|
-
GPU_TYPE_M1_MAX = 68,
|
|
208
|
-
GPU_TYPE_M1_PRO = 69,
|
|
209
|
-
GPU_TYPE_M1 = 70,
|
|
210
|
-
/** GPU_TYPE_RTX_2080_TI - Older but still relevant */
|
|
211
|
-
GPU_TYPE_RTX_2080_TI = 80,
|
|
212
|
-
GPU_TYPE_TITAN_RTX = 81,
|
|
213
|
-
GPU_TYPE_GTX_1080_TI = 82,
|
|
214
|
-
UNRECOGNIZED = -1
|
|
215
|
-
}
|
|
216
|
-
export declare function gPUTypeFromJSON(object: any): GPUType;
|
|
217
|
-
export declare function gPUTypeToJSON(object: GPUType): string;
|
|
218
|
-
/** CPU architecture types */
|
|
219
|
-
export declare enum CPUType {
|
|
220
|
-
CPU_TYPE_UNSPECIFIED = 0,
|
|
221
|
-
CPU_TYPE_X86_64 = 1,
|
|
222
|
-
CPU_TYPE_ARM64 = 2,
|
|
223
|
-
CPU_TYPE_APPLE_SILICON = 3,
|
|
224
|
-
CPU_TYPE_AMD64 = 4,
|
|
225
|
-
CPU_TYPE_GRAVITON = 5,
|
|
226
|
-
UNRECOGNIZED = -1
|
|
227
|
-
}
|
|
228
|
-
export declare function cPUTypeFromJSON(object: any): CPUType;
|
|
229
|
-
export declare function cPUTypeToJSON(object: CPUType): string;
|
|
230
|
-
/** Moderation level */
|
|
231
|
-
export declare enum ModerationLevel {
|
|
232
|
-
MODERATION_LEVEL_UNSPECIFIED = 0,
|
|
233
|
-
MODERATION_LEVEL_NONE = 1,
|
|
234
|
-
MODERATION_LEVEL_LOW = 2,
|
|
235
|
-
MODERATION_LEVEL_MEDIUM = 3,
|
|
236
|
-
MODERATION_LEVEL_HIGH = 4,
|
|
237
|
-
MODERATION_LEVEL_STRICT = 5,
|
|
238
|
-
UNRECOGNIZED = -1
|
|
239
|
-
}
|
|
240
|
-
export declare function moderationLevelFromJSON(object: any): ModerationLevel;
|
|
241
|
-
export declare function moderationLevelToJSON(object: ModerationLevel): string;
|
|
242
|
-
/** License types */
|
|
243
|
-
export declare enum LicenseType {
|
|
244
|
-
LICENSE_TYPE_UNSPECIFIED = 0,
|
|
245
|
-
LICENSE_TYPE_MIT = 1,
|
|
246
|
-
LICENSE_TYPE_APACHE_2_0 = 2,
|
|
247
|
-
LICENSE_TYPE_GPL = 3,
|
|
248
|
-
LICENSE_TYPE_BSD = 4,
|
|
249
|
-
LICENSE_TYPE_COMMERCIAL = 5,
|
|
250
|
-
LICENSE_TYPE_PROPRIETARY = 6,
|
|
251
|
-
LICENSE_TYPE_CUSTOM = 7,
|
|
252
|
-
LICENSE_TYPE_CC_BY = 8,
|
|
253
|
-
LICENSE_TYPE_CC_BY_NC = 9,
|
|
254
|
-
LICENSE_TYPE_CC_BY_SA = 10,
|
|
255
|
-
LICENSE_TYPE_LLAMA = 11,
|
|
256
|
-
LICENSE_TYPE_OPENAI = 12,
|
|
257
|
-
UNRECOGNIZED = -1
|
|
258
|
-
}
|
|
259
|
-
export declare function licenseTypeFromJSON(object: any): LicenseType;
|
|
260
|
-
export declare function licenseTypeToJSON(object: LicenseType): string;
|
|
261
|
-
/** Architecture describes the technical architecture of an AI model. */
|
|
262
|
-
export interface Architecture {
|
|
263
|
-
/**
|
|
264
|
-
* Fundamental architecture type (Transformer, Mamba, etc.).
|
|
265
|
-
* Example: BASE_ARCHITECTURE_TRANSFORMER for GPT models
|
|
266
|
-
*/
|
|
267
|
-
baseArchitecture: BaseArchitecture;
|
|
268
|
-
/**
|
|
269
|
-
* Specific model family/variant.
|
|
270
|
-
* Example: MODEL_ARCHITECTURE_GPT for GPT-4, MODEL_ARCHITECTURE_LLAMA for LLaMA
|
|
271
|
-
*/
|
|
272
|
-
modelArchitecture: ModelArchitecture;
|
|
273
|
-
/**
|
|
274
|
-
* Total learnable parameters in billions.
|
|
275
|
-
* Example: 175 for GPT-3 (175B parameters)
|
|
276
|
-
*/
|
|
277
|
-
parameterCount: number;
|
|
278
|
-
/**
|
|
279
|
-
* Parameters activated per forward pass (for MoE models).
|
|
280
|
-
* Example: 8B active out of 141B total for Mixtral-8x7B
|
|
281
|
-
*/
|
|
282
|
-
activeParameters: number;
|
|
283
|
-
/**
|
|
284
|
-
* Total parameters including non-trainable (embeddings, etc.).
|
|
285
|
-
* May be slightly higher than parameter_count
|
|
286
|
-
*/
|
|
287
|
-
totalParameters: number;
|
|
288
|
-
/**
|
|
289
|
-
* Number of transformer/attention layers.
|
|
290
|
-
* Example: 96 for GPT-3, 32 for 7B models
|
|
291
|
-
*/
|
|
292
|
-
layerCount: number;
|
|
293
|
-
/**
|
|
294
|
-
* Hidden dimension size of the model.
|
|
295
|
-
* Example: 12288 for GPT-3, 4096 for smaller models
|
|
296
|
-
*/
|
|
297
|
-
hiddenSize: number;
|
|
298
|
-
/**
|
|
299
|
-
* Number of attention heads in multi-head attention.
|
|
300
|
-
* Example: 96 for GPT-3, 32 for 7B models
|
|
301
|
-
*/
|
|
302
|
-
attentionHeads: number;
|
|
303
|
-
/**
|
|
304
|
-
* Size of the token vocabulary.
|
|
305
|
-
* Example: 50257 for GPT-2/3, 32000 for LLaMA
|
|
306
|
-
*/
|
|
307
|
-
vocabularySize: number;
|
|
308
|
-
/**
|
|
309
|
-
* Type of position encoding used.
|
|
310
|
-
* Example: POSITION_EMBEDDING_ROTARY for modern models (RoPE)
|
|
311
|
-
*/
|
|
312
|
-
positionEmbedding: PositionEmbedding;
|
|
313
|
-
/**
|
|
314
|
-
* Activation function in feed-forward layers.
|
|
315
|
-
* Example: ACTIVATION_FUNCTION_SWIGLU for LLaMA models
|
|
316
|
-
*/
|
|
317
|
-
activationFunction: ActivationFunction;
|
|
318
|
-
/**
|
|
319
|
-
* Whether this is a Mixture of Experts model.
|
|
320
|
-
* Example: true for Mixtral, GPT-4 (rumored), false for dense models
|
|
321
|
-
*/
|
|
322
|
-
isMixtureOfExperts: boolean;
|
|
323
|
-
/**
|
|
324
|
-
* Total number of expert networks (for MoE).
|
|
325
|
-
* Example: 8 for Mixtral-8x7B
|
|
326
|
-
*/
|
|
327
|
-
expertCount: number;
|
|
328
|
-
/**
|
|
329
|
-
* Number of experts activated per token (for MoE).
|
|
330
|
-
* Example: 2 for Mixtral (2 experts per token out of 8)
|
|
331
|
-
*/
|
|
332
|
-
expertsPerToken: number;
|
|
333
|
-
}
|
|
334
|
-
/** Training contains information about how the model was trained. */
|
|
335
|
-
export interface Training {
|
|
336
|
-
/**
|
|
337
|
-
* Size of training dataset.
|
|
338
|
-
* Examples: "1T tokens", "45TB text", "100B tokens"
|
|
339
|
-
*/
|
|
340
|
-
trainingDataSize: string;
|
|
341
|
-
/**
|
|
342
|
-
* List of data sources used for training.
|
|
343
|
-
* Examples: ["Common Crawl", "Wikipedia", "Books", "GitHub", "ArXiv"]
|
|
344
|
-
*/
|
|
345
|
-
trainingDataSources: string[];
|
|
346
|
-
/**
|
|
347
|
-
* Description of data mixture/proportions.
|
|
348
|
-
* Example: "60% web, 20% books, 10% code, 10% reference"
|
|
349
|
-
*/
|
|
350
|
-
dataMixture: string;
|
|
351
|
-
/**
|
|
352
|
-
* Total training time.
|
|
353
|
-
* Examples: "3 months", "6 weeks", "90 days"
|
|
354
|
-
*/
|
|
355
|
-
trainingDuration: string;
|
|
356
|
-
/**
|
|
357
|
-
* Hardware used for training.
|
|
358
|
-
* Examples: "10000 H100 GPUs", "512 A100 80GB", "TPU v4 pods"
|
|
359
|
-
*/
|
|
360
|
-
trainingHardware: string;
|
|
361
|
-
/**
|
|
362
|
-
* Estimated training cost.
|
|
363
|
-
* Examples: "$100M", "$4.6M", "Not disclosed"
|
|
364
|
-
*/
|
|
365
|
-
trainingCost: string;
|
|
366
|
-
/**
|
|
367
|
-
* Primary training technique used.
|
|
368
|
-
* Example: TRAINING_TECHNIQUE_REINFORCEMENT_LEARNING_FROM_HUMAN_FEEDBACK
|
|
369
|
-
*/
|
|
370
|
-
trainingTechnique: TrainingTechnique;
|
|
371
|
-
/**
|
|
372
|
-
* Training batch size per step.
|
|
373
|
-
* Example: 2048, 4096
|
|
374
|
-
*/
|
|
375
|
-
batchSize: number;
|
|
376
|
-
/**
|
|
377
|
-
* Peak learning rate used.
|
|
378
|
-
* Example: 0.0001, 3e-4
|
|
379
|
-
*/
|
|
380
|
-
learningRate: number;
|
|
381
|
-
/**
|
|
382
|
-
* Total number of training steps/iterations.
|
|
383
|
-
* Example: 1000000
|
|
384
|
-
*/
|
|
385
|
-
trainingSteps: number;
|
|
386
|
-
/**
|
|
387
|
-
* Number of warmup steps for learning rate schedule.
|
|
388
|
-
* Example: 2000
|
|
389
|
-
*/
|
|
390
|
-
warmupSteps: number;
|
|
391
|
-
/**
|
|
392
|
-
* Whether model supports fine-tuning via API.
|
|
393
|
-
* Example: true for GPT-3.5, false for GPT-4
|
|
394
|
-
*/
|
|
395
|
-
fineTuningAvailable: boolean;
|
|
396
|
-
/**
|
|
397
|
-
* Minimum training examples required for fine-tuning.
|
|
398
|
-
* Example: 10 for OpenAI models
|
|
399
|
-
*/
|
|
400
|
-
minFineTuneExamples: number;
|
|
401
|
-
/**
|
|
402
|
-
* Maximum training examples allowed for fine-tuning.
|
|
403
|
-
* Example: 100000 for GPT-3.5
|
|
404
|
-
*/
|
|
405
|
-
maxFineTuneExamples: number;
|
|
406
|
-
/**
|
|
407
|
-
* Supported data formats for fine-tuning.
|
|
408
|
-
* Example: [DATA_FORMAT_JSONL, DATA_FORMAT_CSV]
|
|
409
|
-
*/
|
|
410
|
-
fineTuneFormats: DataFormat[];
|
|
411
|
-
}
|
|
412
|
-
/** TechnicalSpecs contains hardware and software requirements for running the model. */
|
|
413
|
-
export interface TechnicalSpecs {
|
|
414
|
-
/**
|
|
415
|
-
* Minimum system RAM required in GB.
|
|
416
|
-
* Example: 32 for 7B models, 64 for 13B models
|
|
417
|
-
*/
|
|
418
|
-
minMemoryGb: number;
|
|
419
|
-
/**
|
|
420
|
-
* Recommended system RAM for optimal performance in GB.
|
|
421
|
-
* Example: 64 for 7B models, 128 for 13B models
|
|
422
|
-
*/
|
|
423
|
-
recommendedMemoryGb: number;
|
|
424
|
-
/**
|
|
425
|
-
* Minimum GPU VRAM required in GB.
|
|
426
|
-
* Example: 24 for 7B fp16, 48 for 13B fp16
|
|
427
|
-
*/
|
|
428
|
-
minGpuMemoryGb: number;
|
|
429
|
-
/**
|
|
430
|
-
* Disk storage required for model files in GB.
|
|
431
|
-
* Example: 15 for 7B models, 30 for 13B models
|
|
432
|
-
*/
|
|
433
|
-
storageGb: number;
|
|
434
|
-
/**
|
|
435
|
-
* Quantization method if applicable.
|
|
436
|
-
* Example: QUANTIZATION_INT8 for 8-bit quantized models
|
|
437
|
-
*/
|
|
438
|
-
quantization: Quantization;
|
|
439
|
-
/**
|
|
440
|
-
* Numerical precision used.
|
|
441
|
-
* Example: PRECISION_FP16 for half-precision inference
|
|
442
|
-
*/
|
|
443
|
-
precision: Precision;
|
|
444
|
-
/**
|
|
445
|
-
* Optimization level/profile.
|
|
446
|
-
* Examples: "O3", "fast", "balanced", "memory-optimized"
|
|
447
|
-
*/
|
|
448
|
-
optimizationLevel: string;
|
|
449
|
-
/**
|
|
450
|
-
* List of compatible GPU types.
|
|
451
|
-
* Example: [GPU_TYPE_A100, GPU_TYPE_H100, GPU_TYPE_RTX_4090]
|
|
452
|
-
*/
|
|
453
|
-
supportedGpus: GPUType[];
|
|
454
|
-
/**
|
|
455
|
-
* Whether CUDA is required for GPU inference.
|
|
456
|
-
* Example: true for NVIDIA GPUs, false for CPU-only
|
|
457
|
-
*/
|
|
458
|
-
requiresCuda: boolean;
|
|
459
|
-
/**
|
|
460
|
-
* Minimum CUDA version required.
|
|
461
|
-
* Examples: "11.8", "12.0"
|
|
462
|
-
*/
|
|
463
|
-
cudaVersion: string;
|
|
464
|
-
/**
|
|
465
|
-
* Recommended inference framework.
|
|
466
|
-
* Example: INFERENCE_FRAMEWORK_VLLM for high-throughput serving
|
|
467
|
-
*/
|
|
468
|
-
inferenceFramework: InferenceFramework;
|
|
469
|
-
/**
|
|
470
|
-
* Format of distributed model files.
|
|
471
|
-
* Example: MODEL_FORMAT_SAFETENSORS for HuggingFace models
|
|
472
|
-
*/
|
|
473
|
-
modelFormat: ModelFormat;
|
|
474
|
-
/**
|
|
475
|
-
* Total size of model files in GB.
|
|
476
|
-
* Example: 13.5 for 7B model in fp16
|
|
477
|
-
*/
|
|
478
|
-
modelSizeGb: number;
|
|
479
|
-
/**
|
|
480
|
-
* Format of model checkpoint files.
|
|
481
|
-
* Example: CHECKPOINT_FORMAT_PYTORCH for .pt files
|
|
482
|
-
*/
|
|
483
|
-
checkpointFormat: CheckpointFormat;
|
|
484
|
-
/**
|
|
485
|
-
* Number of model shard files.
|
|
486
|
-
* Example: 2 for models split across multiple files
|
|
487
|
-
*/
|
|
488
|
-
numberOfFiles: number;
|
|
489
|
-
/**
|
|
490
|
-
* List of compatible CPU architectures.
|
|
491
|
-
* Example: [CPU_TYPE_X86_64, CPU_TYPE_ARM64]
|
|
492
|
-
*/
|
|
493
|
-
supportedCpus: CPUType[];
|
|
494
|
-
/**
|
|
495
|
-
* Serving framework/stack.
|
|
496
|
-
* Examples: "TGI", "vLLM", "llama.cpp"
|
|
497
|
-
*/
|
|
498
|
-
servingFramework: string;
|
|
499
|
-
/**
|
|
500
|
-
* Pre-built container image if available.
|
|
501
|
-
* Example: "nvcr.io/nvidia/pytorch:23.10-py3"
|
|
502
|
-
*/
|
|
503
|
-
containerImage: string;
|
|
504
|
-
/**
|
|
505
|
-
* Minimum GPU for reasonable performance.
|
|
506
|
-
* Example: GPU_TYPE_RTX_3090 for 7B models
|
|
507
|
-
*/
|
|
508
|
-
minimumGpu: GPUType;
|
|
509
|
-
/**
|
|
510
|
-
* Minimum GPU memory for basic inference in GB.
|
|
511
|
-
* Example: 16 for 7B int8 models
|
|
512
|
-
*/
|
|
513
|
-
minimumGpuMemoryGb: number;
|
|
514
|
-
/**
|
|
515
|
-
* Whether model can run on CPU (even if slowly).
|
|
516
|
-
* Example: true for smaller quantized models
|
|
517
|
-
*/
|
|
518
|
-
cpuCompatible: boolean;
|
|
519
|
-
/**
|
|
520
|
-
* Recommended GPU memory for good performance in GB.
|
|
521
|
-
* Example: 24 for 7B fp16 with reasonable batch size
|
|
522
|
-
*/
|
|
523
|
-
recommendedGpuMemoryGb: number;
|
|
524
|
-
}
|
|
525
|
-
/** Safety contains safety, moderation, and compliance features. */
|
|
526
|
-
export interface Safety {
|
|
527
|
-
/**
|
|
528
|
-
* Built-in content moderation strictness.
|
|
529
|
-
* Example: MODERATION_LEVEL_HIGH for family-friendly models
|
|
530
|
-
*/
|
|
531
|
-
moderationLevel: ModerationLevel;
|
|
532
|
-
/**
|
|
533
|
-
* Whether automatic content filtering is enabled.
|
|
534
|
-
* Example: true for models that block harmful content
|
|
535
|
-
*/
|
|
536
|
-
contentFiltering: boolean;
|
|
537
|
-
/**
|
|
538
|
-
* List of active safety filters.
|
|
539
|
-
* Examples: ["violence", "sexual", "hate", "self-harm", "illegal"]
|
|
540
|
-
*/
|
|
541
|
-
safetyFilters: string[];
|
|
542
|
-
/**
|
|
543
|
-
* Bias assessment rating.
|
|
544
|
-
* Examples: "Low", "Medium", "High", "Evaluated"
|
|
545
|
-
*/
|
|
546
|
-
biasRating: string;
|
|
547
|
-
/**
|
|
548
|
-
* Toxicity score from evaluations (0-1).
|
|
549
|
-
* Example: 0.02 for well-aligned models (lower is better)
|
|
550
|
-
*/
|
|
551
|
-
toxicityScore: number;
|
|
552
|
-
/**
|
|
553
|
-
* GDPR (General Data Protection Regulation) compliance.
|
|
554
|
-
* Example: true for models that don't retain user data
|
|
555
|
-
*/
|
|
556
|
-
gdprCompliant: boolean;
|
|
557
|
-
/**
|
|
558
|
-
* HIPAA (Health Insurance Portability and Accountability Act) compliance.
|
|
559
|
-
* Example: true for medical-safe models
|
|
560
|
-
*/
|
|
561
|
-
hipaaCompliant: boolean;
|
|
562
|
-
/**
|
|
563
|
-
* SOC 2 (Service Organization Control 2) compliance.
|
|
564
|
-
* Example: true for enterprise-grade security
|
|
565
|
-
*/
|
|
566
|
-
soc2Compliant: boolean;
|
|
567
|
-
/**
|
|
568
|
-
* ISO certification status.
|
|
569
|
-
* Example: true for ISO 27001 certified services
|
|
570
|
-
*/
|
|
571
|
-
isoCertified: boolean;
|
|
572
|
-
/**
|
|
573
|
-
* Can refuse harmful or inappropriate requests.
|
|
574
|
-
* Example: true for models trained to decline harmful tasks
|
|
575
|
-
*/
|
|
576
|
-
refusalCapability: boolean;
|
|
577
|
-
/**
|
|
578
|
-
* Whether outputs include watermarking.
|
|
579
|
-
* Example: true for models with detectible AI signatures
|
|
580
|
-
*/
|
|
581
|
-
watermarkOutput: boolean;
|
|
582
|
-
/**
|
|
583
|
-
* Built-in safety guardrails.
|
|
584
|
-
* Examples: ["constitutional-ai", "harmlessness", "helpfulness"]
|
|
585
|
-
*/
|
|
586
|
-
builtInGuardrails: string[];
|
|
587
|
-
/**
|
|
588
|
-
* Supports custom safety guardrails.
|
|
589
|
-
* Example: true if users can add their own safety rules
|
|
590
|
-
*/
|
|
591
|
-
customGuardrails: boolean;
|
|
592
|
-
}
|
|
593
|
-
/** Licensing contains license and usage terms for the model. */
|
|
594
|
-
export interface Licensing {
|
|
595
|
-
/**
|
|
596
|
-
* Type of license governing model use.
|
|
597
|
-
* Example: LICENSE_TYPE_APACHE_2_0, LICENSE_TYPE_PROPRIETARY
|
|
598
|
-
*/
|
|
599
|
-
licenseType: LicenseType;
|
|
600
|
-
/**
|
|
601
|
-
* URL to full license text.
|
|
602
|
-
* Example: "https://github.com/meta-llama/llama/blob/main/LICENSE"
|
|
603
|
-
*/
|
|
604
|
-
licenseUrl: string;
|
|
605
|
-
/**
|
|
606
|
-
* Whether source code is openly available.
|
|
607
|
-
* Example: true for research papers with code
|
|
608
|
-
*/
|
|
609
|
-
isOpenSource: boolean;
|
|
610
|
-
/**
|
|
611
|
-
* Whether model weights are publicly downloadable.
|
|
612
|
-
* Example: true for LLaMA, Mistral; false for GPT-4
|
|
613
|
-
*/
|
|
614
|
-
isOpenWeights: boolean;
|
|
615
|
-
/**
|
|
616
|
-
* Allowed for commercial/business use.
|
|
617
|
-
* Example: true for Apache/MIT licensed models
|
|
618
|
-
*/
|
|
619
|
-
commercialUse: boolean;
|
|
620
|
-
/**
|
|
621
|
-
* Allowed for research purposes.
|
|
622
|
-
* Example: true for most models, even proprietary ones
|
|
623
|
-
*/
|
|
624
|
-
researchUse: boolean;
|
|
625
|
-
/**
|
|
626
|
-
* Must attribute/cite when using.
|
|
627
|
-
* Example: true for CC-BY licenses
|
|
628
|
-
*/
|
|
629
|
-
attributionRequired: boolean;
|
|
630
|
-
/**
|
|
631
|
-
* Derivatives must use same license.
|
|
632
|
-
* Example: true for GPL, CC-BY-SA licenses
|
|
633
|
-
*/
|
|
634
|
-
shareAlikeRequired: boolean;
|
|
635
|
-
/**
|
|
636
|
-
* Specific usage restrictions or conditions.
|
|
637
|
-
* Examples: ["No use for surveillance", "Monthly active user limits",
|
|
638
|
-
* "No competitive use against provider"]
|
|
639
|
-
*/
|
|
640
|
-
usageRestrictions: string[];
|
|
641
|
-
}
|
|
642
|
-
export declare const Architecture: MessageFns<Architecture>;
|
|
643
|
-
export declare const Training: MessageFns<Training>;
|
|
644
|
-
export declare const TechnicalSpecs: MessageFns<TechnicalSpecs>;
|
|
645
|
-
export declare const Safety: MessageFns<Safety>;
|
|
646
|
-
export declare const Licensing: MessageFns<Licensing>;
|
|
647
|
-
type Builtin = Date | Function | Uint8Array | string | number | boolean | undefined;
|
|
648
|
-
export type DeepPartial<T> = T extends Builtin ? T : T extends globalThis.Array<infer U> ? globalThis.Array<DeepPartial<U>> : T extends ReadonlyArray<infer U> ? ReadonlyArray<DeepPartial<U>> : T extends {} ? {
|
|
649
|
-
[K in keyof T]?: DeepPartial<T[K]>;
|
|
650
|
-
} : Partial<T>;
|
|
651
|
-
type KeysOfUnion<T> = T extends T ? keyof T : never;
|
|
652
|
-
export type Exact<P, I extends P> = P extends Builtin ? P : P & {
|
|
653
|
-
[K in keyof P]: Exact<P[K], I[K]>;
|
|
654
|
-
} & {
|
|
655
|
-
[K in Exclude<keyof I, KeysOfUnion<P>>]: never;
|
|
656
|
-
};
|
|
657
|
-
export interface MessageFns<T> {
|
|
658
|
-
encode(message: T, writer?: BinaryWriter): BinaryWriter;
|
|
659
|
-
decode(input: BinaryReader | Uint8Array, length?: number): T;
|
|
660
|
-
fromJSON(object: any): T;
|
|
661
|
-
toJSON(message: T): unknown;
|
|
662
|
-
create<I extends Exact<DeepPartial<T>, I>>(base?: I): T;
|
|
663
|
-
fromPartial<I extends Exact<DeepPartial<T>, I>>(object: I): T;
|
|
664
|
-
}
|
|
665
|
-
export {};
|