sentencepiece-buf 0.2.1-0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/README.md ADDED
@@ -0,0 +1,9 @@
1
+ # sentencepiece-buf
2
+
3
+ This is an NPM package containing a lightweight compiled version of [`sentencepiece.proto`](https://github.com/google/sentencepiece/blob/v0.2.1/src/sentencepiece.proto) for JavaScript using [`@bufbuild/protobuf`](https://github.com/bufbuild/protobuf-es).
4
+
5
+ It can be imported and run from Node.js and Web.
6
+
7
+ ## Versioning
8
+
9
+ The version of this package tracks the `sentencepiece` repository version, with an additional suffix for changes. For instance `v0.2.1-0` of this repository is `v0.2.1` of SentencePiece.
@@ -0,0 +1,717 @@
1
+ // Copyright 2016 Google Inc.
2
+ //
3
+ // Licensed under the Apache License, Version 2.0 (the "License");
4
+ // you may not use this file except in compliance with the License.
5
+ // You may obtain a copy of the License at
6
+ //
7
+ // http://www.apache.org/licenses/LICENSE-2.0
8
+ //
9
+ // Unless required by applicable law or agreed to in writing, software
10
+ // distributed under the License is distributed on an "AS IS" BASIS,
11
+ // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+ // See the License for the specific language governing permissions and
13
+ // limitations under the License.!
14
+
15
+ // @generated by protoc-gen-es v2.10.2 with parameter "target=js+dts,import_extension=js,js_import_style=module"
16
+ // @generated from file sentencepiece_model.proto (package sentencepiece, syntax proto2)
17
+ /* eslint-disable */
18
+
19
+ import type { GenEnum, GenFile, GenMessage } from "@bufbuild/protobuf/codegenv2";
20
+ import type { Message } from "@bufbuild/protobuf";
21
+
22
+ /**
23
+ * Describes the file sentencepiece_model.proto.
24
+ */
25
+ export declare const file_sentencepiece_model: GenFile;
26
+
27
+ /**
28
+ * TrainerSpec encodes a various parameters for SentencePiece training.
29
+ * Next id: 55
30
+ *
31
+ * @generated from message sentencepiece.TrainerSpec
32
+ */
33
+ export declare type TrainerSpec = Message<"sentencepiece.TrainerSpec"> & {
34
+ /**
35
+ * /////////////////////////////////////////////////////////////////
36
+ * General parameters
37
+ *
38
+ * Input corpus files.
39
+ * Trainer accepts the following two formats:
40
+ * A) Monolingual: plain text, one sentence per line.
41
+ * B) Bilingual: TSV, source sentence <tab> target sentence
42
+ * When bilingual data is passed, shared vocabulary model is built.
43
+ * Note that the input file must be raw corpus, not a preprocessed corpus.
44
+ * Trainer only loads the first `input_sentence_size` sentences specified
45
+ * with this parameter.
46
+ *
47
+ * @generated from field: repeated string input = 1;
48
+ */
49
+ input: string[];
50
+
51
+ /**
52
+ * Input corpus format:
53
+ * "text": one-sentence-per-line text format (default)
54
+ * "tsv": sentence <tab> freq
55
+ *
56
+ * @generated from field: optional string input_format = 7;
57
+ */
58
+ inputFormat: string;
59
+
60
+ /**
61
+ * Output model file prefix.
62
+ * <model_prefix>.model and <model_prefix>.vocab are generated.
63
+ *
64
+ * @generated from field: optional string model_prefix = 2;
65
+ */
66
+ modelPrefix: string;
67
+
68
+ /**
69
+ * @generated from field: optional sentencepiece.TrainerSpec.ModelType model_type = 3 [default = UNIGRAM];
70
+ */
71
+ modelType: TrainerSpec_ModelType;
72
+
73
+ /**
74
+ * Vocabulary size. 8k is the default size.
75
+ *
76
+ * @generated from field: optional int32 vocab_size = 4 [default = 8000];
77
+ */
78
+ vocabSize: number;
79
+
80
+ /**
81
+ * List of the languages this model can accept.
82
+ * Since the model is language-agnostic, this field is used as a reference.
83
+ *
84
+ * @generated from field: repeated string accept_language = 5;
85
+ */
86
+ acceptLanguage: string[];
87
+
88
+ /**
89
+ * Size of self-test samples, which are encoded in the model file.
90
+ *
91
+ * @generated from field: optional int32 self_test_sample_size = 6 [default = 0];
92
+ */
93
+ selfTestSampleSize: number;
94
+
95
+ /**
96
+ * Whether to use DP version of sentencepiece. Use it with TSV input format
97
+ * (requires precomputed word tab counts to work).
98
+ *
99
+ * @generated from field: optional bool enable_differential_privacy = 50 [default = false];
100
+ */
101
+ enableDifferentialPrivacy: boolean;
102
+
103
+ /**
104
+ * Set these parameters if you need DP version of sentencepiece.
105
+ * std of noise to add.
106
+ *
107
+ * @generated from field: optional float differential_privacy_noise_level = 51 [default = 0];
108
+ */
109
+ differentialPrivacyNoiseLevel: number;
110
+
111
+ /**
112
+ * Clipping threshold to apply after adding noise. All the words with
113
+ * frequency less than this value are dropped.
114
+ *
115
+ * @generated from field: optional uint64 differential_privacy_clipping_threshold = 52 [default = 0];
116
+ */
117
+ differentialPrivacyClippingThreshold: bigint;
118
+
119
+ /**
120
+ * /////////////////////////////////////////////////////////////////
121
+ * Training parameters.
122
+ *
123
+ * Uses characters which cover the corpus with the ratio of `chars_coverage`.
124
+ * This parameter determines the set of basic Alphabet of sentence piece.
125
+ * 1.0 - `chars_coverage` characters are treated as UNK.
126
+ * See also required_chars field.
127
+ *
128
+ * @generated from field: optional float character_coverage = 10 [default = 0.9995];
129
+ */
130
+ characterCoverage: number;
131
+
132
+ /**
133
+ * Maximum size of sentences the trainer loads from `input` parameter.
134
+ * Trainer simply loads the `input` files in sequence.
135
+ * It is better to shuffle the input corpus randomly.
136
+ *
137
+ * @generated from field: optional uint64 input_sentence_size = 11 [default = 0];
138
+ */
139
+ inputSentenceSize: bigint;
140
+
141
+ /**
142
+ * @generated from field: optional bool shuffle_input_sentence = 19 [default = true];
143
+ */
144
+ shuffleInputSentence: boolean;
145
+
146
+ /**
147
+ * Maximum size of sentences to make seed sentence pieces.
148
+ * Extended suffix array is constructed to extract frequent
149
+ * sub-strings from the corpus. This uses 20N working space,
150
+ * where N is the size of corpus.
151
+ *
152
+ * @generated from field: optional int32 mining_sentence_size = 12 [deprecated = true];
153
+ * @deprecated
154
+ */
155
+ miningSentenceSize: number;
156
+
157
+ /**
158
+ * Maximum size of sentences to train sentence pieces.
159
+ *
160
+ * @generated from field: optional int32 training_sentence_size = 13 [deprecated = true];
161
+ * @deprecated
162
+ */
163
+ trainingSentenceSize: number;
164
+
165
+ /**
166
+ * The size of seed sentencepieces.
167
+ * `seed_sentencepiece_size` must be larger than `vocab_size`.
168
+ *
169
+ * @generated from field: optional int32 seed_sentencepiece_size = 14 [default = 1000000];
170
+ */
171
+ seedSentencepieceSize: number;
172
+
173
+ /**
174
+ * In every EM sub-iterations, keeps top
175
+ * `shrinking_factor` * `current sentencepieces size` with respect to
176
+ * the loss of the sentence piece. This value should be smaller than 1.0.
177
+ *
178
+ * @generated from field: optional float shrinking_factor = 15 [default = 0.75];
179
+ */
180
+ shrinkingFactor: number;
181
+
182
+ /**
183
+ * The maximum sentence length in byte. The sentences with the length
184
+ * larger than `max_sentence_length` is simply ignored.
185
+ * Longer input tends to bring the following risks:
186
+ * * Overflow during EM training (unigram language model only)
187
+ * * Performance drop because of O(n log n) cost in BPE.
188
+ *
189
+ * @generated from field: optional int32 max_sentence_length = 18 [default = 4192];
190
+ */
191
+ maxSentenceLength: number;
192
+
193
+ /**
194
+ * Number of threads in the training.
195
+ *
196
+ * @generated from field: optional int32 num_threads = 16 [default = 16];
197
+ */
198
+ numThreads: number;
199
+
200
+ /**
201
+ * Number of EM sub iterations.
202
+ *
203
+ * @generated from field: optional int32 num_sub_iterations = 17 [default = 2];
204
+ */
205
+ numSubIterations: number;
206
+
207
+ /**
208
+ * /////////////////////////////////////////////////////////////////
209
+ * SentencePiece parameters which control the shapes of sentence piece.
210
+ *
211
+ * Maximum length of sentencepiece.
212
+ *
213
+ * @generated from field: optional int32 max_sentencepiece_length = 20 [default = 16];
214
+ */
215
+ maxSentencepieceLength: number;
216
+
217
+ /**
218
+ * Uses Unicode script to split sentence pieces.
219
+ * When `split_by_unicode_script` is true, we do not allow sentence piece to
220
+ * include multiple Unicode scripts, e.g. "F1" is not a valid piece.
221
+ * Exception: CJ characters (Hiragana/Katakana/Han) are all handled
222
+ * as one script type, since Japanese word can consist of multiple scripts.
223
+ * This exception is always applied regardless of the accept-language
224
+ * parameter.
225
+ *
226
+ * @generated from field: optional bool split_by_unicode_script = 21 [default = true];
227
+ */
228
+ splitByUnicodeScript: boolean;
229
+
230
+ /**
231
+ * When `split_by_number` is true, put a boundary between number and
232
+ * non-number transition. If we want to treat "F1" is one token, set this flag
233
+ * to be false.
234
+ *
235
+ * @generated from field: optional bool split_by_number = 23 [default = true];
236
+ */
237
+ splitByNumber: boolean;
238
+
239
+ /**
240
+ * Use a white space to split sentence pieces.
241
+ * When `split_by_whitespace` is false, we may have the piece containing
242
+ * a white space in the middle. e.g., "in_the".
243
+ *
244
+ * @generated from field: optional bool split_by_whitespace = 22 [default = true];
245
+ */
246
+ splitByWhitespace: boolean;
247
+
248
+ /**
249
+ * Adds whitespace symbol (_) as a suffix instead of prefix. e.g., _hello =>
250
+ * hello_. When `treat_whitespace_as_suffix` is true,
251
+ * NormalizerSpec::add_dummy_prefix will add the dummy whitespace to the end
252
+ * of sentence.
253
+ *
254
+ * @generated from field: optional bool treat_whitespace_as_suffix = 24 [default = false];
255
+ */
256
+ treatWhitespaceAsSuffix: boolean;
257
+
258
+ /**
259
+ * Allows pieces that only contain whitespaces instead of appearing only as
260
+ * prefix or suffix of other pieces.
261
+ *
262
+ * @generated from field: optional bool allow_whitespace_only_pieces = 26 [default = false];
263
+ */
264
+ allowWhitespaceOnlyPieces: boolean;
265
+
266
+ /**
267
+ * Split all digits (0-9) into separate pieces.
268
+ *
269
+ * @generated from field: optional bool split_digits = 25 [default = false];
270
+ */
271
+ splitDigits: boolean;
272
+
273
+ /**
274
+ * Defines the pre-tokenization delimiter.
275
+ * When specified, no pieces crossing this delimiter is not included
276
+ * in the vocab. Then the delimiter string is virtually ignored
277
+ * during the training. This field can allows constraints on the vocabulary
278
+ * selection. Note that this field is available on unigram mode.
279
+ *
280
+ * @generated from field: optional string pretokenization_delimiter = 53 [default = ""];
281
+ */
282
+ pretokenizationDelimiter: string;
283
+
284
+ /**
285
+ * /////////////////////////////////////////////////////////////////
286
+ * Vocabulary management
287
+ *
288
+ * Defines control symbols used as an indicator to
289
+ * change the behavior of the decoder. <s> and </s> are pre-defined.
290
+ * We can use this field to encode various meta information,
291
+ * including language indicator in multilingual model.
292
+ * These symbols are not visible to users, but visible to
293
+ * the decoder. Note that when the input sentence contains control symbols,
294
+ * they are not treated as one token, but segmented into normal pieces.
295
+ * Control symbols must be inserted independently from the segmentation.
296
+ *
297
+ * @generated from field: repeated string control_symbols = 30;
298
+ */
299
+ controlSymbols: string[];
300
+
301
+ /**
302
+ * Defines user defined symbols.
303
+ * These symbols are added with extremely high score
304
+ * so they are always treated as one unique symbol in any context.
305
+ * Typical usage of user_defined_symbols is placeholder for named entities.
306
+ *
307
+ * @generated from field: repeated string user_defined_symbols = 31;
308
+ */
309
+ userDefinedSymbols: string[];
310
+
311
+ /**
312
+ * Defines required characters. Each UTF8 character in this string is included
313
+ * in the character set regardless of character_coverage value. Unlike
314
+ * user_defined_symbols, these characters have scores based on the frequency
315
+ * on input sentences, and the model can form subwords using characters
316
+ * in this field.
317
+ *
318
+ * @generated from field: optional string required_chars = 36;
319
+ */
320
+ requiredChars: string;
321
+
322
+ /**
323
+ * Decomposes unknown pieces into UTF-8 bytes.
324
+ *
325
+ * @generated from field: optional bool byte_fallback = 35 [default = false];
326
+ */
327
+ byteFallback: boolean;
328
+
329
+ /**
330
+ * When creating the vocabulary file, defines whether or not to additionally
331
+ * output the score for each piece.
332
+ *
333
+ * @generated from field: optional bool vocabulary_output_piece_score = 32 [default = true];
334
+ */
335
+ vocabularyOutputPieceScore: boolean;
336
+
337
+ /**
338
+ * `vocab_size` is treated as hard limit. Crash if
339
+ * the model can not produce the vocab of size `vocab_size`,
340
+ * When `hard_vocab_limit` is false, vocab_size is treated
341
+ * as soft limit. Note that when model_type=char,
342
+ * always assumes hard_vocab_limit = false.
343
+ *
344
+ * @generated from field: optional bool hard_vocab_limit = 33 [default = true];
345
+ */
346
+ hardVocabLimit: boolean;
347
+
348
+ /**
349
+ * use all symbols for vocab extraction. This flag is valid
350
+ * if model type is either CHAR or WORD
351
+ *
352
+ * @generated from field: optional bool use_all_vocab = 34 [default = false];
353
+ */
354
+ useAllVocab: boolean;
355
+
356
+ /**
357
+ * /////////////////////////////////////////////////////////////////
358
+ * Reserved special meta tokens.
359
+ * * -1 is not used.
360
+ * * unk_id must not be -1.
361
+ * Id must starts with 0 and be contigous.
362
+ *
363
+ * <unk>
364
+ *
365
+ * @generated from field: optional int32 unk_id = 40 [default = 0];
366
+ */
367
+ unkId: number;
368
+
369
+ /**
370
+ * <s>
371
+ *
372
+ * @generated from field: optional int32 bos_id = 41 [default = 1];
373
+ */
374
+ bosId: number;
375
+
376
+ /**
377
+ * </s>
378
+ *
379
+ * @generated from field: optional int32 eos_id = 42 [default = 2];
380
+ */
381
+ eosId: number;
382
+
383
+ /**
384
+ * <pad> (padding)
385
+ *
386
+ * @generated from field: optional int32 pad_id = 43 [default = -1];
387
+ */
388
+ padId: number;
389
+
390
+ /**
391
+ * @generated from field: optional string unk_piece = 45 [default = "<unk>"];
392
+ */
393
+ unkPiece: string;
394
+
395
+ /**
396
+ * @generated from field: optional string bos_piece = 46 [default = "<s>"];
397
+ */
398
+ bosPiece: string;
399
+
400
+ /**
401
+ * @generated from field: optional string eos_piece = 47 [default = "</s>"];
402
+ */
403
+ eosPiece: string;
404
+
405
+ /**
406
+ * @generated from field: optional string pad_piece = 48 [default = "<pad>"];
407
+ */
408
+ padPiece: string;
409
+
410
+ /**
411
+ * Encodes <unk> into U+2047 (DOUBLE QUESTION MARK),
412
+ * since this character can be useful both for user and
413
+ * developer. We can easily figure out that <unk> is emitted.
414
+ *
415
+ * @generated from field: optional string unk_surface = 44 [default = " ⁇ "];
416
+ */
417
+ unkSurface: string;
418
+
419
+ /**
420
+ * Increase bit depth to allow unigram model training on large
421
+ * (>10M sentences) corpora. A Side-effect of enabling this flag
422
+ * is increased memory usage.
423
+ *
424
+ * @generated from field: optional bool train_extremely_large_corpus = 49 [default = false];
425
+ */
426
+ trainExtremelyLargeCorpus: boolean;
427
+
428
+ /**
429
+ * Path to a seed sentencepieces file, with one tab-separated
430
+ * seed sentencepiece <tab> frequency per line.
431
+ *
432
+ * @generated from field: optional string seed_sentencepieces_file = 54 [default = ""];
433
+ */
434
+ seedSentencepiecesFile: string;
435
+ };
436
+
437
+ /**
438
+ * Describes the message sentencepiece.TrainerSpec.
439
+ * Use `create(TrainerSpecSchema)` to create a new message.
440
+ */
441
+ export declare const TrainerSpecSchema: GenMessage<TrainerSpec>;
442
+
443
+ /**
444
+ * Model type. only have UNIGRAM now.
445
+ *
446
+ * @generated from enum sentencepiece.TrainerSpec.ModelType
447
+ */
448
+ export enum TrainerSpec_ModelType {
449
+ /**
450
+ * Unigram language model with dynamic algorithm
451
+ *
452
+ * @generated from enum value: UNIGRAM = 1;
453
+ */
454
+ UNIGRAM = 1,
455
+
456
+ /**
457
+ * Byte Pair Encoding
458
+ *
459
+ * @generated from enum value: BPE = 2;
460
+ */
461
+ BPE = 2,
462
+
463
+ /**
464
+ * Delimitered by whitespace.
465
+ *
466
+ * @generated from enum value: WORD = 3;
467
+ */
468
+ WORD = 3,
469
+
470
+ /**
471
+ * tokenizes into character sequence
472
+ *
473
+ * @generated from enum value: CHAR = 4;
474
+ */
475
+ CHAR = 4,
476
+ }
477
+
478
+ /**
479
+ * Describes the enum sentencepiece.TrainerSpec.ModelType.
480
+ */
481
+ export declare const TrainerSpec_ModelTypeSchema: GenEnum<TrainerSpec_ModelType>;
482
+
483
+ /**
484
+ * NormalizerSpec encodes a various parameters for string normalizaiton
485
+ *
486
+ * @generated from message sentencepiece.NormalizerSpec
487
+ */
488
+ export declare type NormalizerSpec = Message<"sentencepiece.NormalizerSpec"> & {
489
+ /**
490
+ * name of normalization rule.
491
+ *
492
+ * @generated from field: optional string name = 1;
493
+ */
494
+ name: string;
495
+
496
+ /**
497
+ * Pre-compiled normalization rule created by
498
+ * Builder::GetPrecompiledCharsMap() or Builder::CompileCharsMap() method.
499
+ * Usually this field is set by Builder::GetNormalizerSpec() method.
500
+ *
501
+ * @generated from field: optional bytes precompiled_charsmap = 2;
502
+ */
503
+ precompiledCharsmap: Uint8Array<ArrayBuffer>;
504
+
505
+ /**
506
+ * Adds dummy whitespace at the beginning of text in order to
507
+ * treat "world" in "world" and "hello world" in the same way.
508
+ *
509
+ * @generated from field: optional bool add_dummy_prefix = 3 [default = true];
510
+ */
511
+ addDummyPrefix: boolean;
512
+
513
+ /**
514
+ * Removes leading, trailing, and duplicate internal whitespace.
515
+ *
516
+ * @generated from field: optional bool remove_extra_whitespaces = 4 [default = true];
517
+ */
518
+ removeExtraWhitespaces: boolean;
519
+
520
+ /**
521
+ * Replaces whitespace with meta symbol.
522
+ * This field must be true to train sentence piece model.
523
+ *
524
+ * @generated from field: optional bool escape_whitespaces = 5 [default = true];
525
+ */
526
+ escapeWhitespaces: boolean;
527
+
528
+ /**
529
+ * Custom normalization rule file in TSV format.
530
+ * https://github.com/google/sentencepiece/blob/master/doc/normalization.md
531
+ * This field is only used in SentencePieceTrainer::Train() method, which
532
+ * compiles the rule into the binary rule stored in `precompiled_charsmap`.
533
+ *
534
+ * @generated from field: optional string normalization_rule_tsv = 6;
535
+ */
536
+ normalizationRuleTsv: string;
537
+ };
538
+
539
+ /**
540
+ * Describes the message sentencepiece.NormalizerSpec.
541
+ * Use `create(NormalizerSpecSchema)` to create a new message.
542
+ */
543
+ export declare const NormalizerSpecSchema: GenMessage<NormalizerSpec>;
544
+
545
+ /**
546
+ * Proto to store samples for self-testing.
547
+ *
548
+ * @generated from message sentencepiece.SelfTestData
549
+ */
550
+ export declare type SelfTestData = Message<"sentencepiece.SelfTestData"> & {
551
+ /**
552
+ * @generated from field: repeated sentencepiece.SelfTestData.Sample samples = 1;
553
+ */
554
+ samples: SelfTestData_Sample[];
555
+ };
556
+
557
+ /**
558
+ * Describes the message sentencepiece.SelfTestData.
559
+ * Use `create(SelfTestDataSchema)` to create a new message.
560
+ */
561
+ export declare const SelfTestDataSchema: GenMessage<SelfTestData>;
562
+
563
+ /**
564
+ * @generated from message sentencepiece.SelfTestData.Sample
565
+ */
566
+ export declare type SelfTestData_Sample = Message<"sentencepiece.SelfTestData.Sample"> & {
567
+ /**
568
+ * @generated from field: optional string input = 1;
569
+ */
570
+ input: string;
571
+
572
+ /**
573
+ * @generated from field: optional string expected = 2;
574
+ */
575
+ expected: string;
576
+ };
577
+
578
+ /**
579
+ * Describes the message sentencepiece.SelfTestData.Sample.
580
+ * Use `create(SelfTestData_SampleSchema)` to create a new message.
581
+ */
582
+ export declare const SelfTestData_SampleSchema: GenMessage<SelfTestData_Sample>;
583
+
584
+ /**
585
+ * ModelProto stores model parameters.
586
+ * SentencePieceProcessor is supposed to be self-contained.
587
+ * All settings/parameters which may change the behavior must be encoded
588
+ * in ModelProto.
589
+ *
590
+ * @generated from message sentencepiece.ModelProto
591
+ */
592
+ export declare type ModelProto = Message<"sentencepiece.ModelProto"> & {
593
+ /**
594
+ * Sentence pieces with scores.
595
+ *
596
+ * @generated from field: repeated sentencepiece.ModelProto.SentencePiece pieces = 1;
597
+ */
598
+ pieces: ModelProto_SentencePiece[];
599
+
600
+ /**
601
+ * Spec used to generate this model file.
602
+ *
603
+ * @generated from field: optional sentencepiece.TrainerSpec trainer_spec = 2;
604
+ */
605
+ trainerSpec?: TrainerSpec;
606
+
607
+ /**
608
+ * Spec for text normalization.
609
+ *
610
+ * @generated from field: optional sentencepiece.NormalizerSpec normalizer_spec = 3;
611
+ */
612
+ normalizerSpec?: NormalizerSpec;
613
+
614
+ /**
615
+ * Stores sample input and its expected segmentation to verify the model.
616
+ *
617
+ * @generated from field: optional sentencepiece.SelfTestData self_test_data = 4;
618
+ */
619
+ selfTestData?: SelfTestData;
620
+
621
+ /**
622
+ * Spec for text de-normalization.
623
+ *
624
+ * @generated from field: optional sentencepiece.NormalizerSpec denormalizer_spec = 5;
625
+ */
626
+ denormalizerSpec?: NormalizerSpec;
627
+ };
628
+
629
+ /**
630
+ * Describes the message sentencepiece.ModelProto.
631
+ * Use `create(ModelProtoSchema)` to create a new message.
632
+ */
633
+ export declare const ModelProtoSchema: GenMessage<ModelProto>;
634
+
635
+ /**
636
+ * @generated from message sentencepiece.ModelProto.SentencePiece
637
+ */
638
+ export declare type ModelProto_SentencePiece = Message<"sentencepiece.ModelProto.SentencePiece"> & {
639
+ /**
640
+ * piece must not be empty.
641
+ *
642
+ * @generated from field: optional string piece = 1;
643
+ */
644
+ piece: string;
645
+
646
+ /**
647
+ * @generated from field: optional float score = 2;
648
+ */
649
+ score: number;
650
+
651
+ /**
652
+ * @generated from field: optional sentencepiece.ModelProto.SentencePiece.Type type = 3 [default = NORMAL];
653
+ */
654
+ type: ModelProto_SentencePiece_Type;
655
+ };
656
+
657
+ /**
658
+ * Describes the message sentencepiece.ModelProto.SentencePiece.
659
+ * Use `create(ModelProto_SentencePieceSchema)` to create a new message.
660
+ */
661
+ export declare const ModelProto_SentencePieceSchema: GenMessage<ModelProto_SentencePiece>;
662
+
663
+ /**
664
+ * @generated from enum sentencepiece.ModelProto.SentencePiece.Type
665
+ */
666
+ export enum ModelProto_SentencePiece_Type {
667
+ /**
668
+ * normal symbol
669
+ *
670
+ * @generated from enum value: NORMAL = 1;
671
+ */
672
+ NORMAL = 1,
673
+
674
+ /**
675
+ * unknown symbol. only <unk> for now.
676
+ *
677
+ * @generated from enum value: UNKNOWN = 2;
678
+ */
679
+ UNKNOWN = 2,
680
+
681
+ /**
682
+ * control symbols. </s>, <s>, <2ja> etc.
683
+ *
684
+ * @generated from enum value: CONTROL = 3;
685
+ */
686
+ CONTROL = 3,
687
+
688
+ /**
689
+ * user defined symbols.
690
+ *
691
+ * @generated from enum value: USER_DEFINED = 4;
692
+ */
693
+ USER_DEFINED = 4,
694
+
695
+ /**
696
+ * Typical usage of USER_DEFINED symbol
697
+ * is placeholder.
698
+ *
699
+ * byte symbols. Used when `byte_fallback` is true.
700
+ *
701
+ * @generated from enum value: BYTE = 6;
702
+ */
703
+ BYTE = 6,
704
+
705
+ /**
706
+ * this piece is not used.
707
+ *
708
+ * @generated from enum value: UNUSED = 5;
709
+ */
710
+ UNUSED = 5,
711
+ }
712
+
713
+ /**
714
+ * Describes the enum sentencepiece.ModelProto.SentencePiece.Type.
715
+ */
716
+ export declare const ModelProto_SentencePiece_TypeSchema: GenEnum<ModelProto_SentencePiece_Type>;
717
+
@@ -0,0 +1,94 @@
1
+ // Copyright 2016 Google Inc.
2
+ //
3
+ // Licensed under the Apache License, Version 2.0 (the "License");
4
+ // you may not use this file except in compliance with the License.
5
+ // You may obtain a copy of the License at
6
+ //
7
+ // http://www.apache.org/licenses/LICENSE-2.0
8
+ //
9
+ // Unless required by applicable law or agreed to in writing, software
10
+ // distributed under the License is distributed on an "AS IS" BASIS,
11
+ // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+ // See the License for the specific language governing permissions and
13
+ // limitations under the License.!
14
+
15
+ // @generated by protoc-gen-es v2.10.2 with parameter "target=js+dts,import_extension=js,js_import_style=module"
16
+ // @generated from file sentencepiece_model.proto (package sentencepiece, syntax proto2)
17
+ /* eslint-disable */
18
+
19
+ import { enumDesc, fileDesc, messageDesc, tsEnum } from "@bufbuild/protobuf/codegenv2";
20
+
21
+ /**
22
+ * Describes the file sentencepiece_model.proto.
23
+ */
24
+ export const file_sentencepiece_model = /*@__PURE__*/
25
+ fileDesc("ChlzZW50ZW5jZXBpZWNlX21vZGVsLnByb3RvEg1zZW50ZW5jZXBpZWNlIqQMCgtUcmFpbmVyU3BlYxINCgVpbnB1dBgBIAMoCRIUCgxpbnB1dF9mb3JtYXQYByABKAkSFAoMbW9kZWxfcHJlZml4GAIgASgJEkEKCm1vZGVsX3R5cGUYAyABKA4yJC5zZW50ZW5jZXBpZWNlLlRyYWluZXJTcGVjLk1vZGVsVHlwZToHVU5JR1JBTRIYCgp2b2NhYl9zaXplGAQgASgFOgQ4MDAwEhcKD2FjY2VwdF9sYW5ndWFnZRgFIAMoCRIgChVzZWxmX3Rlc3Rfc2FtcGxlX3NpemUYBiABKAU6ATASKgobZW5hYmxlX2RpZmZlcmVudGlhbF9wcml2YWN5GDIgASgIOgVmYWxzZRIrCiBkaWZmZXJlbnRpYWxfcHJpdmFjeV9ub2lzZV9sZXZlbBgzIAEoAjoBMBIyCidkaWZmZXJlbnRpYWxfcHJpdmFjeV9jbGlwcGluZ190aHJlc2hvbGQYNCABKAQ6ATASIgoSY2hhcmFjdGVyX2NvdmVyYWdlGAogASgCOgYwLjk5OTUSHgoTaW5wdXRfc2VudGVuY2Vfc2l6ZRgLIAEoBDoBMBIkChZzaHVmZmxlX2lucHV0X3NlbnRlbmNlGBMgASgIOgR0cnVlEiAKFG1pbmluZ19zZW50ZW5jZV9zaXplGAwgASgFQgIYARIiChZ0cmFpbmluZ19zZW50ZW5jZV9zaXplGA0gASgFQgIYARIoChdzZWVkX3NlbnRlbmNlcGllY2Vfc2l6ZRgOIAEoBToHMTAwMDAwMBIeChBzaHJpbmtpbmdfZmFjdG9yGA8gASgCOgQwLjc1EiEKE21heF9zZW50ZW5jZV9sZW5ndGgYEiABKAU6BDQxOTISFwoLbnVtX3RocmVhZHMYECABKAU6AjE2Eh0KEm51bV9zdWJfaXRlcmF0aW9ucxgRIAEoBToBMhIkChhtYXhfc2VudGVuY2VwaWVjZV9sZW5ndGgYFCABKAU6AjE2EiUKF3NwbGl0X2J5X3VuaWNvZGVfc2NyaXB0GBUgASgIOgR0cnVlEh0KD3NwbGl0X2J5X251bWJlchgXIAEoCDoEdHJ1ZRIhChNzcGxpdF9ieV93aGl0ZXNwYWNlGBYgASgIOgR0cnVlEikKGnRyZWF0X3doaXRlc3BhY2VfYXNfc3VmZml4GBggASgIOgVmYWxzZRIrChxhbGxvd193aGl0ZXNwYWNlX29ubHlfcGllY2VzGBogASgIOgVmYWxzZRIbCgxzcGxpdF9kaWdpdHMYGSABKAg6BWZhbHNlEiMKGXByZXRva2VuaXphdGlvbl9kZWxpbWl0ZXIYNSABKAk6ABIXCg9jb250cm9sX3N5bWJvbHMYHiADKAkSHAoUdXNlcl9kZWZpbmVkX3N5bWJvbHMYHyADKAkSFgoOcmVxdWlyZWRfY2hhcnMYJCABKAkSHAoNYnl0ZV9mYWxsYmFjaxgjIAEoCDoFZmFsc2USKwoddm9jYWJ1bGFyeV9vdXRwdXRfcGllY2Vfc2NvcmUYICABKAg6BHRydWUSHgoQaGFyZF92b2NhYl9saW1pdBghIAEoCDoEdHJ1ZRIcCg11c2VfYWxsX3ZvY2FiGCIgASgIOgVmYWxzZRIRCgZ1bmtfaWQYKCABKAU6ATASEQoGYm9zX2lkGCkgASgFOgExEhEKBmVvc19pZBgqIAEoBToBMhISCgZwYWRfaWQYKyABKAU6Ai0xEhgKCXVua19waWVjZRgtIAEoCToFPHVuaz4SFgoJYm9zX3BpZWNlGC4gASgJOgM8cz4SFwoJZW9zX3BpZWNlGC8gASgJOgQ8L3M+EhgKCXBhZF9waWVjZRgwIAEoCToFPHBhZD4SGgoLdW5rX3N1cmZhY2UYLCABKAk6BSDigYcgEisKHHRyYWluX2V4dHJlbWVseV9sYXJnZV9jb3JwdXMYMSABKAg6BWZhbHNlEiIKGHNlZWRfc2VudGVuY2VwaWVjZXNfZmlsZRg2IAEoCToAIjUKCU1vZGVsVHlwZRILCgdVTklHUkFNEAESBwoDQlBFEAISCAoEV09SRBADEggKBENIQVIQBCoJCMgBEICAgIACItEBCg5Ob3JtYWxpemVyU3BlYxIMCgRuYW1lGAEgASgJEhwKFHByZWNvbXBpbGVkX2NoYXJzbWFwGAIgASgMEh4KEGFkZF9kdW1teV9wcmVmaXgYAyABKAg6BHRydWUSJgoYcmVtb3ZlX2V4dHJhX3doaXRlc3BhY2VzGAQgASgIOgR0cnVlEiAKEmVzY2FwZV93aGl0ZXNwYWNlcxgFIAEoCDoEdHJ1ZRIeChZub3JtYWxpemF0aW9uX3J1bGVfdHN2GAYgASgJKgkIyAEQgICAgAIieQoMU2VsZlRlc3REYXRhEjMKB3NhbXBsZXMYASADKAsyIi5zZW50ZW5jZXBpZWNlLlNlbGZUZXN0RGF0YS5TYW1wbGUaKQoGU2FtcGxlEg0KBWlucHV0GAEgASgJEhAKCGV4cGVjdGVkGAIgASgJKgkIyAEQgICAgAIi/gMKCk1vZGVsUHJvdG8SNwoGcGllY2VzGAEgAygLMicuc2VudGVuY2VwaWVjZS5Nb2RlbFByb3RvLlNlbnRlbmNlUGllY2USMAoMdHJhaW5lcl9zcGVjGAIgASgLMhouc2VudGVuY2VwaWVjZS5UcmFpbmVyU3BlYxI2Cg9ub3JtYWxpemVyX3NwZWMYAyABKAsyHS5zZW50ZW5jZXBpZWNlLk5vcm1hbGl6ZXJTcGVjEjMKDnNlbGZfdGVzdF9kYXRhGAQgASgLMhsuc2VudGVuY2VwaWVjZS5TZWxmVGVzdERhdGESOAoRZGVub3JtYWxpemVyX3NwZWMYBSABKAsyHS5zZW50ZW5jZXBpZWNlLk5vcm1hbGl6ZXJTcGVjGtIBCg1TZW50ZW5jZVBpZWNlEg0KBXBpZWNlGAEgASgJEg0KBXNjb3JlGAIgASgCEkIKBHR5cGUYAyABKA4yLC5zZW50ZW5jZXBpZWNlLk1vZGVsUHJvdG8uU2VudGVuY2VQaWVjZS5UeXBlOgZOT1JNQUwiVAoEVHlwZRIKCgZOT1JNQUwQARILCgdVTktOT1dOEAISCwoHQ09OVFJPTBADEhAKDFVTRVJfREVGSU5FRBAEEggKBEJZVEUQBhIKCgZVTlVTRUQQBSoJCMgBEICAgIACKgkIyAEQgICAgAJCAkgD");
26
+
27
+ /**
28
+ * Describes the message sentencepiece.TrainerSpec.
29
+ * Use `create(TrainerSpecSchema)` to create a new message.
30
+ */
31
+ export const TrainerSpecSchema = /*@__PURE__*/
32
+ messageDesc(file_sentencepiece_model, 0);
33
+
34
+ /**
35
+ * Describes the enum sentencepiece.TrainerSpec.ModelType.
36
+ */
37
+ export const TrainerSpec_ModelTypeSchema = /*@__PURE__*/
38
+ enumDesc(file_sentencepiece_model, 0, 0);
39
+
40
+ /**
41
+ * Model type. only have UNIGRAM now.
42
+ *
43
+ * @generated from enum sentencepiece.TrainerSpec.ModelType
44
+ */
45
+ export const TrainerSpec_ModelType = /*@__PURE__*/
46
+ tsEnum(TrainerSpec_ModelTypeSchema);
47
+
48
+ /**
49
+ * Describes the message sentencepiece.NormalizerSpec.
50
+ * Use `create(NormalizerSpecSchema)` to create a new message.
51
+ */
52
+ export const NormalizerSpecSchema = /*@__PURE__*/
53
+ messageDesc(file_sentencepiece_model, 1);
54
+
55
+ /**
56
+ * Describes the message sentencepiece.SelfTestData.
57
+ * Use `create(SelfTestDataSchema)` to create a new message.
58
+ */
59
+ export const SelfTestDataSchema = /*@__PURE__*/
60
+ messageDesc(file_sentencepiece_model, 2);
61
+
62
+ /**
63
+ * Describes the message sentencepiece.SelfTestData.Sample.
64
+ * Use `create(SelfTestData_SampleSchema)` to create a new message.
65
+ */
66
+ export const SelfTestData_SampleSchema = /*@__PURE__*/
67
+ messageDesc(file_sentencepiece_model, 2, 0);
68
+
69
+ /**
70
+ * Describes the message sentencepiece.ModelProto.
71
+ * Use `create(ModelProtoSchema)` to create a new message.
72
+ */
73
+ export const ModelProtoSchema = /*@__PURE__*/
74
+ messageDesc(file_sentencepiece_model, 3);
75
+
76
+ /**
77
+ * Describes the message sentencepiece.ModelProto.SentencePiece.
78
+ * Use `create(ModelProto_SentencePieceSchema)` to create a new message.
79
+ */
80
+ export const ModelProto_SentencePieceSchema = /*@__PURE__*/
81
+ messageDesc(file_sentencepiece_model, 3, 0);
82
+
83
+ /**
84
+ * Describes the enum sentencepiece.ModelProto.SentencePiece.Type.
85
+ */
86
+ export const ModelProto_SentencePiece_TypeSchema = /*@__PURE__*/
87
+ enumDesc(file_sentencepiece_model, 3, 0, 0);
88
+
89
+ /**
90
+ * @generated from enum sentencepiece.ModelProto.SentencePiece.Type
91
+ */
92
+ export const ModelProto_SentencePiece_Type = /*@__PURE__*/
93
+ tsEnum(ModelProto_SentencePiece_TypeSchema);
94
+
@@ -0,0 +1,135 @@
1
+ // Copyright 2016 Google Inc.
2
+ //
3
+ // Licensed under the Apache License, Version 2.0 (the "License");
4
+ // you may not use this file except in compliance with the License.
5
+ // You may obtain a copy of the License at
6
+ //
7
+ // http://www.apache.org/licenses/LICENSE-2.0
8
+ //
9
+ // Unless required by applicable law or agreed to in writing, software
10
+ // distributed under the License is distributed on an "AS IS" BASIS,
11
+ // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+ // See the License for the specific language governing permissions and
13
+ // limitations under the License.!
14
+
15
+ // @generated by protoc-gen-es v2.10.2 with parameter "target=js+dts,import_extension=js,js_import_style=module"
16
+ // @generated from file sentencepiece.proto (package sentencepiece, syntax proto2)
17
+ /* eslint-disable */
18
+
19
+ import type { GenFile, GenMessage } from "@bufbuild/protobuf/codegenv2";
20
+ import type { Message } from "@bufbuild/protobuf";
21
+
22
+ /**
23
+ * Describes the file sentencepiece.proto.
24
+ */
25
+ export declare const file_sentencepiece: GenFile;
26
+
27
+ /**
28
+ * SentencePieceText manages a user-facing source sentence,
29
+ * postprocessed target sentence, and internal segmentation
30
+ * with byte offsets.
31
+ *
32
+ * @generated from message sentencepiece.SentencePieceText
33
+ */
34
+ export declare type SentencePieceText = Message<"sentencepiece.SentencePieceText"> & {
35
+ /**
36
+ * User input or postprocessed text. This should be immutable
37
+ * since the byte range in SentencePiece is pointing to a span over this
38
+ * text. Meta symbols for whitespaces are not included.
39
+ *
40
+ * @generated from field: optional string text = 1;
41
+ */
42
+ text: string;
43
+
44
+ /**
45
+ * A sequence of sentence pieces. These `pieces` are guaranteed to be ordered
46
+ * in increasing order of `begin` field. Except for the last piece, `end` of
47
+ * each piece is always equal to `begin` of the next piece.
48
+ *
49
+ * @generated from field: repeated sentencepiece.SentencePieceText.SentencePiece pieces = 2;
50
+ */
51
+ pieces: SentencePieceText_SentencePiece[];
52
+
53
+ /**
54
+ * Score (usually log probability) for MultiSentencePieceText.
55
+ *
56
+ * @generated from field: optional float score = 3;
57
+ */
58
+ score: number;
59
+ };
60
+
61
+ /**
62
+ * Describes the message sentencepiece.SentencePieceText.
63
+ * Use `create(SentencePieceTextSchema)` to create a new message.
64
+ */
65
+ export declare const SentencePieceTextSchema: GenMessage<SentencePieceText>;
66
+
67
+ /**
68
+ * @generated from message sentencepiece.SentencePieceText.SentencePiece
69
+ */
70
+ export declare type SentencePieceText_SentencePiece = Message<"sentencepiece.SentencePieceText.SentencePiece"> & {
71
+ /**
72
+ * Internal representation for the decoder.
73
+ * - Decoder can use |piece| as a basic token.
74
+ * - the piece must be non-empty.
75
+ * - A whitespace is replaced with a meta symbol.
76
+ * - Concatenation of pieces is not always the same as the |text|.
77
+ *
78
+ * @generated from field: optional string piece = 1;
79
+ */
80
+ piece: string;
81
+
82
+ /**
83
+ * Vocabulary id.
84
+ *
85
+ * @generated from field: optional uint32 id = 2;
86
+ */
87
+ id: number;
88
+
89
+ /**
90
+ * External representation for the client.
91
+ * - It is always guaranteed that
92
+ * text.substr(begin, end - begin) == surface.
93
+ * - Concatenation of surface is always the same as the |text|.
94
+ * - |surface| may contain whitespaces.
95
+ * - |surface| may be empty if the piece encodes
96
+ * a control vocabulary. e.g., <s>, </s>, <unk>.
97
+ * - When |surface| is empty, always begin == end. (zero-length span).
98
+ *
99
+ * @generated from field: optional string surface = 3;
100
+ */
101
+ surface: string;
102
+
103
+ /**
104
+ * @generated from field: optional uint32 begin = 4;
105
+ */
106
+ begin: number;
107
+
108
+ /**
109
+ * @generated from field: optional uint32 end = 5;
110
+ */
111
+ end: number;
112
+ };
113
+
114
+ /**
115
+ * Describes the message sentencepiece.SentencePieceText.SentencePiece.
116
+ * Use `create(SentencePieceText_SentencePieceSchema)` to create a new message.
117
+ */
118
+ export declare const SentencePieceText_SentencePieceSchema: GenMessage<SentencePieceText_SentencePiece>;
119
+
120
+ /**
121
+ * @generated from message sentencepiece.NBestSentencePieceText
122
+ */
123
+ export declare type NBestSentencePieceText = Message<"sentencepiece.NBestSentencePieceText"> & {
124
+ /**
125
+ * @generated from field: repeated sentencepiece.SentencePieceText nbests = 1;
126
+ */
127
+ nbests: SentencePieceText[];
128
+ };
129
+
130
+ /**
131
+ * Describes the message sentencepiece.NBestSentencePieceText.
132
+ * Use `create(NBestSentencePieceTextSchema)` to create a new message.
133
+ */
134
+ export declare const NBestSentencePieceTextSchema: GenMessage<NBestSentencePieceText>;
135
+
@@ -0,0 +1,47 @@
1
+ // Copyright 2016 Google Inc.
2
+ //
3
+ // Licensed under the Apache License, Version 2.0 (the "License");
4
+ // you may not use this file except in compliance with the License.
5
+ // You may obtain a copy of the License at
6
+ //
7
+ // http://www.apache.org/licenses/LICENSE-2.0
8
+ //
9
+ // Unless required by applicable law or agreed to in writing, software
10
+ // distributed under the License is distributed on an "AS IS" BASIS,
11
+ // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+ // See the License for the specific language governing permissions and
13
+ // limitations under the License.!
14
+
15
+ // @generated by protoc-gen-es v2.10.2 with parameter "target=js+dts,import_extension=js,js_import_style=module"
16
+ // @generated from file sentencepiece.proto (package sentencepiece, syntax proto2)
17
+ /* eslint-disable */
18
+
19
+ import { fileDesc, messageDesc } from "@bufbuild/protobuf/codegenv2";
20
+
21
+ /**
22
+ * Describes the file sentencepiece.proto.
23
+ */
24
+ export const file_sentencepiece = /*@__PURE__*/
25
+ fileDesc("ChNzZW50ZW5jZXBpZWNlLnByb3RvEg1zZW50ZW5jZXBpZWNlIt8BChFTZW50ZW5jZVBpZWNlVGV4dBIMCgR0ZXh0GAEgASgJEj4KBnBpZWNlcxgCIAMoCzIuLnNlbnRlbmNlcGllY2UuU2VudGVuY2VQaWVjZVRleHQuU2VudGVuY2VQaWVjZRINCgVzY29yZRgDIAEoAhpiCg1TZW50ZW5jZVBpZWNlEg0KBXBpZWNlGAEgASgJEgoKAmlkGAIgASgNEg8KB3N1cmZhY2UYAyABKAkSDQoFYmVnaW4YBCABKA0SCwoDZW5kGAUgASgNKgkIyAEQgICAgAIqCQjIARCAgICAAiJKChZOQmVzdFNlbnRlbmNlUGllY2VUZXh0EjAKBm5iZXN0cxgBIAMoCzIgLnNlbnRlbmNlcGllY2UuU2VudGVuY2VQaWVjZVRleHRCAkgD");
26
+
27
+ /**
28
+ * Describes the message sentencepiece.SentencePieceText.
29
+ * Use `create(SentencePieceTextSchema)` to create a new message.
30
+ */
31
+ export const SentencePieceTextSchema = /*@__PURE__*/
32
+ messageDesc(file_sentencepiece, 0);
33
+
34
+ /**
35
+ * Describes the message sentencepiece.SentencePieceText.SentencePiece.
36
+ * Use `create(SentencePieceText_SentencePieceSchema)` to create a new message.
37
+ */
38
+ export const SentencePieceText_SentencePieceSchema = /*@__PURE__*/
39
+ messageDesc(file_sentencepiece, 0, 0);
40
+
41
+ /**
42
+ * Describes the message sentencepiece.NBestSentencePieceText.
43
+ * Use `create(NBestSentencePieceTextSchema)` to create a new message.
44
+ */
45
+ export const NBestSentencePieceTextSchema = /*@__PURE__*/
46
+ messageDesc(file_sentencepiece, 1);
47
+
package/package.json ADDED
@@ -0,0 +1,49 @@
1
+ {
2
+ "name": "sentencepiece-buf",
3
+ "version": "0.2.1-0",
4
+ "description": "Compiled Protobuf definitions for SentencePiece files",
5
+ "keywords": [
6
+ "sentencepiece",
7
+ "protobuf"
8
+ ],
9
+ "author": {
10
+ "name": "Eric Zhang",
11
+ "email": "ekzhang1@gmail.com",
12
+ "url": "https://www.ekzhang.com"
13
+ },
14
+ "repository": {
15
+ "type": "git",
16
+ "url": "git+https://github.com/ekzhang/protobuf-es-pkg.git"
17
+ },
18
+ "type": "module",
19
+ "files": [
20
+ "/dist/*.{js,d.ts}"
21
+ ],
22
+ "main": "dist/sentencepiece_pb.js",
23
+ "types": "dist/sentencepiece_pb.d.ts",
24
+ "module": "dist/sentencepiece_pb.js",
25
+ "exports": {
26
+ ".": {
27
+ "import": "./dist/sentencepiece_pb.js",
28
+ "types": "./dist/sentencepiece_pb.d.ts"
29
+ },
30
+ "./model": {
31
+ "import": "./dist/sentencepiece_model_pb.js",
32
+ "types": "./dist/sentencepiece_model_pb.d.ts"
33
+ }
34
+ },
35
+ "license": "MIT",
36
+ "dependencies": {
37
+ "@bufbuild/protobuf": "^2.10.2"
38
+ },
39
+ "devDependencies": {
40
+ "@bufbuild/buf": "^1.61.0",
41
+ "@bufbuild/protoc-gen-es": "^2.10.2"
42
+ },
43
+ "engines": {
44
+ "pnpm": ">=10.0.0"
45
+ },
46
+ "scripts": {
47
+ "build": "buf generate && perl -pi -e 's/Uint8Array(?!<)/Uint8Array<ArrayBuffer>/g' dist/*.d.ts"
48
+ }
49
+ }