keras-hub-nightly 0.16.1.dev202410020340__py3-none-any.whl → 0.19.0.dev202501260345__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- keras_hub/api/layers/__init__.py +21 -3
- keras_hub/api/models/__init__.py +71 -12
- keras_hub/api/tokenizers/__init__.py +1 -1
- keras_hub/src/bounding_box/__init__.py +2 -0
- keras_hub/src/bounding_box/converters.py +102 -12
- keras_hub/src/layers/modeling/f_net_encoder.py +1 -1
- keras_hub/src/layers/modeling/masked_lm_head.py +2 -1
- keras_hub/src/layers/modeling/reversible_embedding.py +3 -16
- keras_hub/src/layers/modeling/rms_normalization.py +36 -0
- keras_hub/src/layers/modeling/rotary_embedding.py +3 -2
- keras_hub/src/layers/modeling/token_and_position_embedding.py +1 -1
- keras_hub/src/layers/modeling/transformer_decoder.py +8 -6
- keras_hub/src/layers/modeling/transformer_encoder.py +29 -7
- keras_hub/src/layers/preprocessing/audio_converter.py +3 -7
- keras_hub/src/layers/preprocessing/image_converter.py +170 -34
- keras_hub/src/metrics/bleu.py +4 -3
- keras_hub/src/models/albert/albert_presets.py +4 -12
- keras_hub/src/models/albert/albert_text_classifier.py +7 -7
- keras_hub/src/models/backbone.py +3 -14
- keras_hub/src/models/bart/bart_backbone.py +4 -4
- keras_hub/src/models/bart/bart_presets.py +3 -9
- keras_hub/src/models/bart/bart_seq_2_seq_lm.py +9 -8
- keras_hub/src/models/basnet/__init__.py +5 -0
- keras_hub/src/models/basnet/basnet.py +122 -0
- keras_hub/src/models/basnet/basnet_backbone.py +366 -0
- keras_hub/src/models/basnet/basnet_image_converter.py +8 -0
- keras_hub/src/models/basnet/basnet_preprocessor.py +14 -0
- keras_hub/src/models/basnet/basnet_presets.py +17 -0
- keras_hub/src/models/bert/bert_presets.py +14 -32
- keras_hub/src/models/bert/bert_text_classifier.py +3 -3
- keras_hub/src/models/bloom/bloom_presets.py +8 -24
- keras_hub/src/models/causal_lm.py +56 -12
- keras_hub/src/models/clip/__init__.py +5 -0
- keras_hub/src/models/clip/clip_backbone.py +286 -0
- keras_hub/src/models/clip/clip_encoder_block.py +19 -4
- keras_hub/src/models/clip/clip_image_converter.py +8 -0
- keras_hub/src/models/clip/clip_presets.py +93 -0
- keras_hub/src/models/clip/clip_text_encoder.py +4 -1
- keras_hub/src/models/clip/clip_tokenizer.py +18 -3
- keras_hub/src/models/clip/clip_vision_embedding.py +101 -0
- keras_hub/src/models/clip/clip_vision_encoder.py +159 -0
- keras_hub/src/models/csp_darknet/csp_darknet_backbone.py +2 -1
- keras_hub/src/models/csp_darknet/csp_darknet_image_classifier.py +0 -109
- keras_hub/src/models/deberta_v3/deberta_v3_backbone.py +1 -1
- keras_hub/src/models/deberta_v3/deberta_v3_presets.py +5 -15
- keras_hub/src/models/deberta_v3/deberta_v3_text_classifier.py +4 -4
- keras_hub/src/models/deberta_v3/disentangled_attention_encoder.py +4 -4
- keras_hub/src/models/deberta_v3/disentangled_self_attention.py +3 -2
- keras_hub/src/models/deberta_v3/relative_embedding.py +1 -1
- keras_hub/src/models/deeplab_v3/__init__.py +7 -0
- keras_hub/src/models/deeplab_v3/deeplab_v3_backbone.py +200 -0
- keras_hub/src/models/deeplab_v3/deeplab_v3_image_converter.py +10 -0
- keras_hub/src/models/deeplab_v3/deeplab_v3_image_segmeter_preprocessor.py +16 -0
- keras_hub/src/models/deeplab_v3/deeplab_v3_layers.py +215 -0
- keras_hub/src/models/deeplab_v3/deeplab_v3_presets.py +17 -0
- keras_hub/src/models/deeplab_v3/deeplab_v3_segmenter.py +111 -0
- keras_hub/src/models/densenet/densenet_backbone.py +6 -4
- keras_hub/src/models/densenet/densenet_image_classifier.py +1 -129
- keras_hub/src/models/densenet/densenet_image_converter.py +2 -4
- keras_hub/src/models/densenet/densenet_presets.py +9 -15
- keras_hub/src/models/distil_bert/distil_bert_masked_lm.py +1 -1
- keras_hub/src/models/distil_bert/distil_bert_masked_lm_preprocessor.py +2 -2
- keras_hub/src/models/distil_bert/distil_bert_presets.py +5 -10
- keras_hub/src/models/distil_bert/distil_bert_text_classifier.py +5 -5
- keras_hub/src/models/distil_bert/distil_bert_tokenizer.py +3 -3
- keras_hub/src/models/efficientnet/__init__.py +9 -0
- keras_hub/src/models/efficientnet/cba.py +141 -0
- keras_hub/src/models/efficientnet/efficientnet_backbone.py +160 -61
- keras_hub/src/models/efficientnet/efficientnet_image_classifier.py +14 -0
- keras_hub/src/models/efficientnet/efficientnet_image_classifier_preprocessor.py +16 -0
- keras_hub/src/models/efficientnet/efficientnet_image_converter.py +10 -0
- keras_hub/src/models/efficientnet/efficientnet_presets.py +193 -0
- keras_hub/src/models/efficientnet/fusedmbconv.py +84 -41
- keras_hub/src/models/efficientnet/mbconv.py +53 -22
- keras_hub/src/models/electra/electra_backbone.py +2 -2
- keras_hub/src/models/electra/electra_presets.py +6 -18
- keras_hub/src/models/f_net/f_net_presets.py +2 -6
- keras_hub/src/models/f_net/f_net_text_classifier.py +3 -3
- keras_hub/src/models/f_net/f_net_text_classifier_preprocessor.py +3 -3
- keras_hub/src/models/falcon/falcon_backbone.py +5 -3
- keras_hub/src/models/falcon/falcon_causal_lm.py +18 -8
- keras_hub/src/models/falcon/falcon_presets.py +1 -3
- keras_hub/src/models/falcon/falcon_tokenizer.py +7 -2
- keras_hub/src/models/feature_pyramid_backbone.py +1 -1
- keras_hub/src/models/flux/__init__.py +5 -0
- keras_hub/src/models/flux/flux_layers.py +496 -0
- keras_hub/src/models/flux/flux_maths.py +225 -0
- keras_hub/src/models/flux/flux_model.py +236 -0
- keras_hub/src/models/flux/flux_presets.py +3 -0
- keras_hub/src/models/flux/flux_text_to_image.py +146 -0
- keras_hub/src/models/flux/flux_text_to_image_preprocessor.py +73 -0
- keras_hub/src/models/gemma/gemma_backbone.py +35 -20
- keras_hub/src/models/gemma/gemma_causal_lm.py +2 -2
- keras_hub/src/models/gemma/gemma_decoder_block.py +3 -1
- keras_hub/src/models/gemma/gemma_presets.py +29 -63
- keras_hub/src/models/gpt2/gpt2_causal_lm.py +2 -2
- keras_hub/src/models/gpt2/gpt2_presets.py +5 -14
- keras_hub/src/models/gpt_neo_x/gpt_neo_x_attention.py +2 -1
- keras_hub/src/models/gpt_neo_x/gpt_neo_x_causal_lm.py +3 -3
- keras_hub/src/models/gpt_neo_x/gpt_neo_x_decoder.py +2 -1
- keras_hub/src/models/image_classifier.py +147 -2
- keras_hub/src/models/image_classifier_preprocessor.py +6 -3
- keras_hub/src/models/image_object_detector.py +87 -0
- keras_hub/src/models/image_object_detector_preprocessor.py +57 -0
- keras_hub/src/models/image_segmenter.py +0 -5
- keras_hub/src/models/image_segmenter_preprocessor.py +29 -4
- keras_hub/src/models/image_to_image.py +417 -0
- keras_hub/src/models/inpaint.py +520 -0
- keras_hub/src/models/llama/llama_backbone.py +138 -12
- keras_hub/src/models/llama/llama_causal_lm.py +3 -1
- keras_hub/src/models/llama/llama_presets.py +10 -20
- keras_hub/src/models/llama3/llama3_backbone.py +12 -11
- keras_hub/src/models/llama3/llama3_causal_lm.py +1 -1
- keras_hub/src/models/llama3/llama3_presets.py +4 -12
- keras_hub/src/models/llama3/llama3_tokenizer.py +25 -2
- keras_hub/src/models/mistral/mistral_backbone.py +16 -15
- keras_hub/src/models/mistral/mistral_causal_lm.py +6 -4
- keras_hub/src/models/mistral/mistral_presets.py +3 -9
- keras_hub/src/models/mistral/mistral_transformer_decoder.py +2 -1
- keras_hub/src/models/mit/__init__.py +6 -0
- keras_hub/src/models/{mix_transformer/mix_transformer_backbone.py → mit/mit_backbone.py} +47 -36
- keras_hub/src/models/mit/mit_image_classifier.py +12 -0
- keras_hub/src/models/mit/mit_image_classifier_preprocessor.py +12 -0
- keras_hub/src/models/mit/mit_image_converter.py +8 -0
- keras_hub/src/models/{mix_transformer/mix_transformer_layers.py → mit/mit_layers.py} +20 -13
- keras_hub/src/models/mit/mit_presets.py +139 -0
- keras_hub/src/models/mobilenet/mobilenet_backbone.py +8 -8
- keras_hub/src/models/mobilenet/mobilenet_image_classifier.py +0 -92
- keras_hub/src/models/opt/opt_causal_lm.py +2 -2
- keras_hub/src/models/opt/opt_presets.py +4 -12
- keras_hub/src/models/pali_gemma/pali_gemma_backbone.py +63 -17
- keras_hub/src/models/pali_gemma/pali_gemma_causal_lm.py +3 -1
- keras_hub/src/models/pali_gemma/pali_gemma_decoder_block.py +21 -23
- keras_hub/src/models/pali_gemma/pali_gemma_image_converter.py +2 -4
- keras_hub/src/models/pali_gemma/pali_gemma_presets.py +173 -17
- keras_hub/src/models/pali_gemma/pali_gemma_vit.py +14 -26
- keras_hub/src/models/phi3/phi3_causal_lm.py +3 -1
- keras_hub/src/models/phi3/phi3_decoder.py +0 -1
- keras_hub/src/models/phi3/phi3_presets.py +2 -6
- keras_hub/src/models/phi3/phi3_rotary_embedding.py +1 -1
- keras_hub/src/models/preprocessor.py +25 -11
- keras_hub/src/models/resnet/resnet_backbone.py +3 -14
- keras_hub/src/models/resnet/resnet_image_classifier.py +0 -137
- keras_hub/src/models/resnet/resnet_image_converter.py +2 -4
- keras_hub/src/models/resnet/resnet_presets.py +127 -18
- keras_hub/src/models/retinanet/__init__.py +5 -0
- keras_hub/src/models/retinanet/anchor_generator.py +52 -53
- keras_hub/src/models/retinanet/feature_pyramid.py +103 -39
- keras_hub/src/models/retinanet/non_max_supression.py +1 -0
- keras_hub/src/models/retinanet/prediction_head.py +192 -0
- keras_hub/src/models/retinanet/retinanet_backbone.py +146 -0
- keras_hub/src/models/retinanet/retinanet_image_converter.py +53 -0
- keras_hub/src/models/retinanet/retinanet_label_encoder.py +49 -51
- keras_hub/src/models/retinanet/retinanet_object_detector.py +381 -0
- keras_hub/src/models/retinanet/retinanet_object_detector_preprocessor.py +14 -0
- keras_hub/src/models/retinanet/retinanet_presets.py +16 -0
- keras_hub/src/models/roberta/roberta_backbone.py +2 -2
- keras_hub/src/models/roberta/roberta_presets.py +6 -8
- keras_hub/src/models/roberta/roberta_text_classifier.py +3 -3
- keras_hub/src/models/sam/__init__.py +5 -0
- keras_hub/src/models/sam/sam_backbone.py +2 -3
- keras_hub/src/models/sam/sam_image_converter.py +2 -4
- keras_hub/src/models/sam/sam_image_segmenter.py +16 -16
- keras_hub/src/models/sam/sam_image_segmenter_preprocessor.py +11 -1
- keras_hub/src/models/sam/sam_layers.py +5 -3
- keras_hub/src/models/sam/sam_presets.py +3 -9
- keras_hub/src/models/sam/sam_prompt_encoder.py +4 -2
- keras_hub/src/models/sam/sam_transformer.py +5 -4
- keras_hub/src/models/segformer/__init__.py +8 -0
- keras_hub/src/models/segformer/segformer_backbone.py +167 -0
- keras_hub/src/models/segformer/segformer_image_converter.py +8 -0
- keras_hub/src/models/segformer/segformer_image_segmenter.py +184 -0
- keras_hub/src/models/segformer/segformer_image_segmenter_preprocessor.py +31 -0
- keras_hub/src/models/segformer/segformer_presets.py +136 -0
- keras_hub/src/models/seq_2_seq_lm_preprocessor.py +1 -1
- keras_hub/src/models/stable_diffusion_3/flow_match_euler_discrete_scheduler.py +8 -1
- keras_hub/src/models/stable_diffusion_3/mmdit.py +577 -190
- keras_hub/src/models/stable_diffusion_3/stable_diffusion_3_backbone.py +189 -163
- keras_hub/src/models/stable_diffusion_3/stable_diffusion_3_image_to_image.py +178 -0
- keras_hub/src/models/stable_diffusion_3/stable_diffusion_3_inpaint.py +193 -0
- keras_hub/src/models/stable_diffusion_3/stable_diffusion_3_presets.py +43 -7
- keras_hub/src/models/stable_diffusion_3/stable_diffusion_3_text_to_image.py +25 -14
- keras_hub/src/models/stable_diffusion_3/stable_diffusion_3_text_to_image_preprocessor.py +1 -1
- keras_hub/src/models/t5/t5_backbone.py +5 -4
- keras_hub/src/models/t5/t5_presets.py +47 -19
- keras_hub/src/models/task.py +47 -39
- keras_hub/src/models/text_classifier.py +2 -2
- keras_hub/src/models/text_to_image.py +106 -41
- keras_hub/src/models/vae/__init__.py +1 -0
- keras_hub/src/models/vae/vae_backbone.py +184 -0
- keras_hub/src/models/vae/vae_layers.py +739 -0
- keras_hub/src/models/vgg/__init__.py +5 -0
- keras_hub/src/models/vgg/vgg_backbone.py +4 -24
- keras_hub/src/models/vgg/vgg_image_classifier.py +139 -33
- keras_hub/src/models/vgg/vgg_image_classifier_preprocessor.py +12 -0
- keras_hub/src/models/vgg/vgg_image_converter.py +8 -0
- keras_hub/src/models/vgg/vgg_presets.py +48 -0
- keras_hub/src/models/vit/__init__.py +5 -0
- keras_hub/src/models/vit/vit_backbone.py +152 -0
- keras_hub/src/models/vit/vit_image_classifier.py +187 -0
- keras_hub/src/models/vit/vit_image_classifier_preprocessor.py +12 -0
- keras_hub/src/models/vit/vit_image_converter.py +73 -0
- keras_hub/src/models/vit/vit_layers.py +391 -0
- keras_hub/src/models/vit/vit_presets.py +126 -0
- keras_hub/src/models/vit_det/vit_det_backbone.py +6 -4
- keras_hub/src/models/vit_det/vit_layers.py +3 -3
- keras_hub/src/models/whisper/whisper_audio_converter.py +2 -4
- keras_hub/src/models/whisper/whisper_backbone.py +6 -5
- keras_hub/src/models/whisper/whisper_decoder.py +3 -5
- keras_hub/src/models/whisper/whisper_presets.py +10 -30
- keras_hub/src/models/xlm_roberta/xlm_roberta_masked_lm.py +1 -1
- keras_hub/src/models/xlm_roberta/xlm_roberta_masked_lm_preprocessor.py +2 -2
- keras_hub/src/models/xlm_roberta/xlm_roberta_presets.py +2 -6
- keras_hub/src/models/xlm_roberta/xlm_roberta_text_classifier.py +4 -4
- keras_hub/src/models/xlm_roberta/xlm_roberta_tokenizer.py +2 -1
- keras_hub/src/models/xlnet/relative_attention.py +20 -19
- keras_hub/src/models/xlnet/xlnet_backbone.py +2 -2
- keras_hub/src/models/xlnet/xlnet_content_and_query_embedding.py +3 -5
- keras_hub/src/models/xlnet/xlnet_encoder.py +7 -9
- keras_hub/src/samplers/contrastive_sampler.py +2 -3
- keras_hub/src/samplers/sampler.py +2 -1
- keras_hub/src/tests/test_case.py +41 -6
- keras_hub/src/tokenizers/byte_pair_tokenizer.py +7 -3
- keras_hub/src/tokenizers/byte_tokenizer.py +3 -10
- keras_hub/src/tokenizers/sentence_piece_tokenizer.py +2 -9
- keras_hub/src/tokenizers/sentence_piece_tokenizer_trainer.py +9 -11
- keras_hub/src/tokenizers/tokenizer.py +10 -13
- keras_hub/src/tokenizers/unicode_codepoint_tokenizer.py +9 -7
- keras_hub/src/tokenizers/word_piece_tokenizer_trainer.py +10 -3
- keras_hub/src/utils/keras_utils.py +2 -13
- keras_hub/src/utils/pipeline_model.py +3 -3
- keras_hub/src/utils/preset_utils.py +196 -144
- keras_hub/src/utils/tensor_utils.py +4 -4
- keras_hub/src/utils/timm/convert_densenet.py +6 -4
- keras_hub/src/utils/timm/convert_efficientnet.py +447 -0
- keras_hub/src/utils/timm/convert_resnet.py +1 -1
- keras_hub/src/utils/timm/convert_vgg.py +85 -0
- keras_hub/src/utils/timm/preset_loader.py +14 -9
- keras_hub/src/utils/transformers/convert_llama3.py +21 -5
- keras_hub/src/utils/transformers/convert_vit.py +150 -0
- keras_hub/src/utils/transformers/preset_loader.py +23 -0
- keras_hub/src/utils/transformers/safetensor_utils.py +4 -3
- keras_hub/src/version_utils.py +1 -1
- {keras_hub_nightly-0.16.1.dev202410020340.dist-info → keras_hub_nightly-0.19.0.dev202501260345.dist-info}/METADATA +86 -68
- keras_hub_nightly-0.19.0.dev202501260345.dist-info/RECORD +423 -0
- {keras_hub_nightly-0.16.1.dev202410020340.dist-info → keras_hub_nightly-0.19.0.dev202501260345.dist-info}/WHEEL +1 -1
- keras_hub/src/layers/preprocessing/resizing_image_converter.py +0 -138
- keras_hub/src/models/mix_transformer/__init__.py +0 -0
- keras_hub/src/models/mix_transformer/mix_transformer_classifier.py +0 -119
- keras_hub/src/models/stable_diffusion_3/vae_image_decoder.py +0 -320
- keras_hub_nightly-0.16.1.dev202410020340.dist-info/RECORD +0 -357
- {keras_hub_nightly-0.16.1.dev202410020340.dist-info → keras_hub_nightly-0.19.0.dev202501260345.dist-info}/top_level.txt +0 -0
@@ -0,0 +1,31 @@
|
|
1
|
+
import keras
|
2
|
+
|
3
|
+
from keras_hub.src.api_export import keras_hub_export
|
4
|
+
from keras_hub.src.models.image_segmenter_preprocessor import (
|
5
|
+
ImageSegmenterPreprocessor,
|
6
|
+
)
|
7
|
+
from keras_hub.src.models.segformer.segformer_backbone import SegFormerBackbone
|
8
|
+
from keras_hub.src.models.segformer.segformer_image_converter import (
|
9
|
+
SegFormerImageConverter,
|
10
|
+
)
|
11
|
+
from keras_hub.src.utils.tensor_utils import preprocessing_function
|
12
|
+
|
13
|
+
IMAGENET_DEFAULT_MEAN = [0.485, 0.456, 0.406]
|
14
|
+
IMAGENET_DEFAULT_STD = [0.229, 0.224, 0.225]
|
15
|
+
|
16
|
+
|
17
|
+
@keras_hub_export("keras_hub.models.SegFormerImageSegmenterPreprocessor")
|
18
|
+
class SegFormerImageSegmenterPreprocessor(ImageSegmenterPreprocessor):
|
19
|
+
backbone_cls = SegFormerBackbone
|
20
|
+
image_converter_cls = SegFormerImageConverter
|
21
|
+
|
22
|
+
@preprocessing_function
|
23
|
+
def call(self, x, y=None, sample_weight=None):
|
24
|
+
if self.image_converter:
|
25
|
+
x = self.image_converter(x)
|
26
|
+
y = self.image_converter(y)
|
27
|
+
|
28
|
+
x = x / 255
|
29
|
+
x = (x - IMAGENET_DEFAULT_MEAN) / IMAGENET_DEFAULT_STD
|
30
|
+
|
31
|
+
return keras.utils.pack_x_y_sample_weight(x, y, sample_weight)
|
@@ -0,0 +1,136 @@
|
|
1
|
+
"""SegFormer model preset configurations."""
|
2
|
+
|
3
|
+
presets = {
|
4
|
+
"segformer_b0_ade20k_512": {
|
5
|
+
"metadata": {
|
6
|
+
"description": (
|
7
|
+
"SegFormer model with MiTB0 backbone fine-tuned on ADE20k in "
|
8
|
+
"512x512 resolution."
|
9
|
+
),
|
10
|
+
"params": 3719027,
|
11
|
+
"path": "segformer_b0",
|
12
|
+
},
|
13
|
+
"kaggle_handle": "kaggle://keras/segformer/keras/segformer_b0_ade20k_512/2",
|
14
|
+
},
|
15
|
+
"segformer_b1_ade20k_512": {
|
16
|
+
"metadata": {
|
17
|
+
"description": (
|
18
|
+
"SegFormer model with MiTB1 backbone fine-tuned on ADE20k in "
|
19
|
+
"512x512 resolution."
|
20
|
+
),
|
21
|
+
"params": 13682643,
|
22
|
+
"path": "segformer_b1",
|
23
|
+
},
|
24
|
+
"kaggle_handle": "kaggle://keras/segformer/keras/segformer_b1_ade20k_512/2",
|
25
|
+
},
|
26
|
+
"segformer_b2_ade20k_512": {
|
27
|
+
"metadata": {
|
28
|
+
"description": (
|
29
|
+
"SegFormer model with MiTB2 backbone fine-tuned on ADE20k in "
|
30
|
+
"512x512 resolution."
|
31
|
+
),
|
32
|
+
"params": 24727507,
|
33
|
+
"path": "segformer_b2",
|
34
|
+
},
|
35
|
+
"kaggle_handle": "kaggle://keras/segformer/keras/segformer_b2_ade20k_512/2",
|
36
|
+
},
|
37
|
+
"segformer_b3_ade20k_512": {
|
38
|
+
"metadata": {
|
39
|
+
"description": (
|
40
|
+
"SegFormer model with MiTB3 backbone fine-tuned on ADE20k in "
|
41
|
+
"512x512 resolution."
|
42
|
+
),
|
43
|
+
"params": 44603347,
|
44
|
+
"path": "segformer_b3",
|
45
|
+
},
|
46
|
+
"kaggle_handle": "kaggle://keras/segformer/keras/segformer_b3_ade20k_512/2",
|
47
|
+
},
|
48
|
+
"segformer_b4_ade20k_512": {
|
49
|
+
"metadata": {
|
50
|
+
"description": (
|
51
|
+
"SegFormer model with MiTB4 backbone fine-tuned on ADE20k in "
|
52
|
+
"512x512 resolution."
|
53
|
+
),
|
54
|
+
"params": 61373907,
|
55
|
+
"path": "segformer_b4",
|
56
|
+
},
|
57
|
+
"kaggle_handle": "kaggle://keras/segformer/keras/segformer_b4_ade20k_512/2",
|
58
|
+
},
|
59
|
+
"segformer_b5_ade20k_640": {
|
60
|
+
"metadata": {
|
61
|
+
"description": (
|
62
|
+
"SegFormer model with MiTB5 backbone fine-tuned on ADE20k in "
|
63
|
+
"640x640 resolution."
|
64
|
+
),
|
65
|
+
"params": 81974227,
|
66
|
+
"path": "segformer_b5",
|
67
|
+
},
|
68
|
+
"kaggle_handle": "kaggle://keras/segformer/keras/segformer_b5_ade20k_640/2",
|
69
|
+
},
|
70
|
+
"segformer_b0_cityscapes_1024": {
|
71
|
+
"metadata": {
|
72
|
+
"description": (
|
73
|
+
"SegFormer model with MiTB0 backbone fine-tuned on Cityscapes "
|
74
|
+
"in 1024x1024 resolution."
|
75
|
+
),
|
76
|
+
"params": 3719027,
|
77
|
+
"path": "segformer_b0",
|
78
|
+
},
|
79
|
+
"kaggle_handle": "kaggle://keras/segformer/keras/segformer_b0_cityscapes_1024/2",
|
80
|
+
},
|
81
|
+
"segformer_b1_cityscapes_1024": {
|
82
|
+
"metadata": {
|
83
|
+
"description": (
|
84
|
+
"SegFormer model with MiTB1 backbone fine-tuned on Cityscapes "
|
85
|
+
"in 1024x1024 resolution."
|
86
|
+
),
|
87
|
+
"params": 13682643,
|
88
|
+
"path": "segformer_b1",
|
89
|
+
},
|
90
|
+
"kaggle_handle": "kaggle://keras/segformer/keras/segformer_b1_ade20k_512/2",
|
91
|
+
},
|
92
|
+
"segformer_b2_cityscapes_1024": {
|
93
|
+
"metadata": {
|
94
|
+
"description": (
|
95
|
+
"SegFormer model with MiTB2 backbone fine-tuned on Cityscapes "
|
96
|
+
"in 1024x1024 resolution."
|
97
|
+
),
|
98
|
+
"params": 24727507,
|
99
|
+
"path": "segformer_b2",
|
100
|
+
},
|
101
|
+
"kaggle_handle": "kaggle://keras/segformer/keras/segformer_b2_cityscapes_1024/2",
|
102
|
+
},
|
103
|
+
"segformer_b3_cityscapes_1024": {
|
104
|
+
"metadata": {
|
105
|
+
"description": (
|
106
|
+
"SegFormer model with MiTB3 backbone fine-tuned on Cityscapes "
|
107
|
+
"in 1024x1024 resolution."
|
108
|
+
),
|
109
|
+
"params": 44603347,
|
110
|
+
"path": "segformer_b3",
|
111
|
+
},
|
112
|
+
"kaggle_handle": "kaggle://keras/segformer/keras/segformer_b3_cityscapes_1024/2",
|
113
|
+
},
|
114
|
+
"segformer_b4_cityscapes_1024": {
|
115
|
+
"metadata": {
|
116
|
+
"description": (
|
117
|
+
"SegFormer model with MiTB4 backbone fine-tuned on Cityscapes "
|
118
|
+
"in 1024x1024 resolution."
|
119
|
+
),
|
120
|
+
"params": 61373907,
|
121
|
+
"path": "segformer_b4",
|
122
|
+
},
|
123
|
+
"kaggle_handle": "kaggle://keras/segformer/keras/segformer_b4_cityscapes_1024/2",
|
124
|
+
},
|
125
|
+
"segformer_b5_cityscapes_1024": {
|
126
|
+
"metadata": {
|
127
|
+
"description": (
|
128
|
+
"SegFormer model with MiTB5 backbone fine-tuned on Cityscapes "
|
129
|
+
"in 1024x1024 resolution."
|
130
|
+
),
|
131
|
+
"params": 81974227,
|
132
|
+
"path": "segformer_b5",
|
133
|
+
},
|
134
|
+
"kaggle_handle": "kaggle://keras/segformer/keras/segformer_b5_cityscapes_1024/2",
|
135
|
+
},
|
136
|
+
}
|
@@ -151,7 +151,7 @@ class Seq2SeqLMPreprocessor(Preprocessor):
|
|
151
151
|
# `sequence_length` is an alias for `decoder_sequence_length`
|
152
152
|
sequence_length=None,
|
153
153
|
):
|
154
|
-
"""Convert
|
154
|
+
"""Convert input strings to integer token inputs for generation.
|
155
155
|
|
156
156
|
Similar to calling the layer for training, this method takes in a dict
|
157
157
|
containing `"encoder_text"` and `"decoder_text"`, with strings or tensor
|
@@ -27,7 +27,7 @@ class FlowMatchEulerDiscreteScheduler(layers.Layer):
|
|
27
27
|
https://arxiv.org/abs/2403.03206).
|
28
28
|
"""
|
29
29
|
|
30
|
-
def __init__(self, num_train_timesteps=1000, shift=
|
30
|
+
def __init__(self, num_train_timesteps=1000, shift=3.0, **kwargs):
|
31
31
|
super().__init__(**kwargs)
|
32
32
|
self.num_train_timesteps = int(num_train_timesteps)
|
33
33
|
self.shift = float(shift)
|
@@ -65,6 +65,13 @@ class FlowMatchEulerDiscreteScheduler(layers.Layer):
|
|
65
65
|
timestep = self._sigma_to_timestep(sigma)
|
66
66
|
return sigma, timestep
|
67
67
|
|
68
|
+
def add_noise(self, inputs, noises, step, num_steps):
|
69
|
+
sigma, _ = self(step, num_steps)
|
70
|
+
return ops.add(
|
71
|
+
ops.multiply(sigma, noises),
|
72
|
+
ops.multiply(ops.subtract(1.0, sigma), inputs),
|
73
|
+
)
|
74
|
+
|
68
75
|
def get_config(self):
|
69
76
|
config = super().get_config()
|
70
77
|
config.update(
|