paddlex 2.0.0rc4__py3-none-any.whl → 3.0.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- paddlex/.version +1 -0
- paddlex/__init__.py +35 -18
- paddlex/__main__.py +39 -0
- paddlex/configs/modules/3d_bev_detection/BEVFusion.yaml +38 -0
- paddlex/configs/modules/chart_parsing/PP-Chart2Table.yaml +13 -0
- paddlex/configs/modules/doc_text_orientation/PP-LCNet_x1_0_doc_ori.yaml +41 -0
- paddlex/configs/modules/doc_vlm/PP-DocBee-2B.yaml +14 -0
- paddlex/configs/modules/doc_vlm/PP-DocBee-7B.yaml +14 -0
- paddlex/configs/modules/doc_vlm/PP-DocBee2-3B.yaml +14 -0
- paddlex/configs/modules/face_detection/BlazeFace-FPN-SSH.yaml +40 -0
- paddlex/configs/modules/face_detection/BlazeFace.yaml +40 -0
- paddlex/configs/modules/face_detection/PP-YOLOE_plus-S_face.yaml +40 -0
- paddlex/configs/modules/face_detection/PicoDet_LCNet_x2_5_face.yaml +40 -0
- paddlex/configs/modules/face_feature/MobileFaceNet.yaml +41 -0
- paddlex/configs/modules/face_feature/ResNet50_face.yaml +41 -0
- paddlex/configs/modules/formula_recognition/LaTeX_OCR_rec.yaml +40 -0
- paddlex/configs/modules/formula_recognition/PP-FormulaNet-L.yaml +40 -0
- paddlex/configs/modules/formula_recognition/PP-FormulaNet-S.yaml +40 -0
- paddlex/configs/modules/formula_recognition/PP-FormulaNet_plus-L.yaml +40 -0
- paddlex/configs/modules/formula_recognition/PP-FormulaNet_plus-M.yaml +40 -0
- paddlex/configs/modules/formula_recognition/PP-FormulaNet_plus-S.yaml +40 -0
- paddlex/configs/modules/formula_recognition/UniMERNet.yaml +40 -0
- paddlex/configs/modules/human_detection/PP-YOLOE-L_human.yaml +42 -0
- paddlex/configs/modules/human_detection/PP-YOLOE-S_human.yaml +42 -0
- paddlex/configs/modules/image_anomaly_detection/STFPM.yaml +41 -0
- paddlex/configs/modules/image_classification/CLIP_vit_base_patch16_224.yaml +41 -0
- paddlex/configs/modules/image_classification/CLIP_vit_large_patch14_224.yaml +41 -0
- paddlex/configs/modules/image_classification/ConvNeXt_base_224.yaml +41 -0
- paddlex/configs/modules/image_classification/ConvNeXt_base_384.yaml +41 -0
- paddlex/configs/modules/image_classification/ConvNeXt_large_224.yaml +41 -0
- paddlex/configs/modules/image_classification/ConvNeXt_large_384.yaml +41 -0
- paddlex/configs/modules/image_classification/ConvNeXt_small.yaml +41 -0
- paddlex/configs/modules/image_classification/ConvNeXt_tiny.yaml +41 -0
- paddlex/configs/modules/image_classification/FasterNet-L.yaml +40 -0
- paddlex/configs/modules/image_classification/FasterNet-M.yaml +40 -0
- paddlex/configs/modules/image_classification/FasterNet-S.yaml +40 -0
- paddlex/configs/modules/image_classification/FasterNet-T0.yaml +40 -0
- paddlex/configs/modules/image_classification/FasterNet-T1.yaml +40 -0
- paddlex/configs/modules/image_classification/FasterNet-T2.yaml +40 -0
- paddlex/configs/modules/image_classification/MobileNetV1_x0_25.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV1_x0_5.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV1_x0_75.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV1_x1_0.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV2_x0_25.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV2_x0_5.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV2_x1_0.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV2_x1_5.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV2_x2_0.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV3_large_x0_35.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV3_large_x0_5.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV3_large_x0_75.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV3_large_x1_0.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV3_large_x1_25.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV3_small_x0_35.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV3_small_x0_5.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV3_small_x0_75.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV3_small_x1_0.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV3_small_x1_25.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV4_conv_large.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV4_conv_medium.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV4_conv_small.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV4_hybrid_large.yaml +41 -0
- paddlex/configs/modules/image_classification/MobileNetV4_hybrid_medium.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-HGNetV2-B0.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-HGNetV2-B1.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-HGNetV2-B2.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-HGNetV2-B3.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-HGNetV2-B4.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-HGNetV2-B5.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-HGNetV2-B6.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-HGNet_base.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-HGNet_small.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-HGNet_tiny.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-LCNetV2_base.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-LCNetV2_large.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-LCNetV2_small.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-LCNet_x0_25.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-LCNet_x0_35.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-LCNet_x0_5.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-LCNet_x0_75.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-LCNet_x1_0.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-LCNet_x1_5.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-LCNet_x2_0.yaml +41 -0
- paddlex/configs/modules/image_classification/PP-LCNet_x2_5.yaml +41 -0
- paddlex/configs/modules/image_classification/ResNet101.yaml +41 -0
- paddlex/configs/modules/image_classification/ResNet101_vd.yaml +41 -0
- paddlex/configs/modules/image_classification/ResNet152.yaml +41 -0
- paddlex/configs/modules/image_classification/ResNet152_vd.yaml +41 -0
- paddlex/configs/modules/image_classification/ResNet18.yaml +41 -0
- paddlex/configs/modules/image_classification/ResNet18_vd.yaml +41 -0
- paddlex/configs/modules/image_classification/ResNet200_vd.yaml +41 -0
- paddlex/configs/modules/image_classification/ResNet34.yaml +41 -0
- paddlex/configs/modules/image_classification/ResNet34_vd.yaml +41 -0
- paddlex/configs/modules/image_classification/ResNet50.yaml +41 -0
- paddlex/configs/modules/image_classification/ResNet50_vd.yaml +41 -0
- paddlex/configs/modules/image_classification/StarNet-S1.yaml +41 -0
- paddlex/configs/modules/image_classification/StarNet-S2.yaml +41 -0
- paddlex/configs/modules/image_classification/StarNet-S3.yaml +41 -0
- paddlex/configs/modules/image_classification/StarNet-S4.yaml +41 -0
- paddlex/configs/modules/image_classification/SwinTransformer_base_patch4_window12_384.yaml +41 -0
- paddlex/configs/modules/image_classification/SwinTransformer_base_patch4_window7_224.yaml +41 -0
- paddlex/configs/modules/image_classification/SwinTransformer_large_patch4_window12_384.yaml +41 -0
- paddlex/configs/modules/image_classification/SwinTransformer_large_patch4_window7_224.yaml +41 -0
- paddlex/configs/modules/image_classification/SwinTransformer_small_patch4_window7_224.yaml +41 -0
- paddlex/configs/modules/image_classification/SwinTransformer_tiny_patch4_window7_224.yaml +41 -0
- paddlex/configs/modules/image_feature/PP-ShiTuV2_rec.yaml +42 -0
- paddlex/configs/modules/image_feature/PP-ShiTuV2_rec_CLIP_vit_base.yaml +42 -0
- paddlex/configs/modules/image_feature/PP-ShiTuV2_rec_CLIP_vit_large.yaml +41 -0
- paddlex/configs/modules/image_multilabel_classification/CLIP_vit_base_patch16_448_ML.yaml +41 -0
- paddlex/configs/modules/image_multilabel_classification/PP-HGNetV2-B0_ML.yaml +41 -0
- paddlex/configs/modules/image_multilabel_classification/PP-HGNetV2-B4_ML.yaml +41 -0
- paddlex/configs/modules/image_multilabel_classification/PP-HGNetV2-B6_ML.yaml +41 -0
- paddlex/configs/modules/image_multilabel_classification/PP-LCNet_x1_0_ML.yaml +41 -0
- paddlex/configs/modules/image_multilabel_classification/ResNet50_ML.yaml +41 -0
- paddlex/configs/modules/image_unwarping/UVDoc.yaml +12 -0
- paddlex/configs/modules/instance_segmentation/Cascade-MaskRCNN-ResNet50-FPN.yaml +40 -0
- paddlex/configs/modules/instance_segmentation/Cascade-MaskRCNN-ResNet50-vd-SSLDv2-FPN.yaml +40 -0
- paddlex/configs/modules/instance_segmentation/Mask-RT-DETR-H.yaml +40 -0
- paddlex/configs/modules/instance_segmentation/Mask-RT-DETR-L.yaml +40 -0
- paddlex/configs/modules/instance_segmentation/Mask-RT-DETR-M.yaml +40 -0
- paddlex/configs/modules/instance_segmentation/Mask-RT-DETR-S.yaml +40 -0
- paddlex/configs/modules/instance_segmentation/Mask-RT-DETR-X.yaml +40 -0
- paddlex/configs/modules/instance_segmentation/MaskRCNN-ResNeXt101-vd-FPN.yaml +39 -0
- paddlex/configs/modules/instance_segmentation/MaskRCNN-ResNet101-FPN.yaml +40 -0
- paddlex/configs/modules/instance_segmentation/MaskRCNN-ResNet101-vd-FPN.yaml +40 -0
- paddlex/configs/modules/instance_segmentation/MaskRCNN-ResNet50-FPN.yaml +40 -0
- paddlex/configs/modules/instance_segmentation/MaskRCNN-ResNet50-vd-FPN.yaml +40 -0
- paddlex/configs/modules/instance_segmentation/MaskRCNN-ResNet50.yaml +40 -0
- paddlex/configs/modules/instance_segmentation/PP-YOLOE_seg-S.yaml +40 -0
- paddlex/configs/modules/instance_segmentation/SOLOv2.yaml +40 -0
- paddlex/configs/modules/keypoint_detection/PP-TinyPose_128x96.yaml +40 -0
- paddlex/configs/modules/keypoint_detection/PP-TinyPose_256x192.yaml +40 -0
- paddlex/configs/modules/layout_detection/PP-DocBlockLayout.yaml +40 -0
- paddlex/configs/modules/layout_detection/PP-DocLayout-L.yaml +40 -0
- paddlex/configs/modules/layout_detection/PP-DocLayout-M.yaml +40 -0
- paddlex/configs/modules/layout_detection/PP-DocLayout-S.yaml +40 -0
- paddlex/configs/modules/layout_detection/PP-DocLayout_plus-L.yaml +40 -0
- paddlex/configs/modules/layout_detection/PicoDet-L_layout_17cls.yaml +40 -0
- paddlex/configs/modules/layout_detection/PicoDet-L_layout_3cls.yaml +40 -0
- paddlex/configs/modules/layout_detection/PicoDet-S_layout_17cls.yaml +40 -0
- paddlex/configs/modules/layout_detection/PicoDet-S_layout_3cls.yaml +40 -0
- paddlex/configs/modules/layout_detection/PicoDet_layout_1x.yaml +40 -0
- paddlex/configs/modules/layout_detection/PicoDet_layout_1x_table.yaml +40 -0
- paddlex/configs/modules/layout_detection/RT-DETR-H_layout_17cls.yaml +40 -0
- paddlex/configs/modules/layout_detection/RT-DETR-H_layout_3cls.yaml +40 -0
- paddlex/configs/modules/mainbody_detection/PP-ShiTuV2_det.yaml +41 -0
- paddlex/configs/modules/multilingual_speech_recognition/whisper_base.yaml +12 -0
- paddlex/configs/modules/multilingual_speech_recognition/whisper_large.yaml +12 -0
- paddlex/configs/modules/multilingual_speech_recognition/whisper_medium.yaml +12 -0
- paddlex/configs/modules/multilingual_speech_recognition/whisper_small.yaml +12 -0
- paddlex/configs/modules/multilingual_speech_recognition/whisper_tiny.yaml +12 -0
- paddlex/configs/modules/object_detection/Cascade-FasterRCNN-ResNet50-FPN.yaml +41 -0
- paddlex/configs/modules/object_detection/Cascade-FasterRCNN-ResNet50-vd-SSLDv2-FPN.yaml +42 -0
- paddlex/configs/modules/object_detection/CenterNet-DLA-34.yaml +41 -0
- paddlex/configs/modules/object_detection/CenterNet-ResNet50.yaml +41 -0
- paddlex/configs/modules/object_detection/Co-DINO-R50.yaml +40 -0
- paddlex/configs/modules/object_detection/Co-DINO-Swin-L.yaml +40 -0
- paddlex/configs/modules/object_detection/Co-Deformable-DETR-R50.yaml +40 -0
- paddlex/configs/modules/object_detection/Co-Deformable-DETR-Swin-T.yaml +40 -0
- paddlex/configs/modules/object_detection/DETR-R50.yaml +42 -0
- paddlex/configs/modules/object_detection/FCOS-ResNet50.yaml +41 -0
- paddlex/configs/modules/object_detection/FasterRCNN-ResNeXt101-vd-FPN.yaml +42 -0
- paddlex/configs/modules/object_detection/FasterRCNN-ResNet101-FPN.yaml +42 -0
- paddlex/configs/modules/object_detection/FasterRCNN-ResNet101.yaml +42 -0
- paddlex/configs/modules/object_detection/FasterRCNN-ResNet34-FPN.yaml +42 -0
- paddlex/configs/modules/object_detection/FasterRCNN-ResNet50-FPN.yaml +42 -0
- paddlex/configs/modules/object_detection/FasterRCNN-ResNet50-vd-FPN.yaml +42 -0
- paddlex/configs/modules/object_detection/FasterRCNN-ResNet50-vd-SSLDv2-FPN.yaml +42 -0
- paddlex/configs/modules/object_detection/FasterRCNN-ResNet50.yaml +42 -0
- paddlex/configs/modules/object_detection/FasterRCNN-Swin-Tiny-FPN.yaml +42 -0
- paddlex/configs/modules/object_detection/PP-YOLOE_plus-L.yaml +40 -0
- paddlex/configs/modules/object_detection/PP-YOLOE_plus-M.yaml +40 -0
- paddlex/configs/modules/object_detection/PP-YOLOE_plus-S.yaml +40 -0
- paddlex/configs/modules/object_detection/PP-YOLOE_plus-X.yaml +40 -0
- paddlex/configs/modules/object_detection/PicoDet-L.yaml +40 -0
- paddlex/configs/modules/object_detection/PicoDet-M.yaml +42 -0
- paddlex/configs/modules/object_detection/PicoDet-S.yaml +40 -0
- paddlex/configs/modules/object_detection/PicoDet-XS.yaml +42 -0
- paddlex/configs/modules/object_detection/RT-DETR-H.yaml +40 -0
- paddlex/configs/modules/object_detection/RT-DETR-L.yaml +40 -0
- paddlex/configs/modules/object_detection/RT-DETR-R18.yaml +40 -0
- paddlex/configs/modules/object_detection/RT-DETR-R50.yaml +40 -0
- paddlex/configs/modules/object_detection/RT-DETR-X.yaml +40 -0
- paddlex/configs/modules/object_detection/YOLOX-L.yaml +40 -0
- paddlex/configs/modules/object_detection/YOLOX-M.yaml +40 -0
- paddlex/configs/modules/object_detection/YOLOX-N.yaml +40 -0
- paddlex/configs/modules/object_detection/YOLOX-S.yaml +40 -0
- paddlex/configs/modules/object_detection/YOLOX-T.yaml +40 -0
- paddlex/configs/modules/object_detection/YOLOX-X.yaml +40 -0
- paddlex/configs/modules/object_detection/YOLOv3-DarkNet53.yaml +40 -0
- paddlex/configs/modules/object_detection/YOLOv3-MobileNetV3.yaml +40 -0
- paddlex/configs/modules/object_detection/YOLOv3-ResNet50_vd_DCN.yaml +40 -0
- paddlex/configs/modules/open_vocabulary_detection/GroundingDINO-T.yaml +13 -0
- paddlex/configs/modules/open_vocabulary_detection/YOLO-Worldv2-L.yaml +13 -0
- paddlex/configs/modules/open_vocabulary_segmentation/SAM-H_box.yaml +17 -0
- paddlex/configs/modules/open_vocabulary_segmentation/SAM-H_point.yaml +15 -0
- paddlex/configs/modules/pedestrian_attribute_recognition/PP-LCNet_x1_0_pedestrian_attribute.yaml +41 -0
- paddlex/configs/modules/rotated_object_detection/PP-YOLOE-R-L.yaml +40 -0
- paddlex/configs/modules/seal_text_detection/PP-OCRv4_mobile_seal_det.yaml +40 -0
- paddlex/configs/modules/seal_text_detection/PP-OCRv4_server_seal_det.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/Deeplabv3-R101.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/Deeplabv3-R50.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/Deeplabv3_Plus-R101.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/Deeplabv3_Plus-R50.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/MaskFormer_small.yaml +42 -0
- paddlex/configs/modules/semantic_segmentation/MaskFormer_tiny.yaml +42 -0
- paddlex/configs/modules/semantic_segmentation/OCRNet_HRNet-W18.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/OCRNet_HRNet-W48.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/PP-LiteSeg-B.yaml +41 -0
- paddlex/configs/modules/semantic_segmentation/PP-LiteSeg-T.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/SeaFormer_base.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/SeaFormer_large.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/SeaFormer_small.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/SeaFormer_tiny.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/SegFormer-B0.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/SegFormer-B1.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/SegFormer-B2.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/SegFormer-B3.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/SegFormer-B4.yaml +40 -0
- paddlex/configs/modules/semantic_segmentation/SegFormer-B5.yaml +40 -0
- paddlex/configs/modules/small_object_detection/PP-YOLOE_plus_SOD-L.yaml +42 -0
- paddlex/configs/modules/small_object_detection/PP-YOLOE_plus_SOD-S.yaml +42 -0
- paddlex/configs/modules/small_object_detection/PP-YOLOE_plus_SOD-largesize-L.yaml +42 -0
- paddlex/configs/modules/table_cells_detection/RT-DETR-L_wired_table_cell_det.yaml +40 -0
- paddlex/configs/modules/table_cells_detection/RT-DETR-L_wireless_table_cell_det.yaml +40 -0
- paddlex/configs/modules/table_classification/PP-LCNet_x1_0_table_cls.yaml +41 -0
- paddlex/configs/modules/table_structure_recognition/SLANeXt_wired.yaml +39 -0
- paddlex/configs/modules/table_structure_recognition/SLANeXt_wireless.yaml +39 -0
- paddlex/configs/modules/table_structure_recognition/SLANet.yaml +39 -0
- paddlex/configs/modules/table_structure_recognition/SLANet_plus.yaml +39 -0
- paddlex/configs/modules/text_detection/PP-OCRv3_mobile_det.yaml +40 -0
- paddlex/configs/modules/text_detection/PP-OCRv3_server_det.yaml +40 -0
- paddlex/configs/modules/text_detection/PP-OCRv4_mobile_det.yaml +40 -0
- paddlex/configs/modules/text_detection/PP-OCRv4_server_det.yaml +40 -0
- paddlex/configs/modules/text_detection/PP-OCRv5_mobile_det.yaml +40 -0
- paddlex/configs/modules/text_detection/PP-OCRv5_server_det.yaml +40 -0
- paddlex/configs/modules/text_recognition/PP-OCRv3_mobile_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/PP-OCRv4_mobile_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/PP-OCRv4_server_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/PP-OCRv4_server_rec_doc.yaml +39 -0
- paddlex/configs/modules/text_recognition/PP-OCRv5_mobile_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/PP-OCRv5_server_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/arabic_PP-OCRv3_mobile_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/ch_RepSVTR_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/ch_SVTRv2_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/chinese_cht_PP-OCRv3_mobile_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/cyrillic_PP-OCRv3_mobile_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/devanagari_PP-OCRv3_mobile_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/en_PP-OCRv3_mobile_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/en_PP-OCRv4_mobile_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/japan_PP-OCRv3_mobile_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/ka_PP-OCRv3_mobile_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/korean_PP-OCRv3_mobile_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/latin_PP-OCRv3_mobile_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/ta_PP-OCRv3_mobile_rec.yaml +39 -0
- paddlex/configs/modules/text_recognition/te_PP-OCRv3_mobile_rec.yaml +39 -0
- paddlex/configs/modules/textline_orientation/PP-LCNet_x0_25_textline_ori.yaml +41 -0
- paddlex/configs/modules/ts_anomaly_detection/AutoEncoder_ad.yaml +37 -0
- paddlex/configs/modules/ts_anomaly_detection/DLinear_ad.yaml +37 -0
- paddlex/configs/modules/ts_anomaly_detection/Nonstationary_ad.yaml +37 -0
- paddlex/configs/modules/ts_anomaly_detection/PatchTST_ad.yaml +37 -0
- paddlex/configs/modules/ts_anomaly_detection/TimesNet_ad.yaml +37 -0
- paddlex/configs/modules/ts_classification/TimesNet_cls.yaml +37 -0
- paddlex/configs/modules/ts_forecast/DLinear.yaml +38 -0
- paddlex/configs/modules/ts_forecast/NLinear.yaml +38 -0
- paddlex/configs/modules/ts_forecast/Nonstationary.yaml +38 -0
- paddlex/configs/modules/ts_forecast/PatchTST.yaml +38 -0
- paddlex/configs/modules/ts_forecast/RLinear.yaml +38 -0
- paddlex/configs/modules/ts_forecast/TiDE.yaml +38 -0
- paddlex/configs/modules/ts_forecast/TimesNet.yaml +38 -0
- paddlex/configs/modules/vehicle_attribute_recognition/PP-LCNet_x1_0_vehicle_attribute.yaml +41 -0
- paddlex/configs/modules/vehicle_detection/PP-YOLOE-L_vehicle.yaml +41 -0
- paddlex/configs/modules/vehicle_detection/PP-YOLOE-S_vehicle.yaml +42 -0
- paddlex/configs/modules/video_classification/PP-TSM-R50_8frames_uniform.yaml +42 -0
- paddlex/configs/modules/video_classification/PP-TSMv2-LCNetV2_16frames_uniform.yaml +42 -0
- paddlex/configs/modules/video_classification/PP-TSMv2-LCNetV2_8frames_uniform.yaml +42 -0
- paddlex/configs/modules/video_detection/YOWO.yaml +40 -0
- paddlex/configs/pipelines/3d_bev_detection.yaml +9 -0
- paddlex/configs/pipelines/OCR.yaml +45 -0
- paddlex/configs/pipelines/PP-ChatOCRv3-doc.yaml +151 -0
- paddlex/configs/pipelines/PP-ChatOCRv4-doc.yaml +237 -0
- paddlex/configs/pipelines/PP-ShiTuV2.yaml +18 -0
- paddlex/configs/pipelines/PP-StructureV3.yaml +226 -0
- paddlex/configs/pipelines/anomaly_detection.yaml +8 -0
- paddlex/configs/pipelines/doc_preprocessor.yaml +15 -0
- paddlex/configs/pipelines/doc_understanding.yaml +9 -0
- paddlex/configs/pipelines/face_recognition.yaml +18 -0
- paddlex/configs/pipelines/formula_recognition.yaml +39 -0
- paddlex/configs/pipelines/human_keypoint_detection.yaml +17 -0
- paddlex/configs/pipelines/image_classification.yaml +10 -0
- paddlex/configs/pipelines/image_multilabel_classification.yaml +9 -0
- paddlex/configs/pipelines/instance_segmentation.yaml +10 -0
- paddlex/configs/pipelines/layout_parsing.yaml +102 -0
- paddlex/configs/pipelines/multilingual_speech_recognition.yaml +9 -0
- paddlex/configs/pipelines/object_detection.yaml +10 -0
- paddlex/configs/pipelines/open_vocabulary_detection.yaml +12 -0
- paddlex/configs/pipelines/open_vocabulary_segmentation.yaml +13 -0
- paddlex/configs/pipelines/pedestrian_attribute_recognition.yaml +15 -0
- paddlex/configs/pipelines/rotated_object_detection.yaml +10 -0
- paddlex/configs/pipelines/seal_recognition.yaml +52 -0
- paddlex/configs/pipelines/semantic_segmentation.yaml +10 -0
- paddlex/configs/pipelines/small_object_detection.yaml +10 -0
- paddlex/configs/pipelines/table_recognition.yaml +57 -0
- paddlex/configs/pipelines/table_recognition_v2.yaml +82 -0
- paddlex/configs/pipelines/ts_anomaly_detection.yaml +8 -0
- paddlex/configs/pipelines/ts_classification.yaml +8 -0
- paddlex/configs/pipelines/ts_forecast.yaml +8 -0
- paddlex/configs/pipelines/vehicle_attribute_recognition.yaml +15 -0
- paddlex/configs/pipelines/video_classification.yaml +9 -0
- paddlex/configs/pipelines/video_detection.yaml +10 -0
- paddlex/constants.py +17 -0
- paddlex/engine.py +56 -0
- paddlex/hpip_links.html +31 -0
- paddlex/inference/__init__.py +19 -0
- paddlex/inference/common/__init__.py +13 -0
- paddlex/inference/common/batch_sampler/__init__.py +21 -0
- paddlex/inference/common/batch_sampler/audio_batch_sampler.py +83 -0
- paddlex/inference/common/batch_sampler/base_batch_sampler.py +94 -0
- paddlex/inference/common/batch_sampler/det_3d_batch_sampler.py +144 -0
- paddlex/inference/common/batch_sampler/doc_vlm_batch_sampler.py +87 -0
- paddlex/inference/common/batch_sampler/image_batch_sampler.py +121 -0
- paddlex/inference/common/batch_sampler/ts_batch_sampler.py +109 -0
- paddlex/inference/common/batch_sampler/video_batch_sampler.py +74 -0
- paddlex/inference/common/reader/__init__.py +19 -0
- paddlex/inference/common/reader/audio_reader.py +46 -0
- paddlex/inference/common/reader/det_3d_reader.py +241 -0
- paddlex/inference/common/reader/image_reader.py +73 -0
- paddlex/inference/common/reader/ts_reader.py +46 -0
- paddlex/inference/common/reader/video_reader.py +42 -0
- paddlex/inference/common/result/__init__.py +29 -0
- paddlex/inference/common/result/base_cv_result.py +41 -0
- paddlex/inference/common/result/base_result.py +72 -0
- paddlex/inference/common/result/base_ts_result.py +41 -0
- paddlex/inference/common/result/base_video_result.py +36 -0
- paddlex/inference/common/result/mixin.py +709 -0
- paddlex/inference/models/__init__.py +86 -0
- paddlex/inference/models/anomaly_detection/__init__.py +15 -0
- paddlex/inference/models/anomaly_detection/predictor.py +135 -0
- paddlex/inference/models/anomaly_detection/processors.py +53 -0
- paddlex/inference/models/anomaly_detection/result.py +71 -0
- paddlex/inference/models/base/__init__.py +15 -0
- paddlex/inference/models/base/predictor/__init__.py +15 -0
- paddlex/inference/models/base/predictor/base_predictor.py +414 -0
- paddlex/inference/models/common/__init__.py +26 -0
- paddlex/inference/models/common/static_infer.py +801 -0
- paddlex/inference/models/common/tokenizer/__init__.py +21 -0
- paddlex/inference/models/common/tokenizer/bert_tokenizer.py +655 -0
- paddlex/inference/models/common/tokenizer/clip_tokenizer.py +609 -0
- paddlex/inference/models/common/tokenizer/gpt_tokenizer.py +453 -0
- paddlex/inference/models/common/tokenizer/qwen2_5_tokenizer.py +112 -0
- paddlex/inference/models/common/tokenizer/qwen2_tokenizer.py +438 -0
- paddlex/inference/models/common/tokenizer/qwen_tokenizer.py +288 -0
- paddlex/inference/models/common/tokenizer/tokenizer_utils.py +2149 -0
- paddlex/inference/models/common/tokenizer/tokenizer_utils_base.py +3720 -0
- paddlex/inference/models/common/tokenizer/utils.py +66 -0
- paddlex/inference/models/common/tokenizer/vocab.py +647 -0
- paddlex/inference/models/common/ts/__init__.py +15 -0
- paddlex/inference/models/common/ts/funcs.py +540 -0
- paddlex/inference/models/common/ts/processors.py +322 -0
- paddlex/inference/models/common/vision/__init__.py +23 -0
- paddlex/inference/models/common/vision/funcs.py +98 -0
- paddlex/inference/models/common/vision/processors.py +285 -0
- paddlex/inference/models/common/vlm/__init__.py +13 -0
- paddlex/inference/models/common/vlm/activations.py +189 -0
- paddlex/inference/models/common/vlm/bert_padding.py +127 -0
- paddlex/inference/models/common/vlm/conversion_utils.py +99 -0
- paddlex/inference/models/common/vlm/distributed.py +229 -0
- paddlex/inference/models/common/vlm/flash_attn_utils.py +119 -0
- paddlex/inference/models/common/vlm/fusion_ops.py +205 -0
- paddlex/inference/models/common/vlm/generation/__init__.py +34 -0
- paddlex/inference/models/common/vlm/generation/configuration_utils.py +533 -0
- paddlex/inference/models/common/vlm/generation/logits_process.py +730 -0
- paddlex/inference/models/common/vlm/generation/stopping_criteria.py +106 -0
- paddlex/inference/models/common/vlm/generation/utils.py +2162 -0
- paddlex/inference/models/common/vlm/transformers/__init__.py +16 -0
- paddlex/inference/models/common/vlm/transformers/configuration_utils.py +1037 -0
- paddlex/inference/models/common/vlm/transformers/conversion_utils.py +408 -0
- paddlex/inference/models/common/vlm/transformers/model_outputs.py +1612 -0
- paddlex/inference/models/common/vlm/transformers/model_utils.py +2014 -0
- paddlex/inference/models/common/vlm/transformers/utils.py +178 -0
- paddlex/inference/models/common/vlm/utils.py +109 -0
- paddlex/inference/models/doc_vlm/__init__.py +15 -0
- paddlex/inference/models/doc_vlm/modeling/GOT_ocr_2_0.py +830 -0
- paddlex/inference/models/doc_vlm/modeling/__init__.py +17 -0
- paddlex/inference/models/doc_vlm/modeling/qwen2.py +1606 -0
- paddlex/inference/models/doc_vlm/modeling/qwen2_5_vl.py +3006 -0
- paddlex/inference/models/doc_vlm/modeling/qwen2_vl.py +2495 -0
- paddlex/inference/models/doc_vlm/predictor.py +253 -0
- paddlex/inference/models/doc_vlm/processors/GOT_ocr_2_0.py +97 -0
- paddlex/inference/models/doc_vlm/processors/__init__.py +17 -0
- paddlex/inference/models/doc_vlm/processors/common.py +561 -0
- paddlex/inference/models/doc_vlm/processors/qwen2_5_vl.py +548 -0
- paddlex/inference/models/doc_vlm/processors/qwen2_vl.py +543 -0
- paddlex/inference/models/doc_vlm/result.py +21 -0
- paddlex/inference/models/face_feature/__init__.py +15 -0
- paddlex/inference/models/face_feature/predictor.py +66 -0
- paddlex/inference/models/formula_recognition/__init__.py +15 -0
- paddlex/inference/models/formula_recognition/predictor.py +193 -0
- paddlex/inference/models/formula_recognition/processors.py +1015 -0
- paddlex/inference/models/formula_recognition/result.py +411 -0
- paddlex/inference/models/image_classification/__init__.py +15 -0
- paddlex/inference/models/image_classification/predictor.py +172 -0
- paddlex/inference/models/image_classification/processors.py +89 -0
- paddlex/inference/models/image_classification/result.py +93 -0
- paddlex/inference/models/image_feature/__init__.py +15 -0
- paddlex/inference/models/image_feature/predictor.py +146 -0
- paddlex/inference/models/image_feature/processors.py +31 -0
- paddlex/inference/models/image_feature/result.py +32 -0
- paddlex/inference/models/image_multilabel_classification/__init__.py +15 -0
- paddlex/inference/models/image_multilabel_classification/predictor.py +95 -0
- paddlex/inference/models/image_multilabel_classification/processors.py +89 -0
- paddlex/inference/models/image_multilabel_classification/result.py +96 -0
- paddlex/inference/models/image_unwarping/__init__.py +15 -0
- paddlex/inference/models/image_unwarping/predictor.py +97 -0
- paddlex/inference/models/image_unwarping/processors.py +92 -0
- paddlex/inference/models/image_unwarping/result.py +47 -0
- paddlex/inference/models/instance_segmentation/__init__.py +15 -0
- paddlex/inference/models/instance_segmentation/predictor.py +202 -0
- paddlex/inference/models/instance_segmentation/processors.py +102 -0
- paddlex/inference/models/instance_segmentation/result.py +162 -0
- paddlex/inference/models/keypoint_detection/__init__.py +15 -0
- paddlex/inference/models/keypoint_detection/predictor.py +190 -0
- paddlex/inference/models/keypoint_detection/processors.py +367 -0
- paddlex/inference/models/keypoint_detection/result.py +197 -0
- paddlex/inference/models/m_3d_bev_detection/__init__.py +15 -0
- paddlex/inference/models/m_3d_bev_detection/predictor.py +303 -0
- paddlex/inference/models/m_3d_bev_detection/processors.py +990 -0
- paddlex/inference/models/m_3d_bev_detection/result.py +68 -0
- paddlex/inference/models/m_3d_bev_detection/visualizer_3d.py +169 -0
- paddlex/inference/models/multilingual_speech_recognition/__init__.py +15 -0
- paddlex/inference/models/multilingual_speech_recognition/predictor.py +137 -0
- paddlex/inference/models/multilingual_speech_recognition/processors.py +1933 -0
- paddlex/inference/models/multilingual_speech_recognition/result.py +21 -0
- paddlex/inference/models/object_detection/__init__.py +15 -0
- paddlex/inference/models/object_detection/predictor.py +344 -0
- paddlex/inference/models/object_detection/processors.py +885 -0
- paddlex/inference/models/object_detection/result.py +114 -0
- paddlex/inference/models/object_detection/utils.py +70 -0
- paddlex/inference/models/open_vocabulary_detection/__init__.py +15 -0
- paddlex/inference/models/open_vocabulary_detection/predictor.py +172 -0
- paddlex/inference/models/open_vocabulary_detection/processors/__init__.py +16 -0
- paddlex/inference/models/open_vocabulary_detection/processors/common.py +114 -0
- paddlex/inference/models/open_vocabulary_detection/processors/groundingdino_processors.py +496 -0
- paddlex/inference/models/open_vocabulary_detection/processors/yoloworld_processors.py +209 -0
- paddlex/inference/models/open_vocabulary_segmentation/__init__.py +15 -0
- paddlex/inference/models/open_vocabulary_segmentation/predictor.py +113 -0
- paddlex/inference/models/open_vocabulary_segmentation/processors/__init__.py +15 -0
- paddlex/inference/models/open_vocabulary_segmentation/processors/sam_processer.py +249 -0
- paddlex/inference/models/open_vocabulary_segmentation/results/__init__.py +15 -0
- paddlex/inference/models/open_vocabulary_segmentation/results/sam_result.py +149 -0
- paddlex/inference/models/semantic_segmentation/__init__.py +15 -0
- paddlex/inference/models/semantic_segmentation/predictor.py +158 -0
- paddlex/inference/models/semantic_segmentation/processors.py +117 -0
- paddlex/inference/models/semantic_segmentation/result.py +73 -0
- paddlex/inference/models/table_structure_recognition/__init__.py +15 -0
- paddlex/inference/models/table_structure_recognition/predictor.py +161 -0
- paddlex/inference/models/table_structure_recognition/processors.py +229 -0
- paddlex/inference/models/table_structure_recognition/result.py +63 -0
- paddlex/inference/models/text_detection/__init__.py +15 -0
- paddlex/inference/models/text_detection/predictor.py +191 -0
- paddlex/inference/models/text_detection/processors.py +538 -0
- paddlex/inference/models/text_detection/result.py +46 -0
- paddlex/inference/models/text_recognition/__init__.py +15 -0
- paddlex/inference/models/text_recognition/predictor.py +98 -0
- paddlex/inference/models/text_recognition/processors.py +245 -0
- paddlex/inference/models/text_recognition/result.py +76 -0
- paddlex/inference/models/ts_anomaly_detection/__init__.py +15 -0
- paddlex/inference/models/ts_anomaly_detection/predictor.py +141 -0
- paddlex/inference/models/ts_anomaly_detection/processors.py +98 -0
- paddlex/inference/models/ts_anomaly_detection/result.py +83 -0
- paddlex/inference/models/ts_classification/__init__.py +15 -0
- paddlex/inference/models/ts_classification/predictor.py +122 -0
- paddlex/inference/models/ts_classification/processors.py +122 -0
- paddlex/inference/models/ts_classification/result.py +87 -0
- paddlex/inference/models/ts_forecasting/__init__.py +15 -0
- paddlex/inference/models/ts_forecasting/predictor.py +154 -0
- paddlex/inference/models/ts_forecasting/processors.py +158 -0
- paddlex/inference/models/ts_forecasting/result.py +96 -0
- paddlex/inference/models/video_classification/__init__.py +15 -0
- paddlex/inference/models/video_classification/predictor.py +141 -0
- paddlex/inference/models/video_classification/processors.py +409 -0
- paddlex/inference/models/video_classification/result.py +96 -0
- paddlex/inference/models/video_detection/__init__.py +15 -0
- paddlex/inference/models/video_detection/predictor.py +129 -0
- paddlex/inference/models/video_detection/processors.py +463 -0
- paddlex/inference/models/video_detection/result.py +109 -0
- paddlex/inference/pipelines/__init__.py +239 -0
- paddlex/inference/pipelines/_parallel.py +172 -0
- paddlex/inference/pipelines/anomaly_detection/__init__.py +15 -0
- paddlex/inference/pipelines/anomaly_detection/pipeline.py +82 -0
- paddlex/inference/pipelines/attribute_recognition/__init__.py +15 -0
- paddlex/inference/pipelines/attribute_recognition/pipeline.py +120 -0
- paddlex/inference/pipelines/attribute_recognition/result.py +102 -0
- paddlex/inference/pipelines/base.py +156 -0
- paddlex/inference/pipelines/components/__init__.py +29 -0
- paddlex/inference/pipelines/components/chat_server/__init__.py +16 -0
- paddlex/inference/pipelines/components/chat_server/base.py +39 -0
- paddlex/inference/pipelines/components/chat_server/openai_bot_chat.py +236 -0
- paddlex/inference/pipelines/components/common/__init__.py +19 -0
- paddlex/inference/pipelines/components/common/base_operator.py +37 -0
- paddlex/inference/pipelines/components/common/base_result.py +66 -0
- paddlex/inference/pipelines/components/common/convert_points_and_boxes.py +45 -0
- paddlex/inference/pipelines/components/common/crop_image_regions.py +556 -0
- paddlex/inference/pipelines/components/common/seal_det_warp.py +972 -0
- paddlex/inference/pipelines/components/common/sort_boxes.py +85 -0
- paddlex/inference/pipelines/components/common/warp_image.py +50 -0
- paddlex/inference/pipelines/components/faisser.py +357 -0
- paddlex/inference/pipelines/components/prompt_engineering/__init__.py +16 -0
- paddlex/inference/pipelines/components/prompt_engineering/base.py +35 -0
- paddlex/inference/pipelines/components/prompt_engineering/generate_ensemble_prompt.py +128 -0
- paddlex/inference/pipelines/components/prompt_engineering/generate_kie_prompt.py +148 -0
- paddlex/inference/pipelines/components/retriever/__init__.py +16 -0
- paddlex/inference/pipelines/components/retriever/base.py +228 -0
- paddlex/inference/pipelines/components/retriever/openai_bot_retriever.py +70 -0
- paddlex/inference/pipelines/components/retriever/qianfan_bot_retriever.py +166 -0
- paddlex/inference/pipelines/components/utils/__init__.py +13 -0
- paddlex/inference/pipelines/components/utils/mixin.py +206 -0
- paddlex/inference/pipelines/doc_preprocessor/__init__.py +15 -0
- paddlex/inference/pipelines/doc_preprocessor/pipeline.py +209 -0
- paddlex/inference/pipelines/doc_preprocessor/result.py +98 -0
- paddlex/inference/pipelines/doc_understanding/__init__.py +15 -0
- paddlex/inference/pipelines/doc_understanding/pipeline.py +71 -0
- paddlex/inference/pipelines/face_recognition/__init__.py +15 -0
- paddlex/inference/pipelines/face_recognition/pipeline.py +63 -0
- paddlex/inference/pipelines/face_recognition/result.py +44 -0
- paddlex/inference/pipelines/formula_recognition/__init__.py +15 -0
- paddlex/inference/pipelines/formula_recognition/pipeline.py +347 -0
- paddlex/inference/pipelines/formula_recognition/result.py +282 -0
- paddlex/inference/pipelines/image_classification/__init__.py +15 -0
- paddlex/inference/pipelines/image_classification/pipeline.py +90 -0
- paddlex/inference/pipelines/image_multilabel_classification/__init__.py +15 -0
- paddlex/inference/pipelines/image_multilabel_classification/pipeline.py +97 -0
- paddlex/inference/pipelines/instance_segmentation/__init__.py +15 -0
- paddlex/inference/pipelines/instance_segmentation/pipeline.py +91 -0
- paddlex/inference/pipelines/keypoint_detection/__init__.py +15 -0
- paddlex/inference/pipelines/keypoint_detection/pipeline.py +158 -0
- paddlex/inference/pipelines/layout_parsing/__init__.py +16 -0
- paddlex/inference/pipelines/layout_parsing/pipeline.py +568 -0
- paddlex/inference/pipelines/layout_parsing/pipeline_v2.py +1382 -0
- paddlex/inference/pipelines/layout_parsing/result.py +191 -0
- paddlex/inference/pipelines/layout_parsing/result_v2.py +745 -0
- paddlex/inference/pipelines/layout_parsing/setting.py +87 -0
- paddlex/inference/pipelines/layout_parsing/utils.py +951 -0
- paddlex/inference/pipelines/layout_parsing/xycut_enhanced/__init__.py +16 -0
- paddlex/inference/pipelines/layout_parsing/xycut_enhanced/utils.py +1143 -0
- paddlex/inference/pipelines/layout_parsing/xycut_enhanced/xycuts.py +562 -0
- paddlex/inference/pipelines/m_3d_bev_detection/__init__.py +15 -0
- paddlex/inference/pipelines/m_3d_bev_detection/pipeline.py +74 -0
- paddlex/inference/pipelines/multilingual_speech_recognition/__init__.py +15 -0
- paddlex/inference/pipelines/multilingual_speech_recognition/pipeline.py +78 -0
- paddlex/inference/pipelines/object_detection/__init__.py +15 -0
- paddlex/inference/pipelines/object_detection/pipeline.py +115 -0
- paddlex/inference/pipelines/ocr/__init__.py +15 -0
- paddlex/inference/pipelines/ocr/pipeline.py +463 -0
- paddlex/inference/pipelines/ocr/result.py +255 -0
- paddlex/inference/pipelines/open_vocabulary_detection/__init__.py +15 -0
- paddlex/inference/pipelines/open_vocabulary_detection/pipeline.py +86 -0
- paddlex/inference/pipelines/open_vocabulary_segmentation/__init__.py +15 -0
- paddlex/inference/pipelines/open_vocabulary_segmentation/pipeline.py +100 -0
- paddlex/inference/pipelines/pp_chatocr/__init__.py +16 -0
- paddlex/inference/pipelines/pp_chatocr/pipeline_base.py +111 -0
- paddlex/inference/pipelines/pp_chatocr/pipeline_v3.py +781 -0
- paddlex/inference/pipelines/pp_chatocr/pipeline_v4.py +992 -0
- paddlex/inference/pipelines/pp_shitu_v2/__init__.py +15 -0
- paddlex/inference/pipelines/pp_shitu_v2/pipeline.py +156 -0
- paddlex/inference/pipelines/pp_shitu_v2/result.py +126 -0
- paddlex/inference/pipelines/rotated_object_detection/__init__.py +15 -0
- paddlex/inference/pipelines/rotated_object_detection/pipeline.py +95 -0
- paddlex/inference/pipelines/seal_recognition/__init__.py +15 -0
- paddlex/inference/pipelines/seal_recognition/pipeline.py +335 -0
- paddlex/inference/pipelines/seal_recognition/result.py +89 -0
- paddlex/inference/pipelines/semantic_segmentation/__init__.py +15 -0
- paddlex/inference/pipelines/semantic_segmentation/pipeline.py +95 -0
- paddlex/inference/pipelines/small_object_detection/__init__.py +15 -0
- paddlex/inference/pipelines/small_object_detection/pipeline.py +95 -0
- paddlex/inference/pipelines/table_recognition/__init__.py +16 -0
- paddlex/inference/pipelines/table_recognition/pipeline.py +486 -0
- paddlex/inference/pipelines/table_recognition/pipeline_v2.py +1395 -0
- paddlex/inference/pipelines/table_recognition/result.py +218 -0
- paddlex/inference/pipelines/table_recognition/table_recognition_post_processing.py +366 -0
- paddlex/inference/pipelines/table_recognition/table_recognition_post_processing_v2.py +488 -0
- paddlex/inference/pipelines/table_recognition/utils.py +44 -0
- paddlex/inference/pipelines/ts_anomaly_detection/__init__.py +15 -0
- paddlex/inference/pipelines/ts_anomaly_detection/pipeline.py +72 -0
- paddlex/inference/pipelines/ts_classification/__init__.py +15 -0
- paddlex/inference/pipelines/ts_classification/pipeline.py +72 -0
- paddlex/inference/pipelines/ts_forecasting/__init__.py +15 -0
- paddlex/inference/pipelines/ts_forecasting/pipeline.py +72 -0
- paddlex/inference/pipelines/video_classification/__init__.py +15 -0
- paddlex/inference/pipelines/video_classification/pipeline.py +79 -0
- paddlex/inference/pipelines/video_detection/__init__.py +15 -0
- paddlex/inference/pipelines/video_detection/pipeline.py +86 -0
- paddlex/inference/serving/__init__.py +17 -0
- paddlex/inference/serving/basic_serving/__init__.py +18 -0
- paddlex/inference/serving/basic_serving/_app.py +221 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/__init__.py +44 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/_common/__init__.py +13 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/_common/common.py +104 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/_common/image_recognition.py +36 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/_common/ocr.py +95 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/anomaly_detection.py +67 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/doc_preprocessor.py +100 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/doc_understanding.py +153 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/face_recognition.py +226 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/formula_recognition.py +100 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/human_keypoint_detection.py +81 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/image_classification.py +69 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/image_multilabel_classification.py +73 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/instance_segmentation.py +87 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/layout_parsing.py +117 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/m_3d_bev_detection.py +79 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/multilingual_speech_recognition.py +92 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/object_detection.py +77 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/ocr.py +102 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/open_vocabulary_detection.py +81 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/open_vocabulary_segmentation.py +91 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/pedestrian_attribute_recognition.py +84 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/pp_chatocrv3_doc.py +193 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/pp_chatocrv4_doc.py +223 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/pp_shituv2.py +221 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/pp_structurev3.py +143 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/rotated_object_detection.py +81 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/seal_recognition.py +106 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/semantic_segmentation.py +67 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/small_object_detection.py +72 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/table_recognition.py +108 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/table_recognition_v2.py +113 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/ts_anomaly_detection.py +65 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/ts_classification.py +64 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/ts_forecast.py +65 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/vehicle_attribute_recognition.py +84 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/video_classification.py +76 -0
- paddlex/inference/serving/basic_serving/_pipeline_apps/video_detection.py +92 -0
- paddlex/inference/serving/basic_serving/_server.py +40 -0
- paddlex/inference/serving/infra/__init__.py +13 -0
- paddlex/inference/serving/infra/config.py +36 -0
- paddlex/inference/serving/infra/models.py +79 -0
- paddlex/inference/serving/infra/storage.py +180 -0
- paddlex/inference/serving/infra/utils.py +285 -0
- paddlex/inference/serving/schemas/__init__.py +13 -0
- paddlex/inference/serving/schemas/anomaly_detection.py +39 -0
- paddlex/inference/serving/schemas/doc_preprocessor.py +54 -0
- paddlex/inference/serving/schemas/doc_understanding.py +78 -0
- paddlex/inference/serving/schemas/face_recognition.py +124 -0
- paddlex/inference/serving/schemas/formula_recognition.py +56 -0
- paddlex/inference/serving/schemas/human_keypoint_detection.py +55 -0
- paddlex/inference/serving/schemas/image_classification.py +45 -0
- paddlex/inference/serving/schemas/image_multilabel_classification.py +47 -0
- paddlex/inference/serving/schemas/instance_segmentation.py +53 -0
- paddlex/inference/serving/schemas/layout_parsing.py +71 -0
- paddlex/inference/serving/schemas/m_3d_bev_detection.py +48 -0
- paddlex/inference/serving/schemas/multilingual_speech_recognition.py +57 -0
- paddlex/inference/serving/schemas/object_detection.py +52 -0
- paddlex/inference/serving/schemas/ocr.py +60 -0
- paddlex/inference/serving/schemas/open_vocabulary_detection.py +52 -0
- paddlex/inference/serving/schemas/open_vocabulary_segmentation.py +52 -0
- paddlex/inference/serving/schemas/pedestrian_attribute_recognition.py +61 -0
- paddlex/inference/serving/schemas/pp_chatocrv3_doc.py +133 -0
- paddlex/inference/serving/schemas/pp_chatocrv4_doc.py +150 -0
- paddlex/inference/serving/schemas/pp_shituv2.py +124 -0
- paddlex/inference/serving/schemas/pp_structurev3.py +88 -0
- paddlex/inference/serving/schemas/rotated_object_detection.py +52 -0
- paddlex/inference/serving/schemas/seal_recognition.py +62 -0
- paddlex/inference/serving/schemas/semantic_segmentation.py +45 -0
- paddlex/inference/serving/schemas/shared/__init__.py +13 -0
- paddlex/inference/serving/schemas/shared/classification.py +23 -0
- paddlex/inference/serving/schemas/shared/image_segmentation.py +28 -0
- paddlex/inference/serving/schemas/shared/object_detection.py +24 -0
- paddlex/inference/serving/schemas/shared/ocr.py +25 -0
- paddlex/inference/serving/schemas/small_object_detection.py +52 -0
- paddlex/inference/serving/schemas/table_recognition.py +64 -0
- paddlex/inference/serving/schemas/table_recognition_v2.py +69 -0
- paddlex/inference/serving/schemas/ts_anomaly_detection.py +37 -0
- paddlex/inference/serving/schemas/ts_classification.py +38 -0
- paddlex/inference/serving/schemas/ts_forecast.py +37 -0
- paddlex/inference/serving/schemas/vehicle_attribute_recognition.py +61 -0
- paddlex/inference/serving/schemas/video_classification.py +44 -0
- paddlex/inference/serving/schemas/video_detection.py +56 -0
- paddlex/inference/utils/__init__.py +13 -0
- paddlex/inference/utils/benchmark.py +379 -0
- paddlex/inference/utils/color_map.py +123 -0
- paddlex/inference/utils/get_pipeline_path.py +27 -0
- paddlex/inference/utils/hpi.py +254 -0
- paddlex/inference/utils/hpi_model_info_collection.json +2331 -0
- paddlex/inference/utils/io/__init__.py +36 -0
- paddlex/inference/utils/io/readers.py +504 -0
- paddlex/inference/utils/io/style.py +381 -0
- paddlex/inference/utils/io/tablepyxl.py +157 -0
- paddlex/inference/utils/io/writers.py +458 -0
- paddlex/inference/utils/model_paths.py +48 -0
- paddlex/inference/utils/new_ir_blocklist.py +27 -0
- paddlex/inference/utils/official_models.py +367 -0
- paddlex/inference/utils/pp_option.py +339 -0
- paddlex/inference/utils/trt_blocklist.py +43 -0
- paddlex/inference/utils/trt_config.py +420 -0
- paddlex/model.py +131 -0
- paddlex/modules/__init__.py +115 -0
- paddlex/modules/anomaly_detection/__init__.py +18 -0
- paddlex/modules/anomaly_detection/dataset_checker/__init__.py +94 -0
- paddlex/modules/anomaly_detection/dataset_checker/dataset_src/__init__.py +19 -0
- paddlex/modules/anomaly_detection/dataset_checker/dataset_src/analyse_dataset.py +82 -0
- paddlex/modules/anomaly_detection/dataset_checker/dataset_src/check_dataset.py +91 -0
- paddlex/modules/anomaly_detection/dataset_checker/dataset_src/convert_dataset.py +233 -0
- paddlex/modules/anomaly_detection/dataset_checker/dataset_src/split_dataset.py +87 -0
- paddlex/modules/anomaly_detection/dataset_checker/dataset_src/utils/__init__.py +13 -0
- paddlex/modules/anomaly_detection/dataset_checker/dataset_src/utils/visualizer.py +76 -0
- paddlex/modules/anomaly_detection/evaluator.py +58 -0
- paddlex/modules/anomaly_detection/exportor.py +22 -0
- paddlex/modules/anomaly_detection/model_list.py +16 -0
- paddlex/modules/anomaly_detection/trainer.py +70 -0
- paddlex/modules/base/__init__.py +18 -0
- paddlex/modules/base/build_model.py +33 -0
- paddlex/modules/base/dataset_checker/__init__.py +16 -0
- paddlex/modules/base/dataset_checker/dataset_checker.py +169 -0
- paddlex/modules/base/dataset_checker/utils.py +108 -0
- paddlex/modules/base/evaluator.py +170 -0
- paddlex/modules/base/exportor.py +145 -0
- paddlex/modules/base/trainer.py +144 -0
- paddlex/modules/base/utils/__init__.py +13 -0
- paddlex/modules/base/utils/cinn_setting.py +89 -0
- paddlex/modules/base/utils/coco_eval.py +94 -0
- paddlex/modules/base/utils/topk_eval.py +118 -0
- paddlex/modules/doc_vlm/__init__.py +18 -0
- paddlex/modules/doc_vlm/dataset_checker.py +29 -0
- paddlex/modules/doc_vlm/evaluator.py +29 -0
- paddlex/modules/doc_vlm/exportor.py +29 -0
- paddlex/modules/doc_vlm/model_list.py +16 -0
- paddlex/modules/doc_vlm/trainer.py +41 -0
- paddlex/modules/face_recognition/__init__.py +18 -0
- paddlex/modules/face_recognition/dataset_checker/__init__.py +71 -0
- paddlex/modules/face_recognition/dataset_checker/dataset_src/__init__.py +16 -0
- paddlex/modules/face_recognition/dataset_checker/dataset_src/check_dataset.py +172 -0
- paddlex/modules/face_recognition/dataset_checker/dataset_src/utils/__init__.py +13 -0
- paddlex/modules/face_recognition/dataset_checker/dataset_src/utils/visualizer.py +153 -0
- paddlex/modules/face_recognition/evaluator.py +52 -0
- paddlex/modules/face_recognition/exportor.py +22 -0
- paddlex/modules/face_recognition/model_list.py +15 -0
- paddlex/modules/face_recognition/trainer.py +75 -0
- paddlex/modules/formula_recognition/__init__.py +18 -0
- paddlex/modules/formula_recognition/dataset_checker/__init__.py +113 -0
- paddlex/modules/formula_recognition/dataset_checker/dataset_src/__init__.py +19 -0
- paddlex/modules/formula_recognition/dataset_checker/dataset_src/analyse_dataset.py +158 -0
- paddlex/modules/formula_recognition/dataset_checker/dataset_src/check_dataset.py +76 -0
- paddlex/modules/formula_recognition/dataset_checker/dataset_src/convert_dataset.py +95 -0
- paddlex/modules/formula_recognition/dataset_checker/dataset_src/split_dataset.py +80 -0
- paddlex/modules/formula_recognition/evaluator.py +80 -0
- paddlex/modules/formula_recognition/exportor.py +22 -0
- paddlex/modules/formula_recognition/model_list.py +23 -0
- paddlex/modules/formula_recognition/trainer.py +123 -0
- paddlex/modules/general_recognition/__init__.py +18 -0
- paddlex/modules/general_recognition/dataset_checker/__init__.py +107 -0
- paddlex/modules/general_recognition/dataset_checker/dataset_src/__init__.py +19 -0
- paddlex/modules/general_recognition/dataset_checker/dataset_src/analyse_dataset.py +96 -0
- paddlex/modules/general_recognition/dataset_checker/dataset_src/check_dataset.py +99 -0
- paddlex/modules/general_recognition/dataset_checker/dataset_src/convert_dataset.py +100 -0
- paddlex/modules/general_recognition/dataset_checker/dataset_src/split_dataset.py +82 -0
- paddlex/modules/general_recognition/dataset_checker/dataset_src/utils/__init__.py +13 -0
- paddlex/modules/general_recognition/dataset_checker/dataset_src/utils/visualizer.py +147 -0
- paddlex/modules/general_recognition/evaluator.py +31 -0
- paddlex/modules/general_recognition/exportor.py +22 -0
- paddlex/modules/general_recognition/model_list.py +19 -0
- paddlex/modules/general_recognition/trainer.py +52 -0
- paddlex/modules/image_classification/__init__.py +18 -0
- paddlex/modules/image_classification/dataset_checker/__init__.py +104 -0
- paddlex/modules/image_classification/dataset_checker/dataset_src/__init__.py +19 -0
- paddlex/modules/image_classification/dataset_checker/dataset_src/analyse_dataset.py +92 -0
- paddlex/modules/image_classification/dataset_checker/dataset_src/check_dataset.py +132 -0
- paddlex/modules/image_classification/dataset_checker/dataset_src/convert_dataset.py +51 -0
- paddlex/modules/image_classification/dataset_checker/dataset_src/split_dataset.py +81 -0
- paddlex/modules/image_classification/dataset_checker/dataset_src/utils/__init__.py +13 -0
- paddlex/modules/image_classification/dataset_checker/dataset_src/utils/visualizer.py +153 -0
- paddlex/modules/image_classification/evaluator.py +43 -0
- paddlex/modules/image_classification/exportor.py +22 -0
- paddlex/modules/image_classification/model_list.py +99 -0
- paddlex/modules/image_classification/trainer.py +82 -0
- paddlex/modules/image_unwarping/__init__.py +13 -0
- paddlex/modules/image_unwarping/model_list.py +17 -0
- paddlex/modules/instance_segmentation/__init__.py +18 -0
- paddlex/modules/instance_segmentation/dataset_checker/__init__.py +107 -0
- paddlex/modules/instance_segmentation/dataset_checker/dataset_src/__init__.py +19 -0
- paddlex/modules/instance_segmentation/dataset_checker/dataset_src/analyse_dataset.py +82 -0
- paddlex/modules/instance_segmentation/dataset_checker/dataset_src/check_dataset.py +95 -0
- paddlex/modules/instance_segmentation/dataset_checker/dataset_src/convert_dataset.py +241 -0
- paddlex/modules/instance_segmentation/dataset_checker/dataset_src/split_dataset.py +122 -0
- paddlex/modules/instance_segmentation/dataset_checker/dataset_src/utils/__init__.py +13 -0
- paddlex/modules/instance_segmentation/dataset_checker/dataset_src/utils/visualizer.py +223 -0
- paddlex/modules/instance_segmentation/evaluator.py +32 -0
- paddlex/modules/instance_segmentation/exportor.py +22 -0
- paddlex/modules/instance_segmentation/model_list.py +33 -0
- paddlex/modules/instance_segmentation/trainer.py +31 -0
- paddlex/modules/keypoint_detection/__init__.py +18 -0
- paddlex/modules/keypoint_detection/dataset_checker/__init__.py +56 -0
- paddlex/modules/keypoint_detection/dataset_checker/dataset_src/__init__.py +15 -0
- paddlex/modules/keypoint_detection/dataset_checker/dataset_src/check_dataset.py +91 -0
- paddlex/modules/keypoint_detection/dataset_checker/dataset_src/utils/__init__.py +13 -0
- paddlex/modules/keypoint_detection/dataset_checker/dataset_src/utils/visualizer.py +124 -0
- paddlex/modules/keypoint_detection/evaluator.py +41 -0
- paddlex/modules/keypoint_detection/exportor.py +22 -0
- paddlex/modules/keypoint_detection/model_list.py +16 -0
- paddlex/modules/keypoint_detection/trainer.py +39 -0
- paddlex/modules/m_3d_bev_detection/__init__.py +18 -0
- paddlex/modules/m_3d_bev_detection/dataset_checker/__init__.py +95 -0
- paddlex/modules/m_3d_bev_detection/dataset_checker/dataset_src/__init__.py +17 -0
- paddlex/modules/m_3d_bev_detection/dataset_checker/dataset_src/analyse_dataset.py +106 -0
- paddlex/modules/m_3d_bev_detection/dataset_checker/dataset_src/check_dataset.py +101 -0
- paddlex/modules/m_3d_bev_detection/evaluator.py +46 -0
- paddlex/modules/m_3d_bev_detection/exportor.py +22 -0
- paddlex/modules/m_3d_bev_detection/model_list.py +18 -0
- paddlex/modules/m_3d_bev_detection/trainer.py +68 -0
- paddlex/modules/multilabel_classification/__init__.py +18 -0
- paddlex/modules/multilabel_classification/dataset_checker/__init__.py +106 -0
- paddlex/modules/multilabel_classification/dataset_checker/dataset_src/__init__.py +19 -0
- paddlex/modules/multilabel_classification/dataset_checker/dataset_src/analyse_dataset.py +94 -0
- paddlex/modules/multilabel_classification/dataset_checker/dataset_src/check_dataset.py +132 -0
- paddlex/modules/multilabel_classification/dataset_checker/dataset_src/convert_dataset.py +120 -0
- paddlex/modules/multilabel_classification/dataset_checker/dataset_src/split_dataset.py +81 -0
- paddlex/modules/multilabel_classification/dataset_checker/dataset_src/utils/__init__.py +13 -0
- paddlex/modules/multilabel_classification/dataset_checker/dataset_src/utils/visualizer.py +149 -0
- paddlex/modules/multilabel_classification/evaluator.py +43 -0
- paddlex/modules/multilabel_classification/exportor.py +22 -0
- paddlex/modules/multilabel_classification/model_list.py +24 -0
- paddlex/modules/multilabel_classification/trainer.py +85 -0
- paddlex/modules/multilingual_speech_recognition/__init__.py +18 -0
- paddlex/modules/multilingual_speech_recognition/dataset_checker.py +27 -0
- paddlex/modules/multilingual_speech_recognition/evaluator.py +27 -0
- paddlex/modules/multilingual_speech_recognition/exportor.py +27 -0
- paddlex/modules/multilingual_speech_recognition/model_list.py +22 -0
- paddlex/modules/multilingual_speech_recognition/trainer.py +42 -0
- paddlex/modules/object_detection/__init__.py +18 -0
- paddlex/modules/object_detection/dataset_checker/__init__.py +106 -0
- paddlex/modules/object_detection/dataset_checker/dataset_src/__init__.py +19 -0
- paddlex/modules/object_detection/dataset_checker/dataset_src/analyse_dataset.py +82 -0
- paddlex/modules/object_detection/dataset_checker/dataset_src/check_dataset.py +91 -0
- paddlex/modules/object_detection/dataset_checker/dataset_src/convert_dataset.py +438 -0
- paddlex/modules/object_detection/dataset_checker/dataset_src/split_dataset.py +123 -0
- paddlex/modules/object_detection/dataset_checker/dataset_src/utils/__init__.py +13 -0
- paddlex/modules/object_detection/dataset_checker/dataset_src/utils/visualizer.py +193 -0
- paddlex/modules/object_detection/evaluator.py +57 -0
- paddlex/modules/object_detection/exportor.py +22 -0
- paddlex/modules/object_detection/model_list.py +86 -0
- paddlex/modules/object_detection/trainer.py +98 -0
- paddlex/modules/open_vocabulary_detection/__init__.py +18 -0
- paddlex/modules/open_vocabulary_detection/dataset_checker.py +29 -0
- paddlex/modules/open_vocabulary_detection/evaluator.py +29 -0
- paddlex/modules/open_vocabulary_detection/exportor.py +29 -0
- paddlex/modules/open_vocabulary_detection/model_list.py +16 -0
- paddlex/modules/open_vocabulary_detection/trainer.py +44 -0
- paddlex/modules/open_vocabulary_segmentation/__init__.py +18 -0
- paddlex/modules/open_vocabulary_segmentation/dataset_checker.py +29 -0
- paddlex/modules/open_vocabulary_segmentation/evaluator.py +29 -0
- paddlex/modules/open_vocabulary_segmentation/exportor.py +29 -0
- paddlex/modules/open_vocabulary_segmentation/model_list.py +19 -0
- paddlex/modules/open_vocabulary_segmentation/trainer.py +44 -0
- paddlex/modules/semantic_segmentation/__init__.py +18 -0
- paddlex/modules/semantic_segmentation/dataset_checker/__init__.py +109 -0
- paddlex/modules/semantic_segmentation/dataset_checker/dataset_src/__init__.py +19 -0
- paddlex/modules/semantic_segmentation/dataset_checker/dataset_src/analyse_dataset.py +76 -0
- paddlex/modules/semantic_segmentation/dataset_checker/dataset_src/check_dataset.py +80 -0
- paddlex/modules/semantic_segmentation/dataset_checker/dataset_src/convert_dataset.py +165 -0
- paddlex/modules/semantic_segmentation/dataset_checker/dataset_src/split_dataset.py +87 -0
- paddlex/modules/semantic_segmentation/dataset_checker/dataset_src/utils/__init__.py +13 -0
- paddlex/modules/semantic_segmentation/dataset_checker/dataset_src/utils/visualizer.py +75 -0
- paddlex/modules/semantic_segmentation/evaluator.py +58 -0
- paddlex/modules/semantic_segmentation/exportor.py +31 -0
- paddlex/modules/semantic_segmentation/model_list.py +37 -0
- paddlex/modules/semantic_segmentation/trainer.py +72 -0
- paddlex/modules/table_recognition/__init__.py +18 -0
- paddlex/modules/table_recognition/dataset_checker/__init__.py +98 -0
- paddlex/modules/table_recognition/dataset_checker/dataset_src/__init__.py +18 -0
- paddlex/modules/table_recognition/dataset_checker/dataset_src/analyse_dataset.py +59 -0
- paddlex/modules/table_recognition/dataset_checker/dataset_src/check_dataset.py +87 -0
- paddlex/modules/table_recognition/dataset_checker/dataset_src/split_dataset.py +80 -0
- paddlex/modules/table_recognition/evaluator.py +43 -0
- paddlex/modules/table_recognition/exportor.py +22 -0
- paddlex/modules/table_recognition/model_list.py +21 -0
- paddlex/modules/table_recognition/trainer.py +67 -0
- paddlex/modules/text_detection/__init__.py +18 -0
- paddlex/modules/text_detection/dataset_checker/__init__.py +107 -0
- paddlex/modules/text_detection/dataset_checker/dataset_src/__init__.py +18 -0
- paddlex/modules/text_detection/dataset_checker/dataset_src/analyse_dataset.py +220 -0
- paddlex/modules/text_detection/dataset_checker/dataset_src/check_dataset.py +106 -0
- paddlex/modules/text_detection/dataset_checker/dataset_src/split_dataset.py +140 -0
- paddlex/modules/text_detection/evaluator.py +41 -0
- paddlex/modules/text_detection/exportor.py +22 -0
- paddlex/modules/text_detection/model_list.py +26 -0
- paddlex/modules/text_detection/trainer.py +65 -0
- paddlex/modules/text_recognition/__init__.py +18 -0
- paddlex/modules/text_recognition/dataset_checker/__init__.py +125 -0
- paddlex/modules/text_recognition/dataset_checker/dataset_src/__init__.py +19 -0
- paddlex/modules/text_recognition/dataset_checker/dataset_src/analyse_dataset.py +162 -0
- paddlex/modules/text_recognition/dataset_checker/dataset_src/check_dataset.py +104 -0
- paddlex/modules/text_recognition/dataset_checker/dataset_src/convert_dataset.py +95 -0
- paddlex/modules/text_recognition/dataset_checker/dataset_src/split_dataset.py +80 -0
- paddlex/modules/text_recognition/evaluator.py +64 -0
- paddlex/modules/text_recognition/exportor.py +22 -0
- paddlex/modules/text_recognition/model_list.py +36 -0
- paddlex/modules/text_recognition/trainer.py +105 -0
- paddlex/modules/ts_anomaly_detection/__init__.py +19 -0
- paddlex/modules/ts_anomaly_detection/dataset_checker/__init__.py +111 -0
- paddlex/modules/ts_anomaly_detection/dataset_checker/dataset_src/__init__.py +19 -0
- paddlex/modules/ts_anomaly_detection/dataset_checker/dataset_src/analyse_dataset.py +19 -0
- paddlex/modules/ts_anomaly_detection/dataset_checker/dataset_src/check_dataset.py +64 -0
- paddlex/modules/ts_anomaly_detection/dataset_checker/dataset_src/convert_dataset.py +74 -0
- paddlex/modules/ts_anomaly_detection/dataset_checker/dataset_src/split_dataset.py +63 -0
- paddlex/modules/ts_anomaly_detection/evaluator.py +67 -0
- paddlex/modules/ts_anomaly_detection/exportor.py +44 -0
- paddlex/modules/ts_anomaly_detection/model_list.py +22 -0
- paddlex/modules/ts_anomaly_detection/trainer.py +113 -0
- paddlex/modules/ts_classification/__init__.py +19 -0
- paddlex/modules/ts_classification/dataset_checker/__init__.py +111 -0
- paddlex/modules/ts_classification/dataset_checker/dataset_src/__init__.py +19 -0
- paddlex/modules/ts_classification/dataset_checker/dataset_src/analyse_dataset.py +77 -0
- paddlex/modules/ts_classification/dataset_checker/dataset_src/check_dataset.py +64 -0
- paddlex/modules/ts_classification/dataset_checker/dataset_src/convert_dataset.py +74 -0
- paddlex/modules/ts_classification/dataset_checker/dataset_src/split_dataset.py +88 -0
- paddlex/modules/ts_classification/evaluator.py +66 -0
- paddlex/modules/ts_classification/exportor.py +44 -0
- paddlex/modules/ts_classification/model_list.py +18 -0
- paddlex/modules/ts_classification/trainer.py +108 -0
- paddlex/modules/ts_forecast/__init__.py +19 -0
- paddlex/modules/ts_forecast/dataset_checker/__init__.py +111 -0
- paddlex/modules/ts_forecast/dataset_checker/dataset_src/__init__.py +19 -0
- paddlex/modules/ts_forecast/dataset_checker/dataset_src/analyse_dataset.py +19 -0
- paddlex/modules/ts_forecast/dataset_checker/dataset_src/check_dataset.py +64 -0
- paddlex/modules/ts_forecast/dataset_checker/dataset_src/convert_dataset.py +73 -0
- paddlex/modules/ts_forecast/dataset_checker/dataset_src/split_dataset.py +63 -0
- paddlex/modules/ts_forecast/evaluator.py +66 -0
- paddlex/modules/ts_forecast/exportor.py +44 -0
- paddlex/modules/ts_forecast/model_list.py +24 -0
- paddlex/modules/ts_forecast/trainer.py +108 -0
- paddlex/modules/video_classification/__init__.py +18 -0
- paddlex/modules/video_classification/dataset_checker/__init__.py +93 -0
- paddlex/modules/video_classification/dataset_checker/dataset_src/__init__.py +18 -0
- paddlex/modules/video_classification/dataset_checker/dataset_src/analyse_dataset.py +93 -0
- paddlex/modules/video_classification/dataset_checker/dataset_src/check_dataset.py +120 -0
- paddlex/modules/video_classification/dataset_checker/dataset_src/split_dataset.py +82 -0
- paddlex/modules/video_classification/evaluator.py +44 -0
- paddlex/modules/video_classification/exportor.py +22 -0
- paddlex/modules/video_classification/model_list.py +19 -0
- paddlex/modules/video_classification/trainer.py +88 -0
- paddlex/modules/video_detection/__init__.py +18 -0
- paddlex/modules/video_detection/dataset_checker/__init__.py +86 -0
- paddlex/modules/video_detection/dataset_checker/dataset_src/__init__.py +17 -0
- paddlex/modules/video_detection/dataset_checker/dataset_src/analyse_dataset.py +100 -0
- paddlex/modules/video_detection/dataset_checker/dataset_src/check_dataset.py +132 -0
- paddlex/modules/video_detection/evaluator.py +42 -0
- paddlex/modules/video_detection/exportor.py +22 -0
- paddlex/modules/video_detection/model_list.py +15 -0
- paddlex/modules/video_detection/trainer.py +82 -0
- paddlex/ops/__init__.py +152 -0
- paddlex/ops/iou3d_nms/iou3d_cpu.cpp +266 -0
- paddlex/ops/iou3d_nms/iou3d_cpu.h +28 -0
- paddlex/ops/iou3d_nms/iou3d_nms.cpp +206 -0
- paddlex/ops/iou3d_nms/iou3d_nms.h +35 -0
- paddlex/ops/iou3d_nms/iou3d_nms_api.cpp +114 -0
- paddlex/ops/iou3d_nms/iou3d_nms_kernel.cu +484 -0
- paddlex/ops/setup.py +37 -0
- paddlex/ops/voxel/voxelize_op.cc +194 -0
- paddlex/ops/voxel/voxelize_op.cu +346 -0
- paddlex/paddlex_cli.py +476 -0
- paddlex/repo_apis/Paddle3D_api/__init__.py +17 -0
- paddlex/repo_apis/Paddle3D_api/bev_fusion/__init__.py +18 -0
- paddlex/repo_apis/Paddle3D_api/bev_fusion/config.py +118 -0
- paddlex/repo_apis/Paddle3D_api/bev_fusion/model.py +238 -0
- paddlex/repo_apis/Paddle3D_api/bev_fusion/register.py +55 -0
- paddlex/repo_apis/Paddle3D_api/bev_fusion/runner.py +104 -0
- paddlex/repo_apis/Paddle3D_api/pp3d_config.py +145 -0
- paddlex/repo_apis/PaddleClas_api/__init__.py +17 -0
- paddlex/repo_apis/PaddleClas_api/cls/__init__.py +19 -0
- paddlex/repo_apis/PaddleClas_api/cls/config.py +595 -0
- paddlex/repo_apis/PaddleClas_api/cls/model.py +355 -0
- paddlex/repo_apis/PaddleClas_api/cls/register.py +907 -0
- paddlex/repo_apis/PaddleClas_api/cls/runner.py +218 -0
- paddlex/repo_apis/PaddleClas_api/shitu_rec/__init__.py +18 -0
- paddlex/repo_apis/PaddleClas_api/shitu_rec/config.py +141 -0
- paddlex/repo_apis/PaddleClas_api/shitu_rec/model.py +20 -0
- paddlex/repo_apis/PaddleClas_api/shitu_rec/register.py +68 -0
- paddlex/repo_apis/PaddleClas_api/shitu_rec/runner.py +50 -0
- paddlex/repo_apis/PaddleDetection_api/__init__.py +17 -0
- paddlex/repo_apis/PaddleDetection_api/config_helper.py +280 -0
- paddlex/repo_apis/PaddleDetection_api/instance_seg/__init__.py +18 -0
- paddlex/repo_apis/PaddleDetection_api/instance_seg/config.py +457 -0
- paddlex/repo_apis/PaddleDetection_api/instance_seg/model.py +403 -0
- paddlex/repo_apis/PaddleDetection_api/instance_seg/register.py +262 -0
- paddlex/repo_apis/PaddleDetection_api/instance_seg/runner.py +225 -0
- paddlex/repo_apis/PaddleDetection_api/object_det/__init__.py +19 -0
- paddlex/repo_apis/PaddleDetection_api/object_det/config.py +540 -0
- paddlex/repo_apis/PaddleDetection_api/object_det/model.py +429 -0
- paddlex/repo_apis/PaddleDetection_api/object_det/official_categories.py +245 -0
- paddlex/repo_apis/PaddleDetection_api/object_det/register.py +1135 -0
- paddlex/repo_apis/PaddleDetection_api/object_det/runner.py +225 -0
- paddlex/repo_apis/PaddleNLP_api/__init__.py +13 -0
- paddlex/repo_apis/PaddleOCR_api/__init__.py +22 -0
- paddlex/repo_apis/PaddleOCR_api/config_utils.py +53 -0
- paddlex/repo_apis/PaddleOCR_api/formula_rec/__init__.py +16 -0
- paddlex/repo_apis/PaddleOCR_api/formula_rec/config.py +571 -0
- paddlex/repo_apis/PaddleOCR_api/formula_rec/model.py +398 -0
- paddlex/repo_apis/PaddleOCR_api/formula_rec/register.py +99 -0
- paddlex/repo_apis/PaddleOCR_api/formula_rec/runner.py +239 -0
- paddlex/repo_apis/PaddleOCR_api/table_rec/__init__.py +16 -0
- paddlex/repo_apis/PaddleOCR_api/table_rec/config.py +64 -0
- paddlex/repo_apis/PaddleOCR_api/table_rec/model.py +126 -0
- paddlex/repo_apis/PaddleOCR_api/table_rec/register.py +70 -0
- paddlex/repo_apis/PaddleOCR_api/table_rec/runner.py +51 -0
- paddlex/repo_apis/PaddleOCR_api/text_det/__init__.py +16 -0
- paddlex/repo_apis/PaddleOCR_api/text_det/config.py +62 -0
- paddlex/repo_apis/PaddleOCR_api/text_det/model.py +72 -0
- paddlex/repo_apis/PaddleOCR_api/text_det/register.py +107 -0
- paddlex/repo_apis/PaddleOCR_api/text_det/runner.py +53 -0
- paddlex/repo_apis/PaddleOCR_api/text_rec/__init__.py +16 -0
- paddlex/repo_apis/PaddleOCR_api/text_rec/config.py +564 -0
- paddlex/repo_apis/PaddleOCR_api/text_rec/model.py +398 -0
- paddlex/repo_apis/PaddleOCR_api/text_rec/register.py +216 -0
- paddlex/repo_apis/PaddleOCR_api/text_rec/runner.py +239 -0
- paddlex/repo_apis/PaddleSeg_api/__init__.py +16 -0
- paddlex/repo_apis/PaddleSeg_api/base_seg_config.py +134 -0
- paddlex/repo_apis/PaddleSeg_api/seg/__init__.py +16 -0
- paddlex/repo_apis/PaddleSeg_api/seg/config.py +183 -0
- paddlex/repo_apis/PaddleSeg_api/seg/model.py +491 -0
- paddlex/repo_apis/PaddleSeg_api/seg/register.py +272 -0
- paddlex/repo_apis/PaddleSeg_api/seg/runner.py +261 -0
- paddlex/repo_apis/PaddleTS_api/__init__.py +20 -0
- paddlex/repo_apis/PaddleTS_api/ts_ad/__init__.py +16 -0
- paddlex/repo_apis/PaddleTS_api/ts_ad/config.py +88 -0
- paddlex/repo_apis/PaddleTS_api/ts_ad/register.py +146 -0
- paddlex/repo_apis/PaddleTS_api/ts_ad/runner.py +158 -0
- paddlex/repo_apis/PaddleTS_api/ts_base/__init__.py +13 -0
- paddlex/repo_apis/PaddleTS_api/ts_base/config.py +244 -0
- paddlex/repo_apis/PaddleTS_api/ts_base/model.py +276 -0
- paddlex/repo_apis/PaddleTS_api/ts_base/runner.py +158 -0
- paddlex/repo_apis/PaddleTS_api/ts_cls/__init__.py +16 -0
- paddlex/repo_apis/PaddleTS_api/ts_cls/config.py +72 -0
- paddlex/repo_apis/PaddleTS_api/ts_cls/register.py +59 -0
- paddlex/repo_apis/PaddleTS_api/ts_cls/runner.py +158 -0
- paddlex/repo_apis/PaddleTS_api/ts_fc/__init__.py +16 -0
- paddlex/repo_apis/PaddleTS_api/ts_fc/config.py +136 -0
- paddlex/repo_apis/PaddleTS_api/ts_fc/register.py +186 -0
- paddlex/repo_apis/PaddleVideo_api/__init__.py +17 -0
- paddlex/repo_apis/PaddleVideo_api/config_utils.py +51 -0
- paddlex/repo_apis/PaddleVideo_api/video_cls/__init__.py +19 -0
- paddlex/repo_apis/PaddleVideo_api/video_cls/config.py +548 -0
- paddlex/repo_apis/PaddleVideo_api/video_cls/model.py +346 -0
- paddlex/repo_apis/PaddleVideo_api/video_cls/register.py +70 -0
- paddlex/repo_apis/PaddleVideo_api/video_cls/runner.py +204 -0
- paddlex/repo_apis/PaddleVideo_api/video_det/__init__.py +19 -0
- paddlex/repo_apis/PaddleVideo_api/video_det/config.py +549 -0
- paddlex/repo_apis/PaddleVideo_api/video_det/model.py +298 -0
- paddlex/repo_apis/PaddleVideo_api/video_det/register.py +44 -0
- paddlex/repo_apis/PaddleVideo_api/video_det/runner.py +199 -0
- paddlex/repo_apis/__init__.py +13 -0
- paddlex/repo_apis/base/__init__.py +22 -0
- paddlex/repo_apis/base/config.py +237 -0
- paddlex/repo_apis/base/model.py +563 -0
- paddlex/repo_apis/base/register.py +135 -0
- paddlex/repo_apis/base/runner.py +390 -0
- paddlex/repo_apis/base/utils/__init__.py +13 -0
- paddlex/repo_apis/base/utils/arg.py +64 -0
- paddlex/repo_apis/base/utils/subprocess.py +107 -0
- paddlex/repo_manager/__init__.py +17 -0
- paddlex/repo_manager/core.py +253 -0
- paddlex/repo_manager/meta.py +180 -0
- paddlex/repo_manager/repo.py +425 -0
- paddlex/repo_manager/utils.py +148 -0
- paddlex/utils/__init__.py +1 -12
- paddlex/utils/cache.py +146 -0
- paddlex/utils/config.py +216 -0
- paddlex/utils/custom_device_list.py +311 -0
- paddlex/utils/deps.py +249 -0
- paddlex/utils/device.py +195 -0
- paddlex/utils/download.py +168 -182
- paddlex/utils/env.py +31 -48
- paddlex/utils/errors/__init__.py +17 -0
- paddlex/utils/errors/dataset_checker.py +78 -0
- paddlex/utils/errors/others.py +138 -0
- paddlex/utils/file_interface.py +211 -0
- paddlex/utils/flags.py +70 -0
- paddlex/utils/fonts/__init__.py +97 -0
- paddlex/utils/func_register.py +41 -0
- paddlex/utils/install.py +87 -0
- paddlex/utils/interactive_get_pipeline.py +55 -0
- paddlex/utils/lazy_loader.py +68 -0
- paddlex/utils/logging.py +140 -33
- paddlex/utils/misc.py +201 -0
- paddlex/utils/pipeline_arguments.py +719 -0
- paddlex/utils/result_saver.py +58 -0
- paddlex/utils/subclass_register.py +99 -0
- paddlex/version.py +55 -0
- paddlex-3.0.0.dist-info/METADATA +1168 -0
- paddlex-3.0.0.dist-info/RECORD +1093 -0
- paddlex-3.0.0.dist-info/WHEEL +5 -0
- paddlex-3.0.0.dist-info/entry_points.txt +2 -0
- paddlex-3.0.0.dist-info/licenses/LICENSE +169 -0
- paddlex-3.0.0.dist-info/top_level.txt +1 -0
- PaddleClas/__init__.py +0 -16
- PaddleClas/paddleclas.py +0 -375
- PaddleClas/ppcls/__init__.py +0 -20
- PaddleClas/ppcls/data/__init__.py +0 -15
- PaddleClas/ppcls/data/imaug/__init__.py +0 -94
- PaddleClas/ppcls/data/imaug/autoaugment.py +0 -264
- PaddleClas/ppcls/data/imaug/batch_operators.py +0 -117
- PaddleClas/ppcls/data/imaug/cutout.py +0 -41
- PaddleClas/ppcls/data/imaug/fmix.py +0 -217
- PaddleClas/ppcls/data/imaug/grid.py +0 -89
- PaddleClas/ppcls/data/imaug/hide_and_seek.py +0 -44
- PaddleClas/ppcls/data/imaug/operators.py +0 -244
- PaddleClas/ppcls/data/imaug/randaugment.py +0 -106
- PaddleClas/ppcls/data/imaug/random_erasing.py +0 -55
- PaddleClas/ppcls/data/reader.py +0 -318
- PaddleClas/ppcls/modeling/__init__.py +0 -20
- PaddleClas/ppcls/modeling/architectures/__init__.py +0 -51
- PaddleClas/ppcls/modeling/architectures/alexnet.py +0 -132
- PaddleClas/ppcls/modeling/architectures/darknet.py +0 -161
- PaddleClas/ppcls/modeling/architectures/densenet.py +0 -308
- PaddleClas/ppcls/modeling/architectures/distillation_models.py +0 -65
- PaddleClas/ppcls/modeling/architectures/distilled_vision_transformer.py +0 -196
- PaddleClas/ppcls/modeling/architectures/dpn.py +0 -425
- PaddleClas/ppcls/modeling/architectures/efficientnet.py +0 -901
- PaddleClas/ppcls/modeling/architectures/ghostnet.py +0 -331
- PaddleClas/ppcls/modeling/architectures/googlenet.py +0 -207
- PaddleClas/ppcls/modeling/architectures/hrnet.py +0 -742
- PaddleClas/ppcls/modeling/architectures/inception_v3.py +0 -481
- PaddleClas/ppcls/modeling/architectures/inception_v4.py +0 -455
- PaddleClas/ppcls/modeling/architectures/mixnet.py +0 -782
- PaddleClas/ppcls/modeling/architectures/mobilenet_v1.py +0 -266
- PaddleClas/ppcls/modeling/architectures/mobilenet_v2.py +0 -248
- PaddleClas/ppcls/modeling/architectures/mobilenet_v3.py +0 -359
- PaddleClas/ppcls/modeling/architectures/regnet.py +0 -383
- PaddleClas/ppcls/modeling/architectures/repvgg.py +0 -339
- PaddleClas/ppcls/modeling/architectures/res2net.py +0 -272
- PaddleClas/ppcls/modeling/architectures/res2net_vd.py +0 -295
- PaddleClas/ppcls/modeling/architectures/resnest.py +0 -705
- PaddleClas/ppcls/modeling/architectures/resnet.py +0 -316
- PaddleClas/ppcls/modeling/architectures/resnet_vc.py +0 -309
- PaddleClas/ppcls/modeling/architectures/resnet_vd.py +0 -354
- PaddleClas/ppcls/modeling/architectures/resnext.py +0 -253
- PaddleClas/ppcls/modeling/architectures/resnext101_wsl.py +0 -447
- PaddleClas/ppcls/modeling/architectures/resnext_vd.py +0 -266
- PaddleClas/ppcls/modeling/architectures/rexnet.py +0 -240
- PaddleClas/ppcls/modeling/architectures/se_resnet_vd.py +0 -378
- PaddleClas/ppcls/modeling/architectures/se_resnext.py +0 -290
- PaddleClas/ppcls/modeling/architectures/se_resnext_vd.py +0 -285
- PaddleClas/ppcls/modeling/architectures/shufflenet_v2.py +0 -320
- PaddleClas/ppcls/modeling/architectures/squeezenet.py +0 -154
- PaddleClas/ppcls/modeling/architectures/vgg.py +0 -152
- PaddleClas/ppcls/modeling/architectures/vision_transformer.py +0 -402
- PaddleClas/ppcls/modeling/architectures/xception.py +0 -345
- PaddleClas/ppcls/modeling/architectures/xception_deeplab.py +0 -386
- PaddleClas/ppcls/modeling/loss.py +0 -154
- PaddleClas/ppcls/modeling/utils.py +0 -53
- PaddleClas/ppcls/optimizer/__init__.py +0 -19
- PaddleClas/ppcls/optimizer/learning_rate.py +0 -159
- PaddleClas/ppcls/optimizer/optimizer.py +0 -165
- PaddleClas/ppcls/utils/__init__.py +0 -27
- PaddleClas/ppcls/utils/check.py +0 -151
- PaddleClas/ppcls/utils/config.py +0 -201
- PaddleClas/ppcls/utils/logger.py +0 -120
- PaddleClas/ppcls/utils/metrics.py +0 -107
- PaddleClas/ppcls/utils/misc.py +0 -62
- PaddleClas/ppcls/utils/model_zoo.py +0 -213
- PaddleClas/ppcls/utils/save_load.py +0 -163
- PaddleClas/setup.py +0 -55
- PaddleClas/tools/__init__.py +0 -15
- PaddleClas/tools/download.py +0 -50
- PaddleClas/tools/ema.py +0 -58
- PaddleClas/tools/eval.py +0 -112
- PaddleClas/tools/export_model.py +0 -85
- PaddleClas/tools/export_serving_model.py +0 -76
- PaddleClas/tools/infer/__init__.py +0 -16
- PaddleClas/tools/infer/infer.py +0 -94
- PaddleClas/tools/infer/predict.py +0 -117
- PaddleClas/tools/infer/utils.py +0 -233
- PaddleClas/tools/program.py +0 -444
- PaddleClas/tools/test_hubserving.py +0 -113
- PaddleClas/tools/train.py +0 -141
- paddlex/cls.py +0 -76
- paddlex/command.py +0 -215
- paddlex/cv/__init__.py +0 -17
- paddlex/cv/datasets/__init__.py +0 -18
- paddlex/cv/datasets/coco.py +0 -169
- paddlex/cv/datasets/imagenet.py +0 -88
- paddlex/cv/datasets/seg_dataset.py +0 -91
- paddlex/cv/datasets/voc.py +0 -301
- paddlex/cv/models/__init__.py +0 -18
- paddlex/cv/models/base.py +0 -623
- paddlex/cv/models/classifier.py +0 -814
- paddlex/cv/models/detector.py +0 -1747
- paddlex/cv/models/load_model.py +0 -126
- paddlex/cv/models/segmenter.py +0 -673
- paddlex/cv/models/slim/__init__.py +0 -13
- paddlex/cv/models/slim/prune.py +0 -55
- paddlex/cv/models/utils/__init__.py +0 -13
- paddlex/cv/models/utils/det_metrics/__init__.py +0 -15
- paddlex/cv/models/utils/det_metrics/coco_utils.py +0 -217
- paddlex/cv/models/utils/det_metrics/metrics.py +0 -220
- paddlex/cv/models/utils/ema.py +0 -48
- paddlex/cv/models/utils/seg_metrics.py +0 -62
- paddlex/cv/models/utils/visualize.py +0 -394
- paddlex/cv/transforms/__init__.py +0 -46
- paddlex/cv/transforms/batch_operators.py +0 -286
- paddlex/cv/transforms/box_utils.py +0 -41
- paddlex/cv/transforms/functions.py +0 -193
- paddlex/cv/transforms/operators.py +0 -1402
- paddlex/det.py +0 -43
- paddlex/paddleseg/__init__.py +0 -17
- paddlex/paddleseg/core/__init__.py +0 -20
- paddlex/paddleseg/core/infer.py +0 -289
- paddlex/paddleseg/core/predict.py +0 -145
- paddlex/paddleseg/core/train.py +0 -258
- paddlex/paddleseg/core/val.py +0 -172
- paddlex/paddleseg/cvlibs/__init__.py +0 -17
- paddlex/paddleseg/cvlibs/callbacks.py +0 -279
- paddlex/paddleseg/cvlibs/config.py +0 -359
- paddlex/paddleseg/cvlibs/manager.py +0 -142
- paddlex/paddleseg/cvlibs/param_init.py +0 -91
- paddlex/paddleseg/datasets/__init__.py +0 -21
- paddlex/paddleseg/datasets/ade.py +0 -112
- paddlex/paddleseg/datasets/cityscapes.py +0 -86
- paddlex/paddleseg/datasets/cocostuff.py +0 -79
- paddlex/paddleseg/datasets/dataset.py +0 -164
- paddlex/paddleseg/datasets/mini_deep_globe_road_extraction.py +0 -95
- paddlex/paddleseg/datasets/optic_disc_seg.py +0 -97
- paddlex/paddleseg/datasets/pascal_context.py +0 -80
- paddlex/paddleseg/datasets/voc.py +0 -113
- paddlex/paddleseg/models/__init__.py +0 -39
- paddlex/paddleseg/models/ann.py +0 -436
- paddlex/paddleseg/models/attention_unet.py +0 -189
- paddlex/paddleseg/models/backbones/__init__.py +0 -18
- paddlex/paddleseg/models/backbones/hrnet.py +0 -815
- paddlex/paddleseg/models/backbones/mobilenetv3.py +0 -365
- paddlex/paddleseg/models/backbones/resnet_vd.py +0 -364
- paddlex/paddleseg/models/backbones/xception_deeplab.py +0 -415
- paddlex/paddleseg/models/bisenet.py +0 -311
- paddlex/paddleseg/models/danet.py +0 -220
- paddlex/paddleseg/models/decoupled_segnet.py +0 -233
- paddlex/paddleseg/models/deeplab.py +0 -258
- paddlex/paddleseg/models/dnlnet.py +0 -231
- paddlex/paddleseg/models/emanet.py +0 -219
- paddlex/paddleseg/models/fast_scnn.py +0 -318
- paddlex/paddleseg/models/fcn.py +0 -135
- paddlex/paddleseg/models/gcnet.py +0 -223
- paddlex/paddleseg/models/gscnn.py +0 -357
- paddlex/paddleseg/models/hardnet.py +0 -309
- paddlex/paddleseg/models/isanet.py +0 -202
- paddlex/paddleseg/models/layers/__init__.py +0 -19
- paddlex/paddleseg/models/layers/activation.py +0 -73
- paddlex/paddleseg/models/layers/attention.py +0 -146
- paddlex/paddleseg/models/layers/layer_libs.py +0 -168
- paddlex/paddleseg/models/layers/nonlocal2d.py +0 -155
- paddlex/paddleseg/models/layers/pyramid_pool.py +0 -182
- paddlex/paddleseg/models/losses/__init__.py +0 -27
- paddlex/paddleseg/models/losses/binary_cross_entropy_loss.py +0 -174
- paddlex/paddleseg/models/losses/bootstrapped_cross_entropy.py +0 -73
- paddlex/paddleseg/models/losses/cross_entropy_loss.py +0 -94
- paddlex/paddleseg/models/losses/decoupledsegnet_relax_boundary_loss.py +0 -129
- paddlex/paddleseg/models/losses/dice_loss.py +0 -61
- paddlex/paddleseg/models/losses/edge_attention_loss.py +0 -78
- paddlex/paddleseg/models/losses/gscnn_dual_task_loss.py +0 -141
- paddlex/paddleseg/models/losses/l1_loss.py +0 -76
- paddlex/paddleseg/models/losses/lovasz_loss.py +0 -222
- paddlex/paddleseg/models/losses/mean_square_error_loss.py +0 -65
- paddlex/paddleseg/models/losses/mixed_loss.py +0 -58
- paddlex/paddleseg/models/losses/ohem_cross_entropy_loss.py +0 -99
- paddlex/paddleseg/models/losses/ohem_edge_attention_loss.py +0 -114
- paddlex/paddleseg/models/ocrnet.py +0 -248
- paddlex/paddleseg/models/pspnet.py +0 -147
- paddlex/paddleseg/models/sfnet.py +0 -236
- paddlex/paddleseg/models/shufflenet_slim.py +0 -268
- paddlex/paddleseg/models/u2net.py +0 -574
- paddlex/paddleseg/models/unet.py +0 -155
- paddlex/paddleseg/models/unet_3plus.py +0 -316
- paddlex/paddleseg/models/unet_plusplus.py +0 -237
- paddlex/paddleseg/transforms/__init__.py +0 -16
- paddlex/paddleseg/transforms/functional.py +0 -161
- paddlex/paddleseg/transforms/transforms.py +0 -937
- paddlex/paddleseg/utils/__init__.py +0 -22
- paddlex/paddleseg/utils/config_check.py +0 -60
- paddlex/paddleseg/utils/download.py +0 -163
- paddlex/paddleseg/utils/env/__init__.py +0 -16
- paddlex/paddleseg/utils/env/seg_env.py +0 -56
- paddlex/paddleseg/utils/env/sys_env.py +0 -122
- paddlex/paddleseg/utils/logger.py +0 -48
- paddlex/paddleseg/utils/metrics.py +0 -146
- paddlex/paddleseg/utils/progbar.py +0 -212
- paddlex/paddleseg/utils/timer.py +0 -53
- paddlex/paddleseg/utils/utils.py +0 -120
- paddlex/paddleseg/utils/visualize.py +0 -90
- paddlex/ppcls/__init__.py +0 -20
- paddlex/ppcls/data/__init__.py +0 -15
- paddlex/ppcls/data/imaug/__init__.py +0 -94
- paddlex/ppcls/data/imaug/autoaugment.py +0 -264
- paddlex/ppcls/data/imaug/batch_operators.py +0 -117
- paddlex/ppcls/data/imaug/cutout.py +0 -41
- paddlex/ppcls/data/imaug/fmix.py +0 -217
- paddlex/ppcls/data/imaug/grid.py +0 -89
- paddlex/ppcls/data/imaug/hide_and_seek.py +0 -44
- paddlex/ppcls/data/imaug/operators.py +0 -256
- paddlex/ppcls/data/imaug/randaugment.py +0 -106
- paddlex/ppcls/data/imaug/random_erasing.py +0 -55
- paddlex/ppcls/data/reader.py +0 -318
- paddlex/ppcls/modeling/__init__.py +0 -20
- paddlex/ppcls/modeling/architectures/__init__.py +0 -51
- paddlex/ppcls/modeling/architectures/alexnet.py +0 -132
- paddlex/ppcls/modeling/architectures/darknet.py +0 -161
- paddlex/ppcls/modeling/architectures/densenet.py +0 -308
- paddlex/ppcls/modeling/architectures/distillation_models.py +0 -65
- paddlex/ppcls/modeling/architectures/distilled_vision_transformer.py +0 -196
- paddlex/ppcls/modeling/architectures/dpn.py +0 -425
- paddlex/ppcls/modeling/architectures/efficientnet.py +0 -901
- paddlex/ppcls/modeling/architectures/ghostnet.py +0 -331
- paddlex/ppcls/modeling/architectures/googlenet.py +0 -207
- paddlex/ppcls/modeling/architectures/hrnet.py +0 -742
- paddlex/ppcls/modeling/architectures/inception_v3.py +0 -541
- paddlex/ppcls/modeling/architectures/inception_v4.py +0 -455
- paddlex/ppcls/modeling/architectures/mixnet.py +0 -782
- paddlex/ppcls/modeling/architectures/mobilenet_v1.py +0 -266
- paddlex/ppcls/modeling/architectures/mobilenet_v2.py +0 -248
- paddlex/ppcls/modeling/architectures/mobilenet_v3.py +0 -359
- paddlex/ppcls/modeling/architectures/regnet.py +0 -383
- paddlex/ppcls/modeling/architectures/repvgg.py +0 -339
- paddlex/ppcls/modeling/architectures/res2net.py +0 -272
- paddlex/ppcls/modeling/architectures/res2net_vd.py +0 -295
- paddlex/ppcls/modeling/architectures/resnest.py +0 -705
- paddlex/ppcls/modeling/architectures/resnet.py +0 -317
- paddlex/ppcls/modeling/architectures/resnet_vc.py +0 -309
- paddlex/ppcls/modeling/architectures/resnet_vd.py +0 -354
- paddlex/ppcls/modeling/architectures/resnext.py +0 -259
- paddlex/ppcls/modeling/architectures/resnext101_wsl.py +0 -447
- paddlex/ppcls/modeling/architectures/resnext_vd.py +0 -266
- paddlex/ppcls/modeling/architectures/rexnet.py +0 -240
- paddlex/ppcls/modeling/architectures/se_resnet_vd.py +0 -378
- paddlex/ppcls/modeling/architectures/se_resnext.py +0 -290
- paddlex/ppcls/modeling/architectures/se_resnext_vd.py +0 -285
- paddlex/ppcls/modeling/architectures/shufflenet_v2.py +0 -320
- paddlex/ppcls/modeling/architectures/squeezenet.py +0 -154
- paddlex/ppcls/modeling/architectures/vgg.py +0 -152
- paddlex/ppcls/modeling/architectures/vision_transformer.py +0 -402
- paddlex/ppcls/modeling/architectures/xception.py +0 -345
- paddlex/ppcls/modeling/architectures/xception_deeplab.py +0 -386
- paddlex/ppcls/modeling/loss.py +0 -158
- paddlex/ppcls/modeling/utils.py +0 -53
- paddlex/ppcls/optimizer/__init__.py +0 -19
- paddlex/ppcls/optimizer/learning_rate.py +0 -159
- paddlex/ppcls/optimizer/optimizer.py +0 -165
- paddlex/ppcls/utils/__init__.py +0 -27
- paddlex/ppcls/utils/check.py +0 -151
- paddlex/ppcls/utils/config.py +0 -201
- paddlex/ppcls/utils/logger.py +0 -120
- paddlex/ppcls/utils/metrics.py +0 -112
- paddlex/ppcls/utils/misc.py +0 -62
- paddlex/ppcls/utils/model_zoo.py +0 -213
- paddlex/ppcls/utils/save_load.py +0 -163
- paddlex/ppdet/__init__.py +0 -16
- paddlex/ppdet/core/__init__.py +0 -15
- paddlex/ppdet/core/config/__init__.py +0 -13
- paddlex/ppdet/core/config/schema.py +0 -248
- paddlex/ppdet/core/config/yaml_helpers.py +0 -118
- paddlex/ppdet/core/workspace.py +0 -279
- paddlex/ppdet/data/__init__.py +0 -21
- paddlex/ppdet/data/reader.py +0 -304
- paddlex/ppdet/data/shm_utils.py +0 -67
- paddlex/ppdet/data/source/__init__.py +0 -27
- paddlex/ppdet/data/source/category.py +0 -823
- paddlex/ppdet/data/source/coco.py +0 -243
- paddlex/ppdet/data/source/dataset.py +0 -192
- paddlex/ppdet/data/source/keypoint_coco.py +0 -656
- paddlex/ppdet/data/source/mot.py +0 -360
- paddlex/ppdet/data/source/voc.py +0 -204
- paddlex/ppdet/data/source/widerface.py +0 -180
- paddlex/ppdet/data/transform/__init__.py +0 -28
- paddlex/ppdet/data/transform/autoaugment_utils.py +0 -1593
- paddlex/ppdet/data/transform/batch_operators.py +0 -758
- paddlex/ppdet/data/transform/gridmask_utils.py +0 -83
- paddlex/ppdet/data/transform/keypoint_operators.py +0 -665
- paddlex/ppdet/data/transform/mot_operators.py +0 -636
- paddlex/ppdet/data/transform/op_helper.py +0 -468
- paddlex/ppdet/data/transform/operators.py +0 -2103
- paddlex/ppdet/engine/__init__.py +0 -29
- paddlex/ppdet/engine/callbacks.py +0 -262
- paddlex/ppdet/engine/env.py +0 -47
- paddlex/ppdet/engine/export_utils.py +0 -118
- paddlex/ppdet/engine/tracker.py +0 -425
- paddlex/ppdet/engine/trainer.py +0 -535
- paddlex/ppdet/metrics/__init__.py +0 -23
- paddlex/ppdet/metrics/coco_utils.py +0 -184
- paddlex/ppdet/metrics/json_results.py +0 -151
- paddlex/ppdet/metrics/keypoint_metrics.py +0 -202
- paddlex/ppdet/metrics/map_utils.py +0 -396
- paddlex/ppdet/metrics/metrics.py +0 -300
- paddlex/ppdet/metrics/mot_eval_utils.py +0 -192
- paddlex/ppdet/metrics/mot_metrics.py +0 -184
- paddlex/ppdet/metrics/widerface_utils.py +0 -393
- paddlex/ppdet/model_zoo/__init__.py +0 -18
- paddlex/ppdet/model_zoo/model_zoo.py +0 -86
- paddlex/ppdet/model_zoo/tests/__init__.py +0 -13
- paddlex/ppdet/model_zoo/tests/test_get_model.py +0 -48
- paddlex/ppdet/model_zoo/tests/test_list_model.py +0 -68
- paddlex/ppdet/modeling/__init__.py +0 -41
- paddlex/ppdet/modeling/architectures/__init__.py +0 -40
- paddlex/ppdet/modeling/architectures/cascade_rcnn.py +0 -144
- paddlex/ppdet/modeling/architectures/centernet.py +0 -103
- paddlex/ppdet/modeling/architectures/deepsort.py +0 -111
- paddlex/ppdet/modeling/architectures/fairmot.py +0 -107
- paddlex/ppdet/modeling/architectures/faster_rcnn.py +0 -106
- paddlex/ppdet/modeling/architectures/fcos.py +0 -105
- paddlex/ppdet/modeling/architectures/jde.py +0 -125
- paddlex/ppdet/modeling/architectures/keypoint_hrhrnet.py +0 -286
- paddlex/ppdet/modeling/architectures/keypoint_hrnet.py +0 -203
- paddlex/ppdet/modeling/architectures/mask_rcnn.py +0 -135
- paddlex/ppdet/modeling/architectures/meta_arch.py +0 -45
- paddlex/ppdet/modeling/architectures/s2anet.py +0 -103
- paddlex/ppdet/modeling/architectures/solov2.py +0 -110
- paddlex/ppdet/modeling/architectures/ssd.py +0 -84
- paddlex/ppdet/modeling/architectures/ttfnet.py +0 -98
- paddlex/ppdet/modeling/architectures/yolo.py +0 -104
- paddlex/ppdet/modeling/backbones/__init__.py +0 -37
- paddlex/ppdet/modeling/backbones/blazenet.py +0 -322
- paddlex/ppdet/modeling/backbones/darknet.py +0 -341
- paddlex/ppdet/modeling/backbones/dla.py +0 -244
- paddlex/ppdet/modeling/backbones/ghostnet.py +0 -476
- paddlex/ppdet/modeling/backbones/hrnet.py +0 -724
- paddlex/ppdet/modeling/backbones/mobilenet_v1.py +0 -410
- paddlex/ppdet/modeling/backbones/mobilenet_v3.py +0 -497
- paddlex/ppdet/modeling/backbones/name_adapter.py +0 -69
- paddlex/ppdet/modeling/backbones/res2net.py +0 -358
- paddlex/ppdet/modeling/backbones/resnet.py +0 -606
- paddlex/ppdet/modeling/backbones/senet.py +0 -140
- paddlex/ppdet/modeling/backbones/vgg.py +0 -216
- paddlex/ppdet/modeling/bbox_utils.py +0 -464
- paddlex/ppdet/modeling/heads/__init__.py +0 -41
- paddlex/ppdet/modeling/heads/bbox_head.py +0 -379
- paddlex/ppdet/modeling/heads/cascade_head.py +0 -285
- paddlex/ppdet/modeling/heads/centernet_head.py +0 -194
- paddlex/ppdet/modeling/heads/face_head.py +0 -113
- paddlex/ppdet/modeling/heads/fcos_head.py +0 -270
- paddlex/ppdet/modeling/heads/keypoint_hrhrnet_head.py +0 -108
- paddlex/ppdet/modeling/heads/mask_head.py +0 -253
- paddlex/ppdet/modeling/heads/roi_extractor.py +0 -111
- paddlex/ppdet/modeling/heads/s2anet_head.py +0 -845
- paddlex/ppdet/modeling/heads/solov2_head.py +0 -537
- paddlex/ppdet/modeling/heads/ssd_head.py +0 -175
- paddlex/ppdet/modeling/heads/ttf_head.py +0 -314
- paddlex/ppdet/modeling/heads/yolo_head.py +0 -124
- paddlex/ppdet/modeling/keypoint_utils.py +0 -302
- paddlex/ppdet/modeling/layers.py +0 -1142
- paddlex/ppdet/modeling/losses/__init__.py +0 -35
- paddlex/ppdet/modeling/losses/ctfocal_loss.py +0 -67
- paddlex/ppdet/modeling/losses/fairmot_loss.py +0 -41
- paddlex/ppdet/modeling/losses/fcos_loss.py +0 -225
- paddlex/ppdet/modeling/losses/iou_aware_loss.py +0 -48
- paddlex/ppdet/modeling/losses/iou_loss.py +0 -210
- paddlex/ppdet/modeling/losses/jde_loss.py +0 -182
- paddlex/ppdet/modeling/losses/keypoint_loss.py +0 -228
- paddlex/ppdet/modeling/losses/solov2_loss.py +0 -101
- paddlex/ppdet/modeling/losses/ssd_loss.py +0 -163
- paddlex/ppdet/modeling/losses/yolo_loss.py +0 -212
- paddlex/ppdet/modeling/mot/__init__.py +0 -25
- paddlex/ppdet/modeling/mot/matching/__init__.py +0 -19
- paddlex/ppdet/modeling/mot/matching/deepsort_matching.py +0 -382
- paddlex/ppdet/modeling/mot/matching/jde_matching.py +0 -145
- paddlex/ppdet/modeling/mot/motion/__init__.py +0 -17
- paddlex/ppdet/modeling/mot/motion/kalman_filter.py +0 -270
- paddlex/ppdet/modeling/mot/tracker/__init__.py +0 -23
- paddlex/ppdet/modeling/mot/tracker/base_jde_tracker.py +0 -267
- paddlex/ppdet/modeling/mot/tracker/base_sde_tracker.py +0 -145
- paddlex/ppdet/modeling/mot/tracker/deepsort_tracker.py +0 -165
- paddlex/ppdet/modeling/mot/tracker/jde_tracker.py +0 -262
- paddlex/ppdet/modeling/mot/utils.py +0 -181
- paddlex/ppdet/modeling/mot/visualization.py +0 -130
- paddlex/ppdet/modeling/necks/__init__.py +0 -25
- paddlex/ppdet/modeling/necks/centernet_fpn.py +0 -185
- paddlex/ppdet/modeling/necks/fpn.py +0 -233
- paddlex/ppdet/modeling/necks/hrfpn.py +0 -131
- paddlex/ppdet/modeling/necks/ttf_fpn.py +0 -243
- paddlex/ppdet/modeling/necks/yolo_fpn.py +0 -1034
- paddlex/ppdet/modeling/ops.py +0 -1599
- paddlex/ppdet/modeling/post_process.py +0 -449
- paddlex/ppdet/modeling/proposal_generator/__init__.py +0 -2
- paddlex/ppdet/modeling/proposal_generator/anchor_generator.py +0 -135
- paddlex/ppdet/modeling/proposal_generator/proposal_generator.py +0 -81
- paddlex/ppdet/modeling/proposal_generator/rpn_head.py +0 -269
- paddlex/ppdet/modeling/proposal_generator/target.py +0 -671
- paddlex/ppdet/modeling/proposal_generator/target_layer.py +0 -476
- paddlex/ppdet/modeling/reid/__init__.py +0 -23
- paddlex/ppdet/modeling/reid/fairmot_embedding_head.py +0 -117
- paddlex/ppdet/modeling/reid/jde_embedding_head.py +0 -189
- paddlex/ppdet/modeling/reid/pyramidal_embedding.py +0 -151
- paddlex/ppdet/modeling/reid/resnet.py +0 -320
- paddlex/ppdet/modeling/shape_spec.py +0 -33
- paddlex/ppdet/modeling/tests/__init__.py +0 -13
- paddlex/ppdet/modeling/tests/test_architectures.py +0 -59
- paddlex/ppdet/modeling/tests/test_base.py +0 -75
- paddlex/ppdet/modeling/tests/test_ops.py +0 -839
- paddlex/ppdet/modeling/tests/test_yolov3_loss.py +0 -420
- paddlex/ppdet/optimizer.py +0 -285
- paddlex/ppdet/slim/__init__.py +0 -62
- paddlex/ppdet/slim/distill.py +0 -111
- paddlex/ppdet/slim/prune.py +0 -85
- paddlex/ppdet/slim/quant.py +0 -52
- paddlex/ppdet/utils/__init__.py +0 -13
- paddlex/ppdet/utils/check.py +0 -93
- paddlex/ppdet/utils/checkpoint.py +0 -216
- paddlex/ppdet/utils/cli.py +0 -151
- paddlex/ppdet/utils/colormap.py +0 -56
- paddlex/ppdet/utils/download.py +0 -477
- paddlex/ppdet/utils/logger.py +0 -71
- paddlex/ppdet/utils/stats.py +0 -95
- paddlex/ppdet/utils/visualizer.py +0 -292
- paddlex/ppdet/utils/voc_utils.py +0 -87
- paddlex/seg.py +0 -38
- paddlex/tools/__init__.py +0 -16
- paddlex/tools/convert.py +0 -52
- paddlex/tools/dataset_conversion/__init__.py +0 -24
- paddlex/tools/dataset_conversion/x2coco.py +0 -379
- paddlex/tools/dataset_conversion/x2imagenet.py +0 -82
- paddlex/tools/dataset_conversion/x2seg.py +0 -343
- paddlex/tools/dataset_conversion/x2voc.py +0 -230
- paddlex/tools/dataset_split/__init__.py +0 -23
- paddlex/tools/dataset_split/coco_split.py +0 -69
- paddlex/tools/dataset_split/imagenet_split.py +0 -75
- paddlex/tools/dataset_split/seg_split.py +0 -96
- paddlex/tools/dataset_split/utils.py +0 -75
- paddlex/tools/dataset_split/voc_split.py +0 -91
- paddlex/tools/split.py +0 -41
- paddlex/utils/checkpoint.py +0 -439
- paddlex/utils/shm.py +0 -67
- paddlex/utils/stats.py +0 -68
- paddlex/utils/utils.py +0 -140
- paddlex-2.0.0rc4.dist-info/LICENSE +0 -201
- paddlex-2.0.0rc4.dist-info/METADATA +0 -29
- paddlex-2.0.0rc4.dist-info/RECORD +0 -445
- paddlex-2.0.0rc4.dist-info/WHEEL +0 -5
- paddlex-2.0.0rc4.dist-info/entry_points.txt +0 -3
- paddlex-2.0.0rc4.dist-info/top_level.txt +0 -2
@@ -0,0 +1,1612 @@
|
|
1
|
+
# Copyright (c) 2024 PaddlePaddle Authors. All Rights Reserved.
|
2
|
+
#
|
3
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
4
|
+
# you may not use this file except in compliance with the License.
|
5
|
+
# You may obtain a copy of the License at
|
6
|
+
#
|
7
|
+
# http://www.apache.org/licenses/LICENSE-2.0
|
8
|
+
#
|
9
|
+
# Unless required by applicable law or agreed to in writing, software
|
10
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
11
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
12
|
+
# See the License for the specific language governing permissions and
|
13
|
+
# limitations under the License.
|
14
|
+
|
15
|
+
import functools
|
16
|
+
from collections import OrderedDict
|
17
|
+
from dataclasses import dataclass, fields
|
18
|
+
from typing import Any, Optional, Tuple
|
19
|
+
|
20
|
+
import numpy as np
|
21
|
+
import paddle
|
22
|
+
from paddle import Tensor
|
23
|
+
from paddle.distributed.fleet.utils import recompute
|
24
|
+
from paddle.nn import MultiHeadAttention
|
25
|
+
from paddle.nn.layer.transformer import _convert_attention_mask
|
26
|
+
|
27
|
+
from ...tokenizer.tokenizer_utils import adapt_stale_fwd_patch
|
28
|
+
|
29
|
+
|
30
|
+
def tuple_output(outputs: Tuple[Tensor], loss: Optional[Tensor] = None):
|
31
|
+
"""re-construct the outputs with one method which contains the simple logic
|
32
|
+
|
33
|
+
Args:
|
34
|
+
outputs (Tuple[Tensor]): the source of the outputs
|
35
|
+
loss (Optional[Tensor], optional): the loss of the model. Defaults to None.
|
36
|
+
"""
|
37
|
+
if loss is not None:
|
38
|
+
outputs = (loss,) + outputs
|
39
|
+
if len(outputs) == 1:
|
40
|
+
return outputs[0]
|
41
|
+
return outputs
|
42
|
+
|
43
|
+
|
44
|
+
def convert_encoder_output(encoder_output):
|
45
|
+
"""
|
46
|
+
Convert encoder_output from tuple to class:`~paddlenlp.transformers.model_outputs.BaseModelOutput`.
|
47
|
+
|
48
|
+
Args:
|
49
|
+
encoder_output (tuple or ModelOutput):
|
50
|
+
The output of the encoder, a tuple consists `last_hidden_state`, `hidden_states`(optional), `attentions`(optional).
|
51
|
+
The data type of `last_hidden_state` is float32 and its shape is [batch_size, sequence_length, hidden_size].
|
52
|
+
"""
|
53
|
+
return BaseModelOutput(
|
54
|
+
last_hidden_state=encoder_output[0],
|
55
|
+
hidden_states=encoder_output[1] if len(encoder_output) > 1 else None,
|
56
|
+
attentions=encoder_output[2] if len(encoder_output) > 2 else None,
|
57
|
+
)
|
58
|
+
|
59
|
+
|
60
|
+
def layer_init_wrapper(func):
|
61
|
+
@functools.wraps(func)
|
62
|
+
def _impl(self, *args, **kwargs):
|
63
|
+
enable_recompute = kwargs.pop("enable_recompute", False)
|
64
|
+
func(self, *args, **kwargs)
|
65
|
+
if paddle.in_dynamic_mode():
|
66
|
+
self.enable_recompute = enable_recompute
|
67
|
+
else:
|
68
|
+
self.enable_recompute = False
|
69
|
+
|
70
|
+
return _impl
|
71
|
+
|
72
|
+
|
73
|
+
@paddle.jit.not_to_static
|
74
|
+
def _transformer_encoder_layer_fwd(
|
75
|
+
self, src, src_mask=None, cache=None, output_attentions=False
|
76
|
+
):
|
77
|
+
self.self_attn.need_weights = output_attentions
|
78
|
+
src_mask = _convert_attention_mask(src_mask, src.dtype)
|
79
|
+
|
80
|
+
residual = src
|
81
|
+
if self.normalize_before:
|
82
|
+
src = self.norm1(src)
|
83
|
+
|
84
|
+
attn_outputs = self.self_attn(src, src, src, src_mask, cache)
|
85
|
+
if isinstance(attn_outputs, tuple):
|
86
|
+
src = attn_outputs[0]
|
87
|
+
outputs = attn_outputs[1:]
|
88
|
+
else:
|
89
|
+
src = attn_outputs
|
90
|
+
outputs = None
|
91
|
+
|
92
|
+
src = residual + self.dropout1(src)
|
93
|
+
if not self.normalize_before:
|
94
|
+
src = self.norm1(src)
|
95
|
+
|
96
|
+
residual = src
|
97
|
+
if self.normalize_before:
|
98
|
+
src = self.norm2(src)
|
99
|
+
src = self.linear2(self.dropout(self.activation(self.linear1(src))))
|
100
|
+
src = residual + self.dropout2(src)
|
101
|
+
if not self.normalize_before:
|
102
|
+
src = self.norm2(src)
|
103
|
+
|
104
|
+
return (
|
105
|
+
src if outputs is None else ((src,) + outputs[::-1])
|
106
|
+
) # hidden_states, cache, attentions
|
107
|
+
|
108
|
+
|
109
|
+
@paddle.jit.not_to_static
|
110
|
+
def _transformer_decoder_layer_fwd(
|
111
|
+
self,
|
112
|
+
tgt,
|
113
|
+
memory,
|
114
|
+
tgt_mask=None,
|
115
|
+
memory_mask=None,
|
116
|
+
cache=None,
|
117
|
+
output_attentions=False,
|
118
|
+
):
|
119
|
+
residual = tgt
|
120
|
+
|
121
|
+
# self attention
|
122
|
+
self.self_attn.need_weights = output_attentions
|
123
|
+
tgt_mask = _convert_attention_mask(tgt_mask, tgt.dtype)
|
124
|
+
|
125
|
+
if self.normalize_before:
|
126
|
+
tgt = self.norm1(tgt)
|
127
|
+
|
128
|
+
self_attn_outputs = self.self_attn(
|
129
|
+
tgt, tgt, tgt, tgt_mask, cache[0] if cache else None
|
130
|
+
)
|
131
|
+
# self_attn_outputs = (tgt, attn_weights, incremental_cache) or only tgt
|
132
|
+
if isinstance(self_attn_outputs, type(tgt)):
|
133
|
+
tgt = self_attn_outputs
|
134
|
+
else:
|
135
|
+
tgt = self_attn_outputs[0]
|
136
|
+
if output_attentions:
|
137
|
+
self_attn_weights = self_attn_outputs[1]
|
138
|
+
if cache:
|
139
|
+
incremental_cache = self_attn_outputs[-1]
|
140
|
+
|
141
|
+
tgt = residual + self.dropout1(tgt)
|
142
|
+
if not self.normalize_before:
|
143
|
+
tgt = self.norm1(tgt)
|
144
|
+
|
145
|
+
residual = tgt
|
146
|
+
|
147
|
+
# cross attention
|
148
|
+
if memory is not None:
|
149
|
+
self.cross_attn.need_weights = output_attentions
|
150
|
+
memory_mask = _convert_attention_mask(memory_mask, memory.dtype)
|
151
|
+
|
152
|
+
if self.normalize_before:
|
153
|
+
tgt = self.norm2(tgt)
|
154
|
+
|
155
|
+
cross_attn_outputs = self.cross_attn(
|
156
|
+
tgt, memory, memory, memory_mask, cache[1] if cache else None
|
157
|
+
)
|
158
|
+
if isinstance(cross_attn_outputs, type(tgt)):
|
159
|
+
tgt = cross_attn_outputs
|
160
|
+
else:
|
161
|
+
tgt = cross_attn_outputs[0]
|
162
|
+
if output_attentions:
|
163
|
+
cross_attn_weights = cross_attn_outputs[1]
|
164
|
+
if cache:
|
165
|
+
static_cache = cross_attn_outputs[-1]
|
166
|
+
|
167
|
+
tgt = residual + self.dropout2(tgt)
|
168
|
+
if not self.normalize_before:
|
169
|
+
tgt = self.norm2(tgt)
|
170
|
+
|
171
|
+
residual = tgt
|
172
|
+
|
173
|
+
if self.normalize_before:
|
174
|
+
tgt = self.norm3(tgt)
|
175
|
+
tgt = self.linear2(self.dropout(self.activation(self.linear1(tgt))))
|
176
|
+
tgt = residual + self.dropout3(tgt)
|
177
|
+
if not self.normalize_before:
|
178
|
+
tgt = self.norm3(tgt)
|
179
|
+
|
180
|
+
if not output_attentions and cache is None:
|
181
|
+
return tgt
|
182
|
+
else:
|
183
|
+
outputs = (tgt,)
|
184
|
+
if output_attentions:
|
185
|
+
outputs += (
|
186
|
+
self_attn_weights,
|
187
|
+
cross_attn_weights if memory is not None else None,
|
188
|
+
)
|
189
|
+
if cache:
|
190
|
+
outputs += (
|
191
|
+
(incremental_cache, static_cache if memory is not None else None),
|
192
|
+
)
|
193
|
+
return outputs
|
194
|
+
|
195
|
+
|
196
|
+
@paddle.jit.not_to_static
|
197
|
+
def _transformer_decoder_fwd(
|
198
|
+
self,
|
199
|
+
tgt,
|
200
|
+
memory=None,
|
201
|
+
tgt_mask=None,
|
202
|
+
memory_mask=None,
|
203
|
+
cache=None,
|
204
|
+
output_attentions=False,
|
205
|
+
output_hidden_states=False,
|
206
|
+
return_dict=False,
|
207
|
+
):
|
208
|
+
tgt_mask = _convert_attention_mask(tgt_mask, tgt.dtype)
|
209
|
+
if memory is not None:
|
210
|
+
memory_mask = _convert_attention_mask(memory_mask, memory.dtype)
|
211
|
+
|
212
|
+
new_caches = [] if cache else None
|
213
|
+
all_hidden_states = [tgt] if output_hidden_states else None
|
214
|
+
all_self_attns = [] if output_attentions else None
|
215
|
+
all_cross_attns = [] if output_attentions else None
|
216
|
+
|
217
|
+
for i, mod in enumerate(self.layers):
|
218
|
+
if cache is None:
|
219
|
+
# if output has no gradient, recompute is unnecessary
|
220
|
+
memory_stop_gradient = memory is not None and memory.stop_gradient
|
221
|
+
has_gradient = (not tgt.stop_gradient) or (not memory_stop_gradient)
|
222
|
+
if self.enable_recompute and has_gradient:
|
223
|
+
outputs = recompute(
|
224
|
+
mod, tgt, memory, tgt_mask, memory_mask, None, output_attentions
|
225
|
+
)
|
226
|
+
else:
|
227
|
+
outputs = mod(
|
228
|
+
tgt,
|
229
|
+
memory,
|
230
|
+
tgt_mask=tgt_mask,
|
231
|
+
memory_mask=memory_mask,
|
232
|
+
cache=None,
|
233
|
+
output_attentions=output_attentions,
|
234
|
+
)
|
235
|
+
else:
|
236
|
+
outputs = mod(
|
237
|
+
tgt,
|
238
|
+
memory,
|
239
|
+
tgt_mask=tgt_mask,
|
240
|
+
memory_mask=memory_mask,
|
241
|
+
cache=cache[i] if cache else None,
|
242
|
+
output_attentions=output_attentions,
|
243
|
+
)
|
244
|
+
if isinstance(outputs, type(tgt)):
|
245
|
+
tgt = outputs
|
246
|
+
else:
|
247
|
+
tgt = outputs[0]
|
248
|
+
if cache:
|
249
|
+
new_caches.append(outputs[-1])
|
250
|
+
if output_attentions:
|
251
|
+
all_self_attns.append(outputs[1])
|
252
|
+
all_cross_attns.append(outputs[2])
|
253
|
+
if output_hidden_states:
|
254
|
+
all_hidden_states.append(tgt)
|
255
|
+
|
256
|
+
if self.norm is not None:
|
257
|
+
tgt = self.norm(tgt)
|
258
|
+
if output_hidden_states:
|
259
|
+
all_hidden_states[-1] = tgt
|
260
|
+
|
261
|
+
if not return_dict:
|
262
|
+
if isinstance(outputs, type(tgt)):
|
263
|
+
return tgt
|
264
|
+
|
265
|
+
temp_list = [
|
266
|
+
tgt,
|
267
|
+
new_caches if cache else None,
|
268
|
+
all_hidden_states,
|
269
|
+
all_self_attns,
|
270
|
+
all_cross_attns,
|
271
|
+
]
|
272
|
+
return tuple(v for v in temp_list if v is not None)
|
273
|
+
|
274
|
+
return BaseModelOutputWithPastAndCrossAttentions(
|
275
|
+
last_hidden_state=tgt,
|
276
|
+
past_key_values=new_caches,
|
277
|
+
hidden_states=all_hidden_states,
|
278
|
+
attentions=all_self_attns,
|
279
|
+
cross_attentions=all_cross_attns,
|
280
|
+
)
|
281
|
+
|
282
|
+
|
283
|
+
@paddle.jit.not_to_static
|
284
|
+
def _transformer_encoder_fwd(
|
285
|
+
self,
|
286
|
+
src,
|
287
|
+
src_mask=None,
|
288
|
+
cache=None,
|
289
|
+
output_attentions=False,
|
290
|
+
output_hidden_states=False,
|
291
|
+
return_dict=False,
|
292
|
+
):
|
293
|
+
src_mask = _convert_attention_mask(src_mask, src.dtype)
|
294
|
+
|
295
|
+
output = src
|
296
|
+
# To get cache from None when use_cache is True, which is compatible with HF
|
297
|
+
# while HF requires decoder. The implementation here uses cache update in the
|
298
|
+
# MultiHeadAttention not so efficiently, and maybe optimize it later.
|
299
|
+
if cache is None and getattr(self, "_use_cache", False):
|
300
|
+
cache = [tuple(self.layers[0].gen_cache(src))] * len(self.layers)
|
301
|
+
# To be compatible with `TransformerEncoder.forward`, `_use_cache` defaults
|
302
|
+
# to True when cache is not None.
|
303
|
+
new_caches = [] if cache is not None and getattr(self, "_use_cache", True) else None
|
304
|
+
all_attentions = [] if output_attentions else None
|
305
|
+
# NOTE: Also includes embedding output which is same as HF.
|
306
|
+
all_hidden_states = [output] if output_hidden_states else None
|
307
|
+
for i, mod in enumerate(self.layers):
|
308
|
+
# if output has no gradient, recompute is unnecessary
|
309
|
+
has_gradient = not output.stop_gradient
|
310
|
+
if self.enable_recompute and has_gradient:
|
311
|
+
# Note: recompute do not support pass as **kwargs yet.
|
312
|
+
layer_outputs = recompute(
|
313
|
+
mod,
|
314
|
+
output,
|
315
|
+
src_mask,
|
316
|
+
(
|
317
|
+
None
|
318
|
+
if cache is None
|
319
|
+
else (
|
320
|
+
cache[i]
|
321
|
+
if isinstance(cache[i], MultiHeadAttention.Cache)
|
322
|
+
else MultiHeadAttention.Cache(*cache[i])
|
323
|
+
)
|
324
|
+
),
|
325
|
+
output_attentions,
|
326
|
+
)
|
327
|
+
else:
|
328
|
+
layer_outputs = mod(
|
329
|
+
output,
|
330
|
+
src_mask=src_mask,
|
331
|
+
cache=(
|
332
|
+
None
|
333
|
+
if cache is None
|
334
|
+
else (
|
335
|
+
cache[i]
|
336
|
+
if isinstance(cache[i], MultiHeadAttention.Cache)
|
337
|
+
else MultiHeadAttention.Cache(*cache[i])
|
338
|
+
)
|
339
|
+
),
|
340
|
+
output_attentions=output_attentions,
|
341
|
+
)
|
342
|
+
|
343
|
+
if isinstance(layer_outputs, tuple):
|
344
|
+
output = layer_outputs[0]
|
345
|
+
outputs = layer_outputs[1:]
|
346
|
+
else:
|
347
|
+
output = layer_outputs
|
348
|
+
outputs = None
|
349
|
+
|
350
|
+
if output_hidden_states:
|
351
|
+
all_hidden_states.append(output)
|
352
|
+
if output_attentions:
|
353
|
+
all_attentions.append(outputs[-1])
|
354
|
+
if new_caches is not None:
|
355
|
+
new_caches.append(
|
356
|
+
outputs[0]
|
357
|
+
if isinstance(cache[i], MultiHeadAttention.Cache)
|
358
|
+
else (tuple(outputs[0]))
|
359
|
+
)
|
360
|
+
|
361
|
+
if self.norm is not None:
|
362
|
+
output = self.norm(output)
|
363
|
+
|
364
|
+
if output_hidden_states:
|
365
|
+
all_hidden_states[-1] = output
|
366
|
+
|
367
|
+
if not return_dict:
|
368
|
+
outputs = tuple(
|
369
|
+
tuple(v) if isinstance(v, list) else v
|
370
|
+
for v in [
|
371
|
+
output,
|
372
|
+
new_caches,
|
373
|
+
all_hidden_states,
|
374
|
+
all_attentions,
|
375
|
+
]
|
376
|
+
if v is not None
|
377
|
+
)
|
378
|
+
if len(outputs) == 1:
|
379
|
+
return output
|
380
|
+
else:
|
381
|
+
return outputs
|
382
|
+
|
383
|
+
return BaseModelOutputWithPastAndCrossAttentions(
|
384
|
+
last_hidden_state=output,
|
385
|
+
past_key_values=new_caches,
|
386
|
+
hidden_states=all_hidden_states,
|
387
|
+
attentions=all_attentions,
|
388
|
+
)
|
389
|
+
|
390
|
+
|
391
|
+
_transformer_encoder_fwd.__name__ = "forward"
|
392
|
+
_transformer_encoder_layer_fwd.__name__ = "forward"
|
393
|
+
# patches of paddle.nn.Transformer to get all hidden_states and attentions
|
394
|
+
paddle.nn.TransformerEncoderLayer.forward = _transformer_encoder_layer_fwd
|
395
|
+
paddle.nn.TransformerDecoderLayer.forward = _transformer_decoder_layer_fwd
|
396
|
+
paddle.nn.TransformerEncoder.forward = _transformer_encoder_fwd
|
397
|
+
paddle.nn.TransformerDecoder.forward = _transformer_decoder_fwd
|
398
|
+
|
399
|
+
_encoder_init = paddle.nn.TransformerEncoder.__init__
|
400
|
+
_decoder_init = paddle.nn.TransformerDecoder.__init__
|
401
|
+
paddle.nn.TransformerEncoder.__init__ = layer_init_wrapper(_encoder_init)
|
402
|
+
paddle.nn.TransformerDecoder.__init__ = layer_init_wrapper(_decoder_init)
|
403
|
+
|
404
|
+
|
405
|
+
def _get_wrap_setattr(cls):
|
406
|
+
def _wrap_setattr(self, name, value):
|
407
|
+
value = adapt_stale_fwd_patch(self, name, value)
|
408
|
+
return super(cls, self).__setattr__(name, value)
|
409
|
+
|
410
|
+
return _wrap_setattr
|
411
|
+
|
412
|
+
|
413
|
+
paddle.nn.TransformerEncoderLayer.__setattr__ = functools.wraps(
|
414
|
+
paddle.nn.TransformerEncoderLayer.__setattr__
|
415
|
+
)(_get_wrap_setattr(paddle.nn.TransformerEncoderLayer))
|
416
|
+
paddle.nn.TransformerEncoder.__setattr__ = functools.wraps(
|
417
|
+
paddle.nn.TransformerEncoder.__setattr__
|
418
|
+
)(_get_wrap_setattr(paddle.nn.TransformerEncoder))
|
419
|
+
paddle.nn.TransformerDecoder.__setattr__ = functools.wraps(
|
420
|
+
paddle.nn.TransformerDecoder.__setattr__
|
421
|
+
)(_get_wrap_setattr(paddle.nn.TransformerDecoder))
|
422
|
+
|
423
|
+
|
424
|
+
def is_tensor(x):
|
425
|
+
if isinstance(x, paddle.Tensor):
|
426
|
+
return True
|
427
|
+
|
428
|
+
return isinstance(x, np.ndarray)
|
429
|
+
|
430
|
+
|
431
|
+
class ModelOutput(OrderedDict):
|
432
|
+
"""
|
433
|
+
Base class for all model outputs as dataclass. Has a `__getitem__` that allows indexing by integer or slice (like a
|
434
|
+
tuple) or strings (like a dictionary) that will ignore the `None` attributes. Otherwise behaves like a regular
|
435
|
+
python dictionary.
|
436
|
+
|
437
|
+
<Tip warning={true}>
|
438
|
+
|
439
|
+
You can't unpack a `ModelOutput` directly. Use the [`~utils.ModelOutput.to_tuple`] method to convert it to a tuple
|
440
|
+
before.
|
441
|
+
|
442
|
+
</Tip>
|
443
|
+
"""
|
444
|
+
|
445
|
+
def __post_init__(self):
|
446
|
+
class_fields = fields(self)
|
447
|
+
|
448
|
+
# note(guosheng): Convert list to tuple automatically, and better to
|
449
|
+
# check if it is frozen.
|
450
|
+
# assert not getattr(self, dataclasses._PARAMS).frozen
|
451
|
+
for f in class_fields:
|
452
|
+
value = getattr(self, f.name)
|
453
|
+
if isinstance(value, list):
|
454
|
+
setattr(self, f.name, tuple(value))
|
455
|
+
|
456
|
+
# Safety and consistency checks
|
457
|
+
if not len(class_fields):
|
458
|
+
raise ValueError(f"{self.__class__.__name__} has no fields.")
|
459
|
+
if not all(field.default is None for field in class_fields[1:]):
|
460
|
+
raise ValueError(
|
461
|
+
f"{self.__class__.__name__} should not have more than one required field."
|
462
|
+
)
|
463
|
+
|
464
|
+
first_field = getattr(self, class_fields[0].name)
|
465
|
+
other_fields_are_none = all(
|
466
|
+
getattr(self, field.name) is None for field in class_fields[1:]
|
467
|
+
)
|
468
|
+
|
469
|
+
if other_fields_are_none and not is_tensor(first_field):
|
470
|
+
if isinstance(first_field, dict):
|
471
|
+
iterator = first_field.items()
|
472
|
+
first_field_iterator = True
|
473
|
+
else:
|
474
|
+
try:
|
475
|
+
iterator = iter(first_field)
|
476
|
+
first_field_iterator = True
|
477
|
+
except TypeError:
|
478
|
+
first_field_iterator = False
|
479
|
+
|
480
|
+
# if we provided an iterator as first field and the iterator is a (key, value) iterator
|
481
|
+
# set the associated fields
|
482
|
+
if first_field_iterator:
|
483
|
+
for element in iterator:
|
484
|
+
if (
|
485
|
+
not isinstance(element, (list, tuple))
|
486
|
+
or not len(element) == 2
|
487
|
+
or not isinstance(element[0], str)
|
488
|
+
):
|
489
|
+
break
|
490
|
+
setattr(self, element[0], element[1])
|
491
|
+
if element[1] is not None:
|
492
|
+
self[element[0]] = element[1]
|
493
|
+
elif first_field is not None:
|
494
|
+
self[class_fields[0].name] = first_field
|
495
|
+
else:
|
496
|
+
for field in class_fields:
|
497
|
+
v = getattr(self, field.name)
|
498
|
+
if v is not None:
|
499
|
+
self[field.name] = v
|
500
|
+
|
501
|
+
def __delitem__(self, *args, **kwargs):
|
502
|
+
raise Exception(
|
503
|
+
f"You cannot use ``__delitem__`` on a {self.__class__.__name__} instance."
|
504
|
+
)
|
505
|
+
|
506
|
+
def setdefault(self, *args, **kwargs):
|
507
|
+
raise Exception(
|
508
|
+
f"You cannot use ``setdefault`` on a {self.__class__.__name__} instance."
|
509
|
+
)
|
510
|
+
|
511
|
+
def pop(self, *args, **kwargs):
|
512
|
+
raise Exception(
|
513
|
+
f"You cannot use ``pop`` on a {self.__class__.__name__} instance."
|
514
|
+
)
|
515
|
+
|
516
|
+
def update(self, *args, **kwargs):
|
517
|
+
raise Exception(
|
518
|
+
f"You cannot use ``update`` on a {self.__class__.__name__} instance."
|
519
|
+
)
|
520
|
+
|
521
|
+
def __getitem__(self, k):
|
522
|
+
if isinstance(k, str):
|
523
|
+
inner_dict = {k: v for (k, v) in self.items()}
|
524
|
+
return inner_dict[k]
|
525
|
+
else:
|
526
|
+
return self.to_tuple()[k]
|
527
|
+
|
528
|
+
def __setattr__(self, name, value):
|
529
|
+
if name in self.keys() and value is not None:
|
530
|
+
# Don't call self.__setitem__ to avoid recursion errors
|
531
|
+
super().__setitem__(name, value)
|
532
|
+
super().__setattr__(name, value)
|
533
|
+
|
534
|
+
def __setitem__(self, key, value):
|
535
|
+
# Will raise a KeyException if needed
|
536
|
+
super().__setitem__(key, value)
|
537
|
+
# Don't call self.__setattr__ to avoid recursion errors
|
538
|
+
super().__setattr__(key, value)
|
539
|
+
|
540
|
+
def to_tuple(self) -> Tuple[Any]:
|
541
|
+
"""
|
542
|
+
Convert self to a tuple containing all the attributes/keys that are not `None`.
|
543
|
+
"""
|
544
|
+
# try to fix: https://github.com/PaddlePaddle/PaddleNLP/issues/3355
|
545
|
+
# when trying to get the keys of `OrderedDict`, `keys` method return empty values.
|
546
|
+
# TODO(wj-Mcat): this bug should be fixed in Paddle framework
|
547
|
+
tuples = ()
|
548
|
+
for field in fields(self):
|
549
|
+
if getattr(self, field.name, None) is None:
|
550
|
+
continue
|
551
|
+
tuples = tuples + (getattr(self, field.name),)
|
552
|
+
|
553
|
+
return tuples
|
554
|
+
|
555
|
+
|
556
|
+
@dataclass
|
557
|
+
class BaseModelOutput(ModelOutput):
|
558
|
+
"""
|
559
|
+
Base class for model's outputs, with potential hidden states and attentions.
|
560
|
+
|
561
|
+
Args:
|
562
|
+
last_hidden_state (`paddle.Tensor` of shape `(batch_size, sequence_length, hidden_size)`):
|
563
|
+
Sequence of hidden-states at the output of the last layer of the model.
|
564
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
565
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
566
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
567
|
+
|
568
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
569
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
570
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
571
|
+
sequence_length)`.
|
572
|
+
|
573
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
574
|
+
heads.
|
575
|
+
"""
|
576
|
+
|
577
|
+
last_hidden_state: paddle.Tensor = None
|
578
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
579
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
580
|
+
|
581
|
+
|
582
|
+
@dataclass
|
583
|
+
class BaseModelOutputWithNoAttention(ModelOutput):
|
584
|
+
"""
|
585
|
+
Base class for model's outputs, with potential hidden states.
|
586
|
+
|
587
|
+
Args:
|
588
|
+
last_hidden_state (`paddle.Tensor` of shape `(batch_size, num_channels, height, width)`):
|
589
|
+
Sequence of hidden-states at the output of the last layer of the model.
|
590
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
591
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
592
|
+
one for the output of each layer) of shape `(batch_size, num_channels, height, width)`.
|
593
|
+
|
594
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
595
|
+
"""
|
596
|
+
|
597
|
+
last_hidden_state: paddle.Tensor = None
|
598
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
599
|
+
|
600
|
+
|
601
|
+
@dataclass
|
602
|
+
class BaseModelOutputWithPooling(ModelOutput):
|
603
|
+
"""
|
604
|
+
Base class for model's outputs that also contains a pooling of the last hidden states.
|
605
|
+
|
606
|
+
Args:
|
607
|
+
last_hidden_state (`paddle.Tensor` of shape `(batch_size, sequence_length, hidden_size)`):
|
608
|
+
Sequence of hidden-states at the output of the last layer of the model.
|
609
|
+
pooler_output (`paddle.Tensor` of shape `(batch_size, hidden_size)`):
|
610
|
+
Last layer hidden-state of the first token of the sequence (classification token) after further processing
|
611
|
+
through the layers used for the auxiliary pretraining task. E.g. for BERT-family of models, this returns
|
612
|
+
the classification token after processing through a linear layer and a tanh activation function. The linear
|
613
|
+
layer weights are trained from the next sentence prediction (classification) objective during pretraining.
|
614
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
615
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
616
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
617
|
+
|
618
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
619
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
620
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
621
|
+
sequence_length)`.
|
622
|
+
|
623
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
624
|
+
heads.
|
625
|
+
"""
|
626
|
+
|
627
|
+
last_hidden_state: paddle.Tensor = None
|
628
|
+
pooler_output: paddle.Tensor = None
|
629
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
630
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
631
|
+
|
632
|
+
|
633
|
+
@dataclass
|
634
|
+
class BaseModelOutputWithPast(ModelOutput):
|
635
|
+
"""
|
636
|
+
Base class for model's outputs that may also contain a past key/values (to speed up sequential decoding).
|
637
|
+
|
638
|
+
Args:
|
639
|
+
last_hidden_state (`paddle.Tensor` of shape `(batch_size, sequence_length, hidden_size)`):
|
640
|
+
Sequence of hidden-states at the output of the last layer of the model.
|
641
|
+
|
642
|
+
If `past_key_values` is used only the last hidden-state of the sequences of shape `(batch_size, 1,
|
643
|
+
hidden_size)` is output.
|
644
|
+
past_key_values (`tuple(tuple(paddle.Tensor))`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`):
|
645
|
+
Tuple of `tuple(paddle.Tensor)` of length `config.n_layers`, with each tuple having 2 tensors of shape
|
646
|
+
`(batch_size, num_heads, sequence_length, embed_size_per_head)`) and optionally if
|
647
|
+
`config.is_encoder_decoder=True` 2 additional tensors of shape `(batch_size, num_heads,
|
648
|
+
encoder_sequence_length, embed_size_per_head)`.
|
649
|
+
|
650
|
+
Contains pre-computed hidden-states (key and values in the self-attention blocks and optionally if
|
651
|
+
`config.is_encoder_decoder=True` in the cross-attention blocks) that can be used (see `past_key_values`
|
652
|
+
input) to speed up sequential decoding.
|
653
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
654
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
655
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
656
|
+
|
657
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
658
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
659
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
660
|
+
sequence_length)`.
|
661
|
+
|
662
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
663
|
+
heads.
|
664
|
+
"""
|
665
|
+
|
666
|
+
last_hidden_state: paddle.Tensor = None
|
667
|
+
past_key_values: Optional[Tuple[Tuple[paddle.Tensor]]] = None
|
668
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
669
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
670
|
+
|
671
|
+
|
672
|
+
@dataclass
|
673
|
+
class BaseModelOutputWithPastAndCrossAttentions(ModelOutput):
|
674
|
+
"""
|
675
|
+
Base class for model's outputs that may also contain a past key/values (to speed up sequential decoding).
|
676
|
+
|
677
|
+
Args:
|
678
|
+
last_hidden_state (`paddle.Tensor` of shape `(batch_size, sequence_length, hidden_size)`):
|
679
|
+
Sequence of hidden-states at the output of the last layer of the model.
|
680
|
+
|
681
|
+
If `past_key_values` is used only the last hidden-state of the sequences of shape `(batch_size, 1,
|
682
|
+
hidden_size)` is output.
|
683
|
+
past_key_values (`tuple(tuple(paddle.Tensor))`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`):
|
684
|
+
Tuple of `tuple(paddle.Tensor)` of length `config.n_layers`, with each tuple having 2 tensors of shape
|
685
|
+
`(batch_size, num_heads, sequence_length, embed_size_per_head)`) and optionally if
|
686
|
+
`config.is_encoder_decoder=True` 2 additional tensors of shape `(batch_size, num_heads,
|
687
|
+
encoder_sequence_length, embed_size_per_head)`.
|
688
|
+
|
689
|
+
Contains pre-computed hidden-states (key and values in the self-attention blocks and optionally if
|
690
|
+
`config.is_encoder_decoder=True` in the cross-attention blocks) that can be used (see `past_key_values`
|
691
|
+
input) to speed up sequential decoding.
|
692
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
693
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
694
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
695
|
+
|
696
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
697
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
698
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
699
|
+
sequence_length)`.
|
700
|
+
|
701
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
702
|
+
heads.
|
703
|
+
cross_attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` and `config.add_cross_attention=True` is passed or when `config.output_attentions=True`):
|
704
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
705
|
+
sequence_length)`.
|
706
|
+
|
707
|
+
Attentions weights of the decoder's cross-attention layer, after the attention softmax, used to compute the
|
708
|
+
weighted average in the cross-attention heads.
|
709
|
+
cum_offsets (`tuple(paddle.Tensor)`, *optional*, needed when `return_full_hidden_states=True`:
|
710
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, 1)`.
|
711
|
+
|
712
|
+
Offset of the current batch.
|
713
|
+
"""
|
714
|
+
|
715
|
+
last_hidden_state: paddle.Tensor = None
|
716
|
+
past_key_values: Optional[Tuple[Tuple[paddle.Tensor]]] = None
|
717
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
718
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
719
|
+
cross_attentions: Optional[Tuple[paddle.Tensor]] = None
|
720
|
+
cum_offsets: Optional[Tuple[paddle.Tensor]] = None
|
721
|
+
|
722
|
+
|
723
|
+
@dataclass
|
724
|
+
class BaseModelOutputWithPastAndMTP(ModelOutput):
|
725
|
+
"""
|
726
|
+
Base class for model's outputs that may also contain a past key/values (to speed up sequential decoding).
|
727
|
+
|
728
|
+
Args:
|
729
|
+
last_hidden_state (`paddle.Tensor` of shape `(batch_size, sequence_length, hidden_size)`):
|
730
|
+
Sequence of hidden-states at the output of the last layer of the model.
|
731
|
+
|
732
|
+
If `past_key_values` is used only the last hidden-state of the sequences of shape `(batch_size, 1,
|
733
|
+
hidden_size)` is output.
|
734
|
+
past_key_values (`tuple(tuple(paddle.Tensor))`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`):
|
735
|
+
Tuple of `tuple(paddle.Tensor)` of length `config.n_layers`, with each tuple having 2 tensors of shape
|
736
|
+
`(batch_size, num_heads, sequence_length, embed_size_per_head)`) and optionally if
|
737
|
+
`config.is_encoder_decoder=True` 2 additional tensors of shape `(batch_size, num_heads,
|
738
|
+
encoder_sequence_length, embed_size_per_head)`.
|
739
|
+
|
740
|
+
Contains pre-computed hidden-states (key and values in the self-attention blocks and optionally if
|
741
|
+
`config.is_encoder_decoder=True` in the cross-attention blocks) that can be used (see `past_key_values`
|
742
|
+
input) to speed up sequential decoding.
|
743
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
744
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
745
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
746
|
+
|
747
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
748
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
749
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
750
|
+
sequence_length)`.
|
751
|
+
|
752
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
753
|
+
heads.
|
754
|
+
mtp_outputs (`tuple(paddle.Tensor)`, *optional*):
|
755
|
+
MTP Layers outputs, used to compute the mtp loss.
|
756
|
+
heads.
|
757
|
+
"""
|
758
|
+
|
759
|
+
last_hidden_state: paddle.Tensor = None
|
760
|
+
past_key_values: Optional[Tuple[Tuple[paddle.Tensor]]] = None
|
761
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
762
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
763
|
+
mtp_outputs: Optional[Tuple[paddle.Tensor]] = None
|
764
|
+
|
765
|
+
|
766
|
+
@dataclass
|
767
|
+
class BaseModelOutputWithPoolingAndCrossAttentions(ModelOutput):
|
768
|
+
"""
|
769
|
+
Base class for model's outputs that also contains a pooling of the last hidden states.
|
770
|
+
|
771
|
+
Args:
|
772
|
+
last_hidden_state (`paddle.Tensor` of shape `(batch_size, sequence_length, hidden_size)`):
|
773
|
+
Sequence of hidden-states at the output of the last layer of the model.
|
774
|
+
pooler_output (`paddle.Tensor` of shape `(batch_size, hidden_size)`):
|
775
|
+
Last layer hidden-state of the first token of the sequence (classification token) after further processing
|
776
|
+
through the layers used for the auxiliary pretraining task. E.g. for BERT-family of models, this returns
|
777
|
+
the classification token after processing through a linear layer and a tanh activation function. The linear
|
778
|
+
layer weights are trained from the next sentence prediction (classification) objective during pretraining.
|
779
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
780
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
781
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
782
|
+
|
783
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
784
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
785
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
786
|
+
sequence_length)`.
|
787
|
+
|
788
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
789
|
+
heads.
|
790
|
+
cross_attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` and `config.add_cross_attention=True` is passed or when `config.output_attentions=True`):
|
791
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
792
|
+
sequence_length)`.
|
793
|
+
|
794
|
+
Attentions weights of the decoder's cross-attention layer, after the attention softmax, used to compute the
|
795
|
+
weighted average in the cross-attention heads.
|
796
|
+
past_key_values (`tuple(tuple(paddle.Tensor))`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`):
|
797
|
+
Tuple of `tuple(paddle.Tensor)` of length `config.n_layers`, with each tuple having 2 tensors of shape
|
798
|
+
`(batch_size, num_heads, sequence_length, embed_size_per_head)`) and optionally if
|
799
|
+
`config.is_encoder_decoder=True` 2 additional tensors of shape `(batch_size, num_heads,
|
800
|
+
encoder_sequence_length, embed_size_per_head)`.
|
801
|
+
|
802
|
+
Contains pre-computed hidden-states (key and values in the self-attention blocks and optionally if
|
803
|
+
`config.is_encoder_decoder=True` in the cross-attention blocks) that can be used (see `past_key_values`
|
804
|
+
input) to speed up sequential decoding.
|
805
|
+
"""
|
806
|
+
|
807
|
+
last_hidden_state: paddle.Tensor = None
|
808
|
+
pooler_output: paddle.Tensor = None
|
809
|
+
past_key_values: Optional[Tuple[Tuple[paddle.Tensor]]] = None
|
810
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
811
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
812
|
+
cross_attentions: Optional[Tuple[paddle.Tensor]] = None
|
813
|
+
|
814
|
+
|
815
|
+
@dataclass
|
816
|
+
class SequenceClassifierOutput(ModelOutput):
|
817
|
+
"""
|
818
|
+
Base class for outputs of sentence classification models.
|
819
|
+
|
820
|
+
Args:
|
821
|
+
loss (`paddle.Tensor` of shape `(1,)`, *optional*, returned when `labels` is provided):
|
822
|
+
Classification (or regression if config.num_labels==1) loss.
|
823
|
+
logits (`paddle.Tensor` of shape `(batch_size, config.num_labels)`):
|
824
|
+
Classification (or regression if config.num_labels==1) scores (before SoftMax).
|
825
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
826
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
827
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
828
|
+
|
829
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
830
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
831
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
832
|
+
sequence_length)`.
|
833
|
+
|
834
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
835
|
+
heads.
|
836
|
+
"""
|
837
|
+
|
838
|
+
loss: Optional[paddle.Tensor] = None
|
839
|
+
logits: paddle.Tensor = None
|
840
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
841
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
842
|
+
|
843
|
+
|
844
|
+
@dataclass
|
845
|
+
class TokenClassifierOutput(ModelOutput):
|
846
|
+
"""
|
847
|
+
Base class for outputs of token classification models.
|
848
|
+
|
849
|
+
Args:
|
850
|
+
loss (`paddle.Tensor` of shape `(1,)`, *optional*, returned when `labels` is provided) :
|
851
|
+
Classification loss.
|
852
|
+
logits (`paddle.Tensor` of shape `(batch_size, sequence_length, config.num_labels)`):
|
853
|
+
Classification scores (before SoftMax).
|
854
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
855
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
856
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
857
|
+
|
858
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
859
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
860
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
861
|
+
sequence_length)`.
|
862
|
+
|
863
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
864
|
+
heads.
|
865
|
+
"""
|
866
|
+
|
867
|
+
loss: Optional[paddle.Tensor] = None
|
868
|
+
logits: paddle.Tensor = None
|
869
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
870
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
871
|
+
|
872
|
+
|
873
|
+
@dataclass
|
874
|
+
class QuestionAnsweringModelOutput(ModelOutput):
|
875
|
+
"""
|
876
|
+
Base class for outputs of question answering models.
|
877
|
+
|
878
|
+
Args:
|
879
|
+
loss (`paddle.Tensor` of shape `(1,)`, *optional*, returned when `labels` is provided):
|
880
|
+
Total span extraction loss is the sum of a Cross-Entropy for the start and end positions.
|
881
|
+
start_logits (`paddle.Tensor` of shape `(batch_size, sequence_length)`):
|
882
|
+
Span-start scores (before SoftMax).
|
883
|
+
end_logits (`paddle.Tensor` of shape `(batch_size, sequence_length)`):
|
884
|
+
Span-end scores (before SoftMax).
|
885
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
886
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
887
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
888
|
+
|
889
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
890
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
891
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
892
|
+
sequence_length)`.
|
893
|
+
|
894
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
895
|
+
heads.
|
896
|
+
"""
|
897
|
+
|
898
|
+
loss: Optional[paddle.Tensor] = None
|
899
|
+
start_logits: paddle.Tensor = None
|
900
|
+
end_logits: paddle.Tensor = None
|
901
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
902
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
903
|
+
|
904
|
+
|
905
|
+
@dataclass
|
906
|
+
class MultipleChoiceModelOutput(ModelOutput):
|
907
|
+
"""
|
908
|
+
Base class for outputs of multiple choice models.
|
909
|
+
|
910
|
+
Args:
|
911
|
+
loss (`paddle.Tensor` of shape *(1,)*, *optional*, returned when `labels` is provided):
|
912
|
+
Classification loss.
|
913
|
+
logits (`paddle.Tensor` of shape `(batch_size, num_choices)`):
|
914
|
+
*num_choices* is the second dimension of the input tensors. (see *input_ids* above).
|
915
|
+
|
916
|
+
Classification scores (before SoftMax).
|
917
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
918
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
919
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
920
|
+
|
921
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
922
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
923
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
924
|
+
sequence_length)`.
|
925
|
+
|
926
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
927
|
+
heads.
|
928
|
+
"""
|
929
|
+
|
930
|
+
loss: Optional[paddle.Tensor] = None
|
931
|
+
logits: paddle.Tensor = None
|
932
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
933
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
934
|
+
|
935
|
+
|
936
|
+
@dataclass
|
937
|
+
class MaskedLMOutput(ModelOutput):
|
938
|
+
"""
|
939
|
+
Base class for masked language models outputs.
|
940
|
+
|
941
|
+
Args:
|
942
|
+
loss (`paddle.Tensor` of shape `(1,)`, *optional*, returned when `labels` is provided):
|
943
|
+
Masked language modeling (MLM) loss.
|
944
|
+
logits (`paddle.Tensor` of shape `(batch_size, sequence_length, config.vocab_size)`):
|
945
|
+
Prediction scores of the language modeling head (scores for each vocabulary token before SoftMax).
|
946
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
947
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
948
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
949
|
+
|
950
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
951
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
952
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
953
|
+
sequence_length)`.
|
954
|
+
|
955
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
956
|
+
heads.
|
957
|
+
"""
|
958
|
+
|
959
|
+
loss: Optional[paddle.Tensor] = None
|
960
|
+
logits: paddle.Tensor = None
|
961
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
962
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
963
|
+
|
964
|
+
|
965
|
+
@dataclass
|
966
|
+
class CausalLMOutputWithPast(ModelOutput):
|
967
|
+
"""
|
968
|
+
Base class for causal language model (or autoregressive) outputs.
|
969
|
+
|
970
|
+
Args:
|
971
|
+
loss (`paddle.Tensor` of shape `(1,)`, *optional*, returned when `labels` is provided):
|
972
|
+
Language modeling loss (for next-token prediction).
|
973
|
+
logits (`paddle.Tensor` of shape `(batch_size, sequence_length, config.vocab_size)`):
|
974
|
+
Prediction scores of the language modeling head (scores for each vocabulary token before SoftMax).
|
975
|
+
past_key_values (`tuple(tuple(paddle.Tensor))`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`):
|
976
|
+
Tuple of `paddle.Tensor` tuples of length `config.n_layers`, with each tuple containing the cached key,
|
977
|
+
value states of the self-attention and the cross-attention layers if model is used in encoder-decoder
|
978
|
+
setting. Only relevant if `config.is_decoder = True`.
|
979
|
+
|
980
|
+
Contains pre-computed hidden-states (key and values in the attention blocks) that can be used (see
|
981
|
+
`past_key_values` input) to speed up sequential decoding.
|
982
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
983
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
984
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
985
|
+
|
986
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
987
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
988
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
989
|
+
sequence_length)`.
|
990
|
+
|
991
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
992
|
+
heads.
|
993
|
+
"""
|
994
|
+
|
995
|
+
loss: Optional[paddle.Tensor] = None
|
996
|
+
logits: paddle.Tensor = None
|
997
|
+
past_key_values: Optional[Tuple[Tuple[paddle.Tensor]]] = None
|
998
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
999
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
1000
|
+
|
1001
|
+
|
1002
|
+
@dataclass
|
1003
|
+
class CausalLMOutputWithCrossAttentions(ModelOutput):
|
1004
|
+
"""
|
1005
|
+
Base class for causal language model (or autoregressive) outputs.
|
1006
|
+
|
1007
|
+
Args:
|
1008
|
+
loss (`paddle.Tensor` of shape `(1,)`, *optional*, returned when `labels` is provided):
|
1009
|
+
Language modeling loss (for next-token prediction).
|
1010
|
+
logits (`paddle.Tensor` of shape `(batch_size, sequence_length, config.vocab_size)`):
|
1011
|
+
Prediction scores of the language modeling head (scores for each vocabulary token before SoftMax).
|
1012
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
1013
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1014
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
1015
|
+
|
1016
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
1017
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
1018
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1019
|
+
sequence_length)`.
|
1020
|
+
|
1021
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
1022
|
+
heads.
|
1023
|
+
cross_attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
1024
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1025
|
+
sequence_length)`.
|
1026
|
+
|
1027
|
+
Cross attentions weights after the attention softmax, used to compute the weighted average in the
|
1028
|
+
cross-attention heads.
|
1029
|
+
past_key_values (`tuple(tuple(paddle.Tensor))`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`):
|
1030
|
+
Tuple of `paddle.Tensor` tuples of length `config.n_layers`, with each tuple containing the cached key,
|
1031
|
+
value states of the self-attention and the cross-attention layers if model is used in encoder-decoder
|
1032
|
+
setting. Only relevant if `config.is_decoder = True`.
|
1033
|
+
|
1034
|
+
Contains pre-computed hidden-states (key and values in the attention blocks) that can be used (see
|
1035
|
+
`past_key_values` input) to speed up sequential decoding.
|
1036
|
+
"""
|
1037
|
+
|
1038
|
+
loss: Optional[paddle.Tensor] = None
|
1039
|
+
logits: paddle.Tensor = None
|
1040
|
+
past_key_values: Optional[Tuple[Tuple[paddle.Tensor]]] = None
|
1041
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1042
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
1043
|
+
cross_attentions: Optional[Tuple[paddle.Tensor]] = None
|
1044
|
+
|
1045
|
+
|
1046
|
+
@dataclass
|
1047
|
+
class Seq2SeqModelOutput(ModelOutput):
|
1048
|
+
"""
|
1049
|
+
Base class for model encoder's outputs that also contains : pre-computed hidden states that can speed up sequential
|
1050
|
+
decoding.
|
1051
|
+
|
1052
|
+
Args:
|
1053
|
+
last_hidden_state (`paddle.Tensor`):
|
1054
|
+
Sequence of hidden-states at the output of the last layer of the decoder of the model, whose shape is `(batch_size, Sequence_length, hidden_size)`.
|
1055
|
+
|
1056
|
+
If `past_key_values` is used only the last hidden-state of the sequences of shape `(batch_size, 1,
|
1057
|
+
hidden_size)` is output.
|
1058
|
+
past_key_values (`tuple(tuple(paddle.Tensor))`, optional):
|
1059
|
+
Tuple of `tuple(paddle.Tensor)` of length `config.n_layers`, with each tuple having 2 tensors of shape
|
1060
|
+
`(batch_size, num_heads, sequence_length, embed_size_per_head)`) and 2 additional tensors of shape
|
1061
|
+
`(batch_size, num_heads, encoder_sequence_length, embed_size_per_head)`.
|
1062
|
+
Returned when `use_cache=True` is passed or when `config.use_cache=True`.
|
1063
|
+
|
1064
|
+
Contains pre-computed hidden-states (key and values in the self-attention blocks and in the cross-attention
|
1065
|
+
blocks) that can be used (see `past_key_values` input) to speed up sequential decoding.
|
1066
|
+
decoder_hidden_states (`tuple(paddle.Tensor)`, optional):
|
1067
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1068
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
1069
|
+
Returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`.
|
1070
|
+
|
1071
|
+
Hidden-states of the decoder at the output of each layer plus the optional initial embedding outputs.
|
1072
|
+
decoder_attentions (`tuple(paddle.Tensor)`, optional):
|
1073
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1074
|
+
sequence_length)`.
|
1075
|
+
Returned when `output_attentions=True` is passed or when `config.output_attentions=True`.
|
1076
|
+
|
1077
|
+
Attentions weights of the decoder, after the attention softmax, used to compute the weighted average in the
|
1078
|
+
self-attention heads.
|
1079
|
+
cross_attentions (`tuple(paddle.Tensor)`, optional):
|
1080
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1081
|
+
sequence_length)`.
|
1082
|
+
Returned when `output_attentions=True` is passed or when `config.output_attentions=True`.
|
1083
|
+
|
1084
|
+
Attentions weights of the decoder's cross-attention layer, after the attention softmax, used to compute the
|
1085
|
+
weighted average in the cross-attention heads.
|
1086
|
+
encoder_last_hidden_state (`paddle.Tensor`, optional):
|
1087
|
+
Sequence of hidden-states at the output of the last layer of the encoder of the model whose shape is `(batch_size, sequence_length, hidden_size)`,
|
1088
|
+
encoder_hidden_states (`tuple(paddle.Tensor)`, optional):
|
1089
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1090
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
1091
|
+
Returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`.
|
1092
|
+
|
1093
|
+
Hidden-states of the encoder at the output of each layer plus the optional initial embedding outputs.
|
1094
|
+
encoder_attentions (`tuple(paddle.Tensor)`, optional):
|
1095
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1096
|
+
sequence_length)`.
|
1097
|
+
Returned when `output_attentions=True` is passed or when `config.output_attentions=True`.
|
1098
|
+
|
1099
|
+
Attentions weights of the encoder, after the attention softmax, used to compute the weighted average in the
|
1100
|
+
self-attention heads.
|
1101
|
+
"""
|
1102
|
+
|
1103
|
+
last_hidden_state: paddle.Tensor = None
|
1104
|
+
past_key_values: Optional[Tuple[Tuple[paddle.Tensor]]] = None
|
1105
|
+
decoder_hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1106
|
+
decoder_attentions: Optional[Tuple[paddle.Tensor]] = None
|
1107
|
+
cross_attentions: Optional[Tuple[paddle.Tensor]] = None
|
1108
|
+
encoder_last_hidden_state: Optional[paddle.Tensor] = None
|
1109
|
+
encoder_hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1110
|
+
encoder_attentions: Optional[Tuple[paddle.Tensor]] = None
|
1111
|
+
|
1112
|
+
|
1113
|
+
@dataclass
|
1114
|
+
class Seq2SeqLMOutput(ModelOutput):
|
1115
|
+
"""
|
1116
|
+
Base class for sequence-to-sequence language models outputs.
|
1117
|
+
|
1118
|
+
Args:
|
1119
|
+
loss (`paddle.Tensor`, optional):
|
1120
|
+
Language modeling loss whose shape is `(1,)`. Returned when `labels` is provided.
|
1121
|
+
logits (`paddle.Tensor`):
|
1122
|
+
Prediction scores of the language modeling head (scores for each vocabulary token before SoftMax) whose shape is `(batch_size, sequence_length, config.vocab_size)`).
|
1123
|
+
past_key_values (`tuple(tuple(paddle.Tensor))`, optional):
|
1124
|
+
Tuple of `tuple(paddle.Tensor)` of length `config.n_layers`, with each tuple having 2 tensors of shape
|
1125
|
+
`(batch_size, num_heads, sequence_length, embed_size_per_head)`) and 2 additional tensors of shape
|
1126
|
+
`(batch_size, num_heads, encoder_sequence_length, embed_size_per_head)`.
|
1127
|
+
Returned when `use_cache=True` is passed or when `config.use_cache=True`.
|
1128
|
+
|
1129
|
+
Contains pre-computed hidden-states (key and values in the self-attention blocks and in the cross-attention
|
1130
|
+
blocks) that can be used (see `past_key_values` input) to speed up sequential decoding.
|
1131
|
+
decoder_hidden_states (`tuple(paddle.Tensor)`, optional):
|
1132
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1133
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
1134
|
+
Returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`.
|
1135
|
+
|
1136
|
+
Hidden-states of the decoder at the output of each layer plus the initial embedding outputs.
|
1137
|
+
decoder_attentions (`tuple(paddle.Tensor)`, optional):
|
1138
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1139
|
+
sequence_length)`.
|
1140
|
+
Returned when `output_attentions=True` is passed or when `config.output_attentions=True`.
|
1141
|
+
|
1142
|
+
Attentions weights of the decoder, after the attention softmax, used to compute the weighted average in the
|
1143
|
+
self-attention heads.
|
1144
|
+
cross_attentions (`tuple(paddle.Tensor)`, optional):
|
1145
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1146
|
+
sequence_length)`.
|
1147
|
+
Returned when `output_attentions=True` is passed or when `config.output_attentions=True`.
|
1148
|
+
|
1149
|
+
Attentions weights of the decoder's cross-attention layer, after the attention softmax, used to compute the
|
1150
|
+
weighted average in the cross-attention heads.
|
1151
|
+
encoder_last_hidden_state (`paddle.Tensor`, optional):
|
1152
|
+
Sequence of hidden-states at the output of the last layer of the encoder of the model whose shape is `(batch_size, sequence_length, hidden_size)`.
|
1153
|
+
encoder_hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
1154
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1155
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
1156
|
+
|
1157
|
+
Hidden-states of the encoder at the output of each layer plus the initial embedding outputs.
|
1158
|
+
encoder_attentions (`tuple(paddle.Tensor)`, optional):
|
1159
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1160
|
+
sequence_length)`.
|
1161
|
+
Returned when `output_attentions=True` is passed or when `config.output_attentions=True`.
|
1162
|
+
|
1163
|
+
Attentions weights of the encoder, after the attention softmax, used to compute the weighted average in the
|
1164
|
+
self-attention heads.
|
1165
|
+
"""
|
1166
|
+
|
1167
|
+
loss: Optional[paddle.Tensor] = None
|
1168
|
+
logits: paddle.Tensor = None
|
1169
|
+
past_key_values: Optional[Tuple[Tuple[paddle.Tensor]]] = None
|
1170
|
+
decoder_hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1171
|
+
decoder_attentions: Optional[Tuple[paddle.Tensor]] = None
|
1172
|
+
cross_attentions: Optional[Tuple[paddle.Tensor]] = None
|
1173
|
+
encoder_last_hidden_state: Optional[paddle.Tensor] = None
|
1174
|
+
encoder_hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1175
|
+
encoder_attentions: Optional[Tuple[paddle.Tensor]] = None
|
1176
|
+
|
1177
|
+
|
1178
|
+
@dataclass
|
1179
|
+
class Seq2SeqQuestionAnsweringModelOutput(ModelOutput):
|
1180
|
+
"""
|
1181
|
+
Base class for outputs of sequence-to-sequence question answering models.
|
1182
|
+
Args:
|
1183
|
+
loss (`paddle.Tensor` ,optional):
|
1184
|
+
Total span extraction loss is the sum of a Cross-Entropy for the start and end positions.
|
1185
|
+
A Tensor of shape `(1,)`, returned when `labels` is provided.
|
1186
|
+
start_logits (`paddle.Tensor`):
|
1187
|
+
Span-start scores (before SoftMax). Tensor of shape `(batch_size, sequence_length)`).
|
1188
|
+
end_logits (`paddle.Tensor`):
|
1189
|
+
Span-end scores (before SoftMax). Tensor of shape `(batch_size, sequence_length)`).
|
1190
|
+
past_key_values (`tuple(tuple(paddle.Tensor))`, optional):
|
1191
|
+
Tuple of `tuple(paddle.Tensor)` of length `n_layers`, with each tuple having 2 tensors of shape
|
1192
|
+
`(batch_size, num_heads, sequence_length, embed_size_per_head)`) and 2 additional tensors of shape
|
1193
|
+
`(batch_size, num_heads, encoder_sequence_length, embed_size_per_head)`.
|
1194
|
+
Returned when `use_cache=True` is passed.
|
1195
|
+
Contains pre-computed hidden-states (key and values in the self-attention blocks and in the cross-attention
|
1196
|
+
blocks) that can be used (see `past_key_values` input) to speed up sequential decoding.
|
1197
|
+
decoder_hidden_states (`tuple(paddle.Tensor)`, optional):
|
1198
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1199
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
1200
|
+
Returned when `output_hidden_states=True` is passed.
|
1201
|
+
Hidden-states of the decoder at the output of each layer plus the initial embedding outputs.
|
1202
|
+
decoder_attentions (`tuple(paddle.Tensor)`, optional):
|
1203
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1204
|
+
sequence_length)`. Returned when `output_attentions=True` is passed.
|
1205
|
+
Attentions weights of the decoder, after the attention softmax, used to compute the weighted average in the
|
1206
|
+
self-attention heads.
|
1207
|
+
cross_attentions (`tuple(paddle.Tensor)`, optional):
|
1208
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1209
|
+
sequence_length)`. Returned when `output_attentions=True` is passed.
|
1210
|
+
Attentions weights of the decoder's cross-attention layer, after the attention softmax, used to compute the
|
1211
|
+
weighted average in the cross-attention heads.
|
1212
|
+
encoder_last_hidden_state (`paddle.Tensor` optional):
|
1213
|
+
Sequence of hidden-states at the output of the last layer of the encoder of the model.
|
1214
|
+
Tensor of shape `(batch_size, sequence_length, hidden_size)`.
|
1215
|
+
encoder_hidden_states (`tuple(paddle.Tensor)`, optional):
|
1216
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1217
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
1218
|
+
Returned when `output_hidden_states=True` is passed.
|
1219
|
+
Hidden-states of the encoder at the output of each layer plus the initial embedding outputs.
|
1220
|
+
encoder_attentions (`tuple(paddle.Tensor)`, optional):
|
1221
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1222
|
+
sequence_length)`. Returned when `output_attentions=True` is passed.
|
1223
|
+
Attentions weights of the encoder, after the attention softmax, used to compute the weighted average in the
|
1224
|
+
self-attention heads.
|
1225
|
+
"""
|
1226
|
+
|
1227
|
+
loss: Optional[paddle.Tensor] = None
|
1228
|
+
start_logits: paddle.Tensor = None
|
1229
|
+
end_logits: paddle.Tensor = None
|
1230
|
+
past_key_values: Optional[Tuple[Tuple[paddle.Tensor]]] = None
|
1231
|
+
decoder_hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1232
|
+
decoder_attentions: Optional[Tuple[paddle.Tensor]] = None
|
1233
|
+
cross_attentions: Optional[Tuple[paddle.Tensor]] = None
|
1234
|
+
encoder_last_hidden_state: Optional[paddle.Tensor] = None
|
1235
|
+
encoder_hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1236
|
+
encoder_attentions: Optional[Tuple[paddle.Tensor]] = None
|
1237
|
+
|
1238
|
+
|
1239
|
+
@dataclass
|
1240
|
+
class Seq2SeqSequenceClassifierOutput(ModelOutput):
|
1241
|
+
"""
|
1242
|
+
Base class for outputs of sequence-to-sequence sentence classification models.
|
1243
|
+
Args:
|
1244
|
+
loss (`paddle.Tensor` optional):
|
1245
|
+
Classification (or regression if config.num_labels==1) loss of shape `(1,)`. Returned when `label` is provided).
|
1246
|
+
logits (`paddle.Tensor`):
|
1247
|
+
Classification (or regression if config.num_labels==1) scores (before SoftMax) of shape `(batch_size, config.num_labels)`
|
1248
|
+
past_key_values (`tuple(tuple(paddle.Tensor))`, optional):
|
1249
|
+
Tuple of `tuple(paddle.Tensor)` of length `config.n_layers`, with each tuple having 2 tensors of shape
|
1250
|
+
`(batch_size, num_heads, sequence_length, embed_size_per_head)`) and 2 additional tensors of shape
|
1251
|
+
`(batch_size, num_heads, encoder_sequence_length, embed_size_per_head)`.
|
1252
|
+
Returned when `use_cache=True` is passed.
|
1253
|
+
Contains pre-computed hidden-states (key and values in the self-attention blocks and in the cross-attention
|
1254
|
+
blocks) that can be used (see `past_key_values` input) to speed up sequential decoding.
|
1255
|
+
decoder_hidden_states (`tuple(paddle.Tensor)`, optional):
|
1256
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1257
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
1258
|
+
Returned when `output_hidden_states=True` is passed.
|
1259
|
+
Hidden-states of the decoder at the output of each layer plus the initial embedding outputs.
|
1260
|
+
decoder_attentions (`tuple(paddle.Tensor)`, optional):
|
1261
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1262
|
+
sequence_length)`. Returned when `output_attentions=True` is passed.
|
1263
|
+
Attentions weights of the decoder, after the attention softmax, used to compute the weighted average in the
|
1264
|
+
self-attention heads.
|
1265
|
+
cross_attentions (`tuple(paddle.Tensor)`, optional):
|
1266
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1267
|
+
sequence_length)`. Returned when `output_attentions=True` is passed.
|
1268
|
+
Attentions weights of the decoder's cross-attention layer, after the attention softmax, used to compute the
|
1269
|
+
weighted average in the cross-attention heads.
|
1270
|
+
encoder_last_hidden_state (`paddle.Tensor`, optional):
|
1271
|
+
Sequence of hidden-states at the output of the last layer of the encoder of the model.
|
1272
|
+
Tensor of shape `(batch_size, sequence_length, hidden_size)`.
|
1273
|
+
encoder_hidden_states (`tuple(paddle.Tensor)`, optional):
|
1274
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1275
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
1276
|
+
Returned when `output_hidden_states=True` is passed.
|
1277
|
+
Hidden-states of the encoder at the output of each layer plus the initial embedding outputs.
|
1278
|
+
encoder_attentions (`tuple(paddle.Tensor)`, optional):
|
1279
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1280
|
+
sequence_length)`.
|
1281
|
+
Returned when `output_attentions=True` is passed.
|
1282
|
+
Attentions weights of the encoder, after the attention softmax, used to compute the weighted average in the
|
1283
|
+
self-attention heads.
|
1284
|
+
"""
|
1285
|
+
|
1286
|
+
loss: Optional[paddle.Tensor] = None
|
1287
|
+
logits: paddle.Tensor = None
|
1288
|
+
past_key_values: Optional[Tuple[Tuple[paddle.Tensor]]] = None
|
1289
|
+
decoder_hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1290
|
+
decoder_attentions: Optional[Tuple[paddle.Tensor]] = None
|
1291
|
+
cross_attentions: Optional[Tuple[paddle.Tensor]] = None
|
1292
|
+
encoder_last_hidden_state: Optional[paddle.Tensor] = None
|
1293
|
+
encoder_hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1294
|
+
encoder_attentions: Optional[Tuple[paddle.Tensor]] = None
|
1295
|
+
|
1296
|
+
|
1297
|
+
@dataclass
|
1298
|
+
class SequenceClassifierOutputWithPast(ModelOutput):
|
1299
|
+
"""
|
1300
|
+
Base class for outputs of sentence classification models.
|
1301
|
+
Args:
|
1302
|
+
loss (`paddle.Tensor`, optional):
|
1303
|
+
Classification (or regression if config.num_labels==1) loss whose shape is `(1,)`.
|
1304
|
+
Returned when `labels` is provided.
|
1305
|
+
logits (`paddle.Tensor`):
|
1306
|
+
Classification (or regression if config.num_labels==1) scores (before SoftMax)
|
1307
|
+
whose shape is `(batch_size, num_labels)`
|
1308
|
+
past_key_values (`tuple(tuple(paddle.Tensor))`, optional):
|
1309
|
+
Tuple of `tuple(paddle.Tensor)` of length `config.n_layers`, with each tuple having 2 tensors of shape
|
1310
|
+
`(batch_size, num_heads, sequence_length, embed_size_per_head)`)
|
1311
|
+
Returned when `use_cache=True` is passed or when `config.use_cache=True`).
|
1312
|
+
Contains pre-computed hidden-states (key and values in the self-attention blocks) that can be used (see
|
1313
|
+
`past_key_values` input) to speed up sequential decoding.
|
1314
|
+
hidden_states (`tuple(paddle.Tensor)`, optional):
|
1315
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1316
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
1317
|
+
Returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`).
|
1318
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
1319
|
+
attentions (`tuple(paddle.Tensor)`, optional):
|
1320
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1321
|
+
sequence_length)`. Returned when `output_attentions=True` is passed or when `config.output_attentions=True`).
|
1322
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
1323
|
+
heads.
|
1324
|
+
"""
|
1325
|
+
|
1326
|
+
loss: Optional[paddle.Tensor] = None
|
1327
|
+
logits: paddle.Tensor = None
|
1328
|
+
past_key_values: Optional[Tuple[Tuple[paddle.Tensor]]] = None
|
1329
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1330
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
1331
|
+
|
1332
|
+
|
1333
|
+
@dataclass
|
1334
|
+
class BackboneOutput(ModelOutput):
|
1335
|
+
"""
|
1336
|
+
Base class for outputs of backbones.
|
1337
|
+
|
1338
|
+
Args:
|
1339
|
+
feature_maps (`tuple(paddle.Tensor)` of shape `(batch_size, num_channels, height, width)`):
|
1340
|
+
Feature maps of the stages.
|
1341
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
1342
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings + one for the output of each layer) of
|
1343
|
+
shape `(batch_size, sequence_length, hidden_size)` or `(batch_size, num_channels, height, width)`,
|
1344
|
+
depending on the backbone.
|
1345
|
+
|
1346
|
+
Hidden-states of the model at the output of each stage plus the initial embedding outputs.
|
1347
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
1348
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1349
|
+
sequence_length)`. Only applicable if the backbone uses attention.
|
1350
|
+
|
1351
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
1352
|
+
heads.
|
1353
|
+
"""
|
1354
|
+
|
1355
|
+
feature_maps: Tuple[paddle.Tensor] = None
|
1356
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1357
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
1358
|
+
|
1359
|
+
|
1360
|
+
@dataclass
|
1361
|
+
class BaseModelOutputWithPoolingAndNoAttention(ModelOutput):
|
1362
|
+
"""
|
1363
|
+
Base class for model's outputs that also contains a pooling of the last hidden states.
|
1364
|
+
|
1365
|
+
Args:
|
1366
|
+
last_hidden_state (`paddle.Tensor` of shape `(batch_size, num_channels, height, width)`):
|
1367
|
+
Sequence of hidden-states at the output of the last layer of the model.
|
1368
|
+
pooler_output (`paddle.Tensor` of shape `(batch_size, hidden_size)`):
|
1369
|
+
Last layer hidden-state after a pooling operation on the spatial dimensions.
|
1370
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
1371
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1372
|
+
one for the output of each layer) of shape `(batch_size, num_channels, height, width)`.
|
1373
|
+
|
1374
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
1375
|
+
"""
|
1376
|
+
|
1377
|
+
last_hidden_state: paddle.Tensor = None
|
1378
|
+
pooler_output: paddle.Tensor = None
|
1379
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1380
|
+
|
1381
|
+
|
1382
|
+
@dataclass
|
1383
|
+
class ImageClassifierOutputWithNoAttention(ModelOutput):
|
1384
|
+
"""
|
1385
|
+
Base class for outputs of image classification models.
|
1386
|
+
|
1387
|
+
Args:
|
1388
|
+
loss (`paddle.Tensor` of shape `(1,)`, *optional*, returned when `labels` is provided):
|
1389
|
+
Classification (or regression if config.num_labels==1) loss.
|
1390
|
+
logits (`paddle.Tensor` of shape `(batch_size, config.num_labels)`):
|
1391
|
+
Classification (or regression if config.num_labels==1) scores (before SoftMax).
|
1392
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
1393
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1394
|
+
one for the output of each stage) of shape `(batch_size, num_channels, height, width)`. Hidden-states (also
|
1395
|
+
called feature maps) of the model at the output of each stage.
|
1396
|
+
"""
|
1397
|
+
|
1398
|
+
loss: Optional[paddle.Tensor] = None
|
1399
|
+
logits: paddle.Tensor = None
|
1400
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1401
|
+
|
1402
|
+
|
1403
|
+
@dataclass
|
1404
|
+
class DepthEstimatorOutput(ModelOutput):
|
1405
|
+
"""
|
1406
|
+
Base class for outputs of depth estimation models.
|
1407
|
+
|
1408
|
+
Args:
|
1409
|
+
loss (`paddle.Tensor` of shape `(1,)`, *optional*, returned when `labels` is provided):
|
1410
|
+
Classification (or regression if config.num_labels==1) loss.
|
1411
|
+
predicted_depth (`paddle.Tensor` of shape `(batch_size, height, width)`):
|
1412
|
+
Predicted depth for each pixel.
|
1413
|
+
|
1414
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
1415
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1416
|
+
one for the output of each layer) of shape `(batch_size, num_channels, height, width)`.
|
1417
|
+
|
1418
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
1419
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
1420
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, patch_size,
|
1421
|
+
sequence_length)`.
|
1422
|
+
|
1423
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
1424
|
+
heads.
|
1425
|
+
"""
|
1426
|
+
|
1427
|
+
loss: Optional[paddle.Tensor] = None
|
1428
|
+
predicted_depth: paddle.Tensor = None
|
1429
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1430
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
1431
|
+
|
1432
|
+
|
1433
|
+
@dataclass
|
1434
|
+
class SemanticSegmenterOutput(ModelOutput):
|
1435
|
+
"""
|
1436
|
+
Base class for outputs of semantic segmentation models.
|
1437
|
+
Args:
|
1438
|
+
loss (`paddle.Tensor` of shape `(1,)`, *optional*, returned when `labels` is provided):
|
1439
|
+
Classification (or regression if config.num_labels==1) loss.
|
1440
|
+
logits (`paddle.Tensor` of shape `(batch_size, config.num_labels, logits_height, logits_width)`):
|
1441
|
+
Classification scores for each pixel.
|
1442
|
+
<Tip warning={true}>
|
1443
|
+
The logits returned do not necessarily have the same size as the `pixel_values` passed as inputs. This is
|
1444
|
+
to avoid doing two interpolations and lose some quality when a user needs to resize the logits to the
|
1445
|
+
original image size as post-processing. You should always check your logits shape and resize as needed.
|
1446
|
+
</Tip>
|
1447
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
1448
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1449
|
+
one for the output of each layer) of shape `(batch_size, patch_size, hidden_size)`.
|
1450
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
1451
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
1452
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, patch_size,
|
1453
|
+
sequence_length)`.
|
1454
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
1455
|
+
heads.
|
1456
|
+
"""
|
1457
|
+
|
1458
|
+
loss: Optional[paddle.Tensor] = None
|
1459
|
+
logits: paddle.Tensor = None
|
1460
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1461
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
1462
|
+
|
1463
|
+
|
1464
|
+
@dataclass
|
1465
|
+
class Seq2SeqSpectrogramOutput(ModelOutput):
|
1466
|
+
"""
|
1467
|
+
Base class for sequence-to-sequence spectrogram outputs.
|
1468
|
+
|
1469
|
+
Args:
|
1470
|
+
loss (`paddle.Tensor` of shape `(1,)`, *optional*, returned when `labels` is provided):
|
1471
|
+
Spectrogram generation loss.
|
1472
|
+
spectrogram (`paddle.Tensor` of shape `(batch_size, sequence_length, num_bins)`):
|
1473
|
+
The predicted spectrogram.
|
1474
|
+
past_key_values (`tuple(tuple(paddle.Tensor))`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`):
|
1475
|
+
Tuple of `tuple(paddle.Tensor)` of length `config.n_layers`, with each tuple having 2 tensors of shape
|
1476
|
+
`(batch_size, num_heads, sequence_length, embed_size_per_head)`) and 2 additional tensors of shape
|
1477
|
+
`(batch_size, num_heads, encoder_sequence_length, embed_size_per_head)`.
|
1478
|
+
|
1479
|
+
Contains pre-computed hidden-states (key and values in the self-attention blocks and in the cross-attention
|
1480
|
+
blocks) that can be used (see `past_key_values` input) to speed up sequential decoding.
|
1481
|
+
decoder_hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
1482
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1483
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
1484
|
+
|
1485
|
+
Hidden-states of the decoder at the output of each layer plus the initial embedding outputs.
|
1486
|
+
decoder_attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
1487
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1488
|
+
sequence_length)`.
|
1489
|
+
|
1490
|
+
Attentions weights of the decoder, after the attention softmax, used to compute the weighted average in the
|
1491
|
+
self-attention heads.
|
1492
|
+
cross_attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
1493
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1494
|
+
sequence_length)`.
|
1495
|
+
|
1496
|
+
Attentions weights of the decoder's cross-attention layer, after the attention softmax, used to compute the
|
1497
|
+
weighted average in the cross-attention heads.
|
1498
|
+
encoder_last_hidden_state (`paddle.Tensor` of shape `(batch_size, sequence_length, hidden_size)`, *optional*):
|
1499
|
+
Sequence of hidden-states at the output of the last layer of the encoder of the model.
|
1500
|
+
encoder_hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
1501
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1502
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
1503
|
+
|
1504
|
+
Hidden-states of the encoder at the output of each layer plus the initial embedding outputs.
|
1505
|
+
encoder_attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
1506
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1507
|
+
sequence_length)`.
|
1508
|
+
|
1509
|
+
Attentions weights of the encoder, after the attention softmax, used to compute the weighted average in the
|
1510
|
+
self-attention heads.
|
1511
|
+
"""
|
1512
|
+
|
1513
|
+
loss: Optional[paddle.Tensor] = None
|
1514
|
+
spectrogram: paddle.Tensor = None
|
1515
|
+
past_key_values: Optional[Tuple[Tuple[paddle.Tensor]]] = None
|
1516
|
+
decoder_hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1517
|
+
decoder_attentions: Optional[Tuple[paddle.Tensor]] = None
|
1518
|
+
cross_attentions: Optional[Tuple[paddle.Tensor]] = None
|
1519
|
+
encoder_last_hidden_state: Optional[paddle.Tensor] = None
|
1520
|
+
encoder_hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1521
|
+
encoder_attentions: Optional[Tuple[paddle.Tensor]] = None
|
1522
|
+
|
1523
|
+
|
1524
|
+
@dataclass
|
1525
|
+
class MoEModelOutputWithPast(ModelOutput):
|
1526
|
+
"""
|
1527
|
+
Base class for model's outputs, with potential hidden states and attentions.
|
1528
|
+
|
1529
|
+
Args:
|
1530
|
+
last_hidden_state (`paddle.Tensor` of shape `(batch_size, sequence_length, hidden_size)`):
|
1531
|
+
Sequence of hidden-states at the output of the last layer of the model.
|
1532
|
+
past_key_values (`tuple(tuple(paddle.Tensor))`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`):
|
1533
|
+
Tuple of `tuple(paddle.Tensor)` of length `config.n_layers`, with each tuple having 2 tensors of shape
|
1534
|
+
`(batch_size, num_heads, sequence_length, embed_size_per_head)`) and optionally if
|
1535
|
+
`config.is_encoder_decoder=True` 2 additional tensors of shape `(batch_size, num_heads,
|
1536
|
+
encoder_sequence_length, embed_size_per_head)`.
|
1537
|
+
|
1538
|
+
Contains pre-computed hidden-states (key and values in the self-attention blocks and optionally if
|
1539
|
+
`config.is_encoder_decoder=True` in the cross-attention blocks) that can be used (see `past_key_values`
|
1540
|
+
input) to speed up sequential decoding.
|
1541
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
1542
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1543
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
1544
|
+
|
1545
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
1546
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
1547
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1548
|
+
sequence_length)`.
|
1549
|
+
|
1550
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
1551
|
+
heads.
|
1552
|
+
router_logits (`tuple(paddle.Tensor)`, *optional*, returned when `output_router_probs=True` and `config.add_router_probs=True` is passed or when `config.output_router_probs=True`):
|
1553
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, sequence_length, num_experts)`.
|
1554
|
+
|
1555
|
+
Raw router logtis (post-softmax) that are computed by MoE routers, these terms are used to compute the auxiliary
|
1556
|
+
loss for Mixture of Experts models.
|
1557
|
+
"""
|
1558
|
+
|
1559
|
+
last_hidden_state: paddle.Tensor = None
|
1560
|
+
past_key_values: Optional[Tuple[Tuple[paddle.Tensor]]] = None
|
1561
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1562
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
1563
|
+
router_logits: Optional[Tuple[paddle.Tensor]] = None
|
1564
|
+
|
1565
|
+
|
1566
|
+
@dataclass
|
1567
|
+
class MoECausalLMOutputWithPast(ModelOutput):
|
1568
|
+
"""
|
1569
|
+
Base class for causal language model (or autoregressive) with mixture of experts outputs.
|
1570
|
+
|
1571
|
+
Args:
|
1572
|
+
loss (`paddle.Tensor` of shape `(1,)`, *optional*, returned when `labels` is provided):
|
1573
|
+
Language modeling loss (for next-token prediction).
|
1574
|
+
|
1575
|
+
logits (`paddle.Tensor` of shape `(batch_size, sequence_length, config.vocab_size)`):
|
1576
|
+
Prediction scores of the language modeling head (scores for each vocabulary token before SoftMax).
|
1577
|
+
|
1578
|
+
aux_loss (`paddle.Tensor`, *optional*, returned when `labels` is provided):
|
1579
|
+
aux_loss for the sparse modules.
|
1580
|
+
|
1581
|
+
router_logits (`tuple(paddle.Tensor)`, *optional*, returned when `output_router_probs=True` and `config.add_router_probs=True` is passed or when `config.output_router_probs=True`):
|
1582
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, sequence_length, num_experts)`.
|
1583
|
+
|
1584
|
+
Raw router logtis (post-softmax) that are computed by MoE routers, these terms are used to compute the auxiliary
|
1585
|
+
loss for Mixture of Experts models.
|
1586
|
+
|
1587
|
+
past_key_values (`tuple(tuple(paddle.Tensor))`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`):
|
1588
|
+
Tuple of `tuple(paddle.Tensor)` of length `config.n_layers`, with each tuple having 2 tensors of shape
|
1589
|
+
`(batch_size, num_heads, sequence_length, embed_size_per_head)`)
|
1590
|
+
|
1591
|
+
Contains pre-computed hidden-states (key and values in the self-attention blocks) that can be used (see
|
1592
|
+
`past_key_values` input) to speed up sequential decoding.
|
1593
|
+
hidden_states (`tuple(paddle.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`):
|
1594
|
+
Tuple of `paddle.Tensor` (one for the output of the embeddings, if the model has an embedding layer, +
|
1595
|
+
one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`.
|
1596
|
+
|
1597
|
+
Hidden-states of the model at the output of each layer plus the optional initial embedding outputs.
|
1598
|
+
attentions (`tuple(paddle.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`):
|
1599
|
+
Tuple of `paddle.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length,
|
1600
|
+
sequence_length)`.
|
1601
|
+
|
1602
|
+
Attentions weights after the attention softmax, used to compute the weighted average in the self-attention
|
1603
|
+
heads.
|
1604
|
+
"""
|
1605
|
+
|
1606
|
+
loss: Optional[paddle.Tensor] = None
|
1607
|
+
aux_loss: Optional[paddle.Tensor] = None
|
1608
|
+
logits: paddle.Tensor = None
|
1609
|
+
past_key_values: Optional[Tuple[Tuple[paddle.Tensor]]] = None
|
1610
|
+
hidden_states: Optional[Tuple[paddle.Tensor]] = None
|
1611
|
+
attentions: Optional[Tuple[paddle.Tensor]] = None
|
1612
|
+
router_logits: Optional[Tuple[paddle.Tensor]] = None
|