PaddleX Model List (Hygon DCU)¶
PaddleX incorporates multiple pipelines, each containing several modules, and each module encompasses various models. The specific models to use can be selected based on the benchmark data below. If you prioritize model accuracy, choose models with higher accuracy. If you prioritize model storage size, select models with smaller storage sizes.
Image Classification Module¶
Model Name | Top-1 Accuracy (%) | Model Storage Size (M) | Model Download Link |
---|---|---|---|
ResNet18 | 71.0 | 41.5 M | Inference Model/Trained Model |
ResNet34 | 74.6 | 77.3 M | Inference Model/Trained Model |
ResNet50 | 76.5 | 90.8 M | Inference Model/Trained Model |
ResNet101 | 77.6 | 158.7 M | Inference Model/Trained Model |
ResNet152 | 78.3 | 214.2 M | Inference Model/Trained Model |
PP-LCNet_x1_0 | 71.32 | 10.5 M | Inference Model/Trained Model |
Note: The above accuracy metrics are Top-1 Accuracy on the ImageNet-1k validation set.
Image Multi-label Classification Module¶
Model Name | mAP (%) | Model Storage Size | Model Download Link |
---|---|---|---|
CLIP_vit_base_patch16_448_ML | 89.15 | 325.6 M | Inference Model/Trained Model |
PP-HGNetV2-B0_ML | 80.98 | 39.6 M | Inference Model/Trained Model |
PP-HGNetV2-B4_ML | 87.96 | 88.5 M | Inference Model/Trained Model |
PP-HGNetV2-B6_ML | 91.06 | 286.5 M | Inference Model/Trained Model |
Note: The above accuracy metrics are for the multi-label classification task mAP of COCO2017.
Image Feature Module¶
Model Name | recall@1(%) | Model Size | Model Download Link |
---|---|---|---|
PP-ShiTuV2_rec_CLIP_vit_base | 88.69 | 306.6 M | Inference Model/Trained Model |
Note: The above accuracy metrics are for AliProducts recall@1。
Object Detection Module¶
Model Name | mAP (%) | Model Size (M) | Model Download Link |
---|---|---|---|
PicoDet-L | 42.6 | 20.9 M | Inference Model/Trained Model |
PicoDet-M | 37.5 | 16.8 M | Inference Model/Trained Model |
PicoDet-S | 29.1 | 4.4 M | Inference Model/Trained Model |
PicoDet-XS | 26.2 | 5.7M | Inference Model/Trained Model |
PP-YOLOE_plus-L | 52.9 | 185.3 M | Inference Model/Trained Model |
PP-YOLOE_plus-M | 49.8 | 83.2 M | Inference Model/Trained Model |
PP-YOLOE_plus-S | 43.7 | 28.3 M | Inference Model/Trained Model |
PP-YOLOE_plus-X | 54.7 | 349.4 M | Inference Model/Trained Model |
RT-DETR-R18 | 46.5 | 70.7 M | Inference Model/Trained Model |
FCOS-ResNet50 | 39.6 | 124.2 M | Inference Model/Trained Model |
YOLOX-N | 26.1 | 3.4M | Inference Model/Trained Model |
FasterRCNN-ResNet34-FPN | 37.8 | 137.5 M | Inference Model/Trained Model |
YOLOv3-DarkNet53 | 39.1 | 219.7 M | Inference ModelInference Model/Trained Model |
Cascade-FasterRCNN-ResNet50-FPN | 41.1 | 245.4 M | Inference Model/Trained Model |
Note: The above accuracy metrics are mAP(0.5:0.95) on the COCO2017 validation set.
Small Object Detection Module¶
Model Name | mAP(%) | Model Size | Model Download Link |
---|---|---|---|
PP-YOLOE_plus_SOD-S | 25.1 | 77.3 M | Inference Model/Trained Model |
PP-YOLOE_plus_SOD-L | 31.9 | 325.0 M | Inference Model/Trained Model |
PP-YOLOE_plus_SOD-largesize-L | 42.7 | 340.5 M | Inference Model/Trained Model |
Note: The above accuracy metrics are for VisDrone-DET validation set mAP(0.5:0.95)。
Semantic Segmentation Module¶
Model Name | mIoU (%) | Model Storage Size (M) | Model Download Link |
---|---|---|---|
Deeplabv3_Plus-R50 | 80.36 | 94.9 M | Inference Model/Trained Model |
Deeplabv3_Plus-R101 | 81.10 | 162.5 M | Inference Model/Trained Model |
PP-LiteSeg-T | 73.10 | 28.5 M | Inference Model/Trained Model |
Note: The above accuracy metrics are mIoU on the Cityscapes dataset.
Abnormality Detection Module¶
Model Name | Avg(%) | Model Size | Model Download Link |
---|---|---|---|
STFPM | 96.2 | 21.5 M | Inference Model/Trained Model |
Note: The above accuracy metrics are evaluated on the MVTec AD dataset using the average anomaly score.
Face Detection Module¶
Model Name | AP (%) Easy/Medium/Hard |
Model Size | Model Download Link |
---|---|---|---|
PicoDet_LCNet_x2_5_face | 93.7/90.7/68.1 | 28.9 M | Inference Model/Trained Model |
Note: The above accuracy metrics are evaluated on the WIDER-FACE validation set with an input size of 640*640.
Text Detection Module¶
Model Name | Detection Hmean (%) | Model Size (M) | Model Download Link |
---|---|---|---|
PP-OCRv4_mobile_det | 77.79 | 4.2 M | Inference Model/Trained Model |
PP-OCRv4_server_det | 82.69 | 100.1M | Inference Model/Trained Model |
Note: The evaluation set for the above accuracy metrics is PaddleOCR's self-built Chinese dataset, covering street scenes, web images, documents, handwriting, and more scenarios, with 500 images for detection.
Text Recognition Module¶
Model Name | Recognition Avg Accuracy (%) | Model Size (M) | Model Download Link |
---|---|---|---|
PP-OCRv4_mobile_rec | 78.20 | 10.6 M | Inference Model/Trained Model |
PP-OCRv4_server_rec | 79.20 | 71.2 M | Inference Model/Trained Model |
Note: The evaluation set for the above accuracy metrics is PaddleOCR's self-built Chinese dataset, covering street scenes, web images, documents, handwriting, and more scenarios, with 11,000 images for text recognition.
Time Series Forecasting Module¶
Model Name | mse | mae | Model Size (M) | Model Download Link |
---|---|---|---|---|
DLinear | 0.382 | 0.394 | 72K | Inference Model/Trained Model |
NLinear | 0.386 | 0.392 | 40K | Inference Model/Trained Model |
RLinear | 0.384 | 0.392 | 40K | Inference Model/Trained Model |
Note: The above accuracy metrics are measured on the ETTH1 dataset (evaluation results on the test set test.csv).