浏览代码

Merge branch 'develop' of https://github.com/PaddlePaddle/PaddleX into develop_fix

FlyingQianMM 4 年之前
父节点
当前提交
3512d7d5c5

+ 12 - 9
dygraph/paddlex/cv/models/base.py

@@ -122,7 +122,14 @@ class BaseModel:
         info = dict()
         info['pruner'] = self.pruner.__class__.__name__
         info['pruning_ratios'] = self.pruning_ratios
-        info['pruner_inputs'] = self.pruner.inputs
+        pruner_inputs = self.pruner.inputs
+        if self.model_type == 'detector':
+            pruner_inputs = {
+                k: v.tolist()
+                for k, v in pruner_inputs[0].items()
+            }
+        info['pruner_inputs'] = pruner_inputs
+
         return info
 
     def get_quant_info(self):
@@ -333,12 +340,13 @@ class BaseModel:
             eval_epoch_tic = time.time()
             if (i + 1) % save_interval_epochs == 0 or i == num_epochs - 1:
                 if eval_dataset is not None and eval_dataset.num_samples > 0:
-                    self.eval_metrics = self.evaluate(
+                    eval_result = self.evaluate(
                         eval_dataset,
                         batch_size=eval_batch_size,
-                        return_details=False)
+                        return_details=True)
                     # 保存最优模型
                     if local_rank == 0:
+                        self.eval_metrics, self.eval_details = eval_result
                         logging.info('[EVAL] Finished, Epoch={}, {} .'.format(
                             i + 1, dict2str(self.eval_metrics)))
                         best_accuracy_key = list(self.eval_metrics.keys())[0]
@@ -426,12 +434,7 @@ class BaseModel:
         pre_pruning_flops = flops(self.net, self.pruner.inputs)
         logging.info("Pre-pruning FLOPs: {}. Pruning starts...".format(
             pre_pruning_flops))
-        skip_vars = []
-        for param in self.net.parameters():
-            if param.shape[0] <= 8:
-                skip_vars.append(param.name)
-        _, self.pruning_ratios = sensitive_prune(self.pruner, pruned_flops,
-                                                 skip_vars)
+        _, self.pruning_ratios = sensitive_prune(self.pruner, pruned_flops)
         post_pruning_flops = flops(self.net, self.pruner.inputs)
         logging.info("Pruning is complete. Post-pruning FLOPs: {}".format(
             post_pruning_flops))

+ 6 - 2
dygraph/paddlex/cv/models/classifier.py

@@ -104,6 +104,7 @@ class BaseClassifier(BaseModel):
             acc1 = paddle.metric.accuracy(softmax_out, label=labels)
             k = min(5, self.num_classes)
             acck = paddle.metric.accuracy(softmax_out, label=labels, k=k)
+            prediction = softmax_out
             # multi cards eval
             if paddle.distributed.get_world_size() > 1:
                 acc1 = paddle.distributed.all_reduce(
@@ -112,9 +113,12 @@ class BaseClassifier(BaseModel):
                 acck = paddle.distributed.all_reduce(
                     acck, op=paddle.distributed.ReduceOp.
                     SUM) / paddle.distributed.get_world_size()
+                prediction = []
+                paddle.distributed.all_gather(prediction, softmax_out)
+                prediction = paddle.concat(prediction, axis=0)
 
             outputs = OrderedDict([('acc1', acc1), ('acc{}'.format(k), acck),
-                                   ('prediction', softmax_out)])
+                                   ('prediction', prediction)])
 
         else:
             # mode == 'train'
@@ -361,7 +365,7 @@ class BaseClassifier(BaseModel):
             for step, data in enumerate(self.eval_data_loader()):
                 outputs = self.run(self.net, data, mode='eval')
                 if return_details:
-                    eval_details.append(outputs['prediction'].numpy())
+                    eval_details.append(outputs['prediction'].tolist())
                 outputs.pop('prediction')
                 eval_metrics.update(outputs)
         if return_details:

+ 5 - 0
dygraph/paddlex/cv/models/load_model.py

@@ -92,6 +92,11 @@ def load_model(model_dir):
             with open(osp.join(model_dir, "prune.yml")) as f:
                 pruning_info = yaml.load(f.read(), Loader=yaml.Loader)
                 inputs = pruning_info['pruner_inputs']
+                if model.model_type == 'detector':
+                    inputs = [{
+                        k: paddle.to_tensor(v)
+                        for k, v in inputs.items()
+                    }]
                 model.pruner = getattr(paddleslim, pruning_info['pruner'])(
                     model.net, inputs=inputs)
                 model.pruning_ratios = pruning_info['pruning_ratios']

+ 23 - 11
dygraph/paddlex/cv/models/segmenter.py

@@ -112,11 +112,13 @@ class BaseSegmenter(BaseModel):
             origin_shape = [label.shape[-2:]]
             # TODO: 替换cv2后postprocess移出run
             pred = self._postprocess(pred, origin_shape, transforms=inputs[2])
-            intersect_area, pred_area, label_area = metrics.calculate_area(
+            intersect_area, pred_area, label_area = paddleseg.utils.metrics.calculate_area(
                 pred, label, self.num_classes)
             outputs['intersect_area'] = intersect_area
             outputs['pred_area'] = pred_area
             outputs['label_area'] = label_area
+            outputs['conf_mat'] = metrics.confusion_matrix(pred, label,
+                                                           self.num_classes)
         if mode == 'train':
             loss_list = metrics.loss_computation(
                 logits_list=net_out, labels=inputs[1], losses=self.losses)
@@ -355,6 +357,7 @@ class BaseSegmenter(BaseModel):
         intersect_area_all = 0
         pred_area_all = 0
         label_area_all = 0
+        conf_mat_all = []
         logging.info(
             "Start to evaluate(total_samples={}, total_steps={})...".format(
                 eval_dataset.num_samples,
@@ -366,16 +369,19 @@ class BaseSegmenter(BaseModel):
                 pred_area = outputs['pred_area']
                 label_area = outputs['label_area']
                 intersect_area = outputs['intersect_area']
+                conf_mat = outputs['conf_mat']
 
                 # Gather from all ranks
                 if nranks > 1:
                     intersect_area_list = []
                     pred_area_list = []
                     label_area_list = []
+                    conf_mat_list = []
                     paddle.distributed.all_gather(intersect_area_list,
                                                   intersect_area)
                     paddle.distributed.all_gather(pred_area_list, pred_area)
                     paddle.distributed.all_gather(label_area_list, label_area)
+                    paddle.distributed.all_gather(conf_mat_list, conf_mat)
 
                     # Some image has been evaluated and should be eliminated in last iter
                     if (step + 1) * nranks > len(eval_dataset):
@@ -383,23 +389,25 @@ class BaseSegmenter(BaseModel):
                         intersect_area_list = intersect_area_list[:valid]
                         pred_area_list = pred_area_list[:valid]
                         label_area_list = label_area_list[:valid]
+                        conf_mat_list = conf_mat_list[:valid]
 
-                    for i in range(len(intersect_area_list)):
-                        intersect_area_all = intersect_area_all + intersect_area_list[
-                            i]
-                        pred_area_all = pred_area_all + pred_area_list[i]
-                        label_area_all = label_area_all + label_area_list[i]
+                    intersect_area_all += sum(intersect_area_list)
+                    pred_area_all += sum(pred_area_list)
+                    label_area_all += sum(label_area_list)
+                    conf_mat_all.extend(conf_mat_list)
 
                 else:
                     intersect_area_all = intersect_area_all + intersect_area
                     pred_area_all = pred_area_all + pred_area
                     label_area_all = label_area_all + label_area
-        class_iou, miou = metrics.mean_iou(intersect_area_all, pred_area_all,
-                                           label_area_all)
+                    conf_mat_all.append(conf_mat)
+        class_iou, miou = paddleseg.utils.metrics.mean_iou(
+            intersect_area_all, pred_area_all, label_area_all)
         # TODO 确认是按oacc还是macc
-        class_acc, oacc = metrics.accuracy(intersect_area_all, pred_area_all)
-        kappa = metrics.kappa(intersect_area_all, pred_area_all,
-                              label_area_all)
+        class_acc, oacc = paddleseg.utils.metrics.accuracy(intersect_area_all,
+                                                           pred_area_all)
+        kappa = paddleseg.utils.metrics.kappa(intersect_area_all,
+                                              pred_area_all, label_area_all)
         category_f1score = metrics.f1_score(intersect_area_all, pred_area_all,
                                             label_area_all)
         eval_metrics = OrderedDict(
@@ -408,6 +416,10 @@ class BaseSegmenter(BaseModel):
                 'category_F1-score'
             ], [miou, class_iou, oacc, class_acc, kappa, category_f1score]))
 
+        if return_details:
+            conf_mat = sum(conf_mat_all)
+            eval_details = {'confusion_matrix': conf_mat.tolist()}
+            return eval_metrics, eval_details
         return eval_metrics
 
     def predict(self, img_file, transforms=None):

+ 18 - 130
dygraph/paddlex/cv/models/utils/seg_metrics.py

@@ -14,7 +14,6 @@
 
 import numpy as np
 import paddle
-import paddle.nn.functional as F
 
 
 def loss_computation(logits_list, labels, losses):
@@ -27,135 +26,6 @@ def loss_computation(logits_list, labels, losses):
     return loss_list
 
 
-def calculate_area(pred, label, num_classes, ignore_index=255):
-    """
-    Calculate intersect, prediction and label area
-
-    Args:
-        pred (Tensor): The prediction by model.
-        label (Tensor): The ground truth of image.
-        num_classes (int): The unique number of target classes.
-        ignore_index (int): Specifies a target value that is ignored. Default: 255.
-
-    Returns:
-        Tensor: The intersection area of prediction and the ground on all class.
-        Tensor: The prediction area on all class.
-        Tensor: The ground truth area on all class
-    """
-    if len(pred.shape) == 4:
-        pred = paddle.squeeze(pred, axis=1)
-    if len(label.shape) == 4:
-        label = paddle.squeeze(label, axis=1)
-    if not pred.shape == label.shape:
-        raise ValueError('Shape of `pred` and `label should be equal, '
-                         'but there are {} and {}.'.format(pred.shape,
-                                                           label.shape))
-
-    # Delete ignore_index
-    mask = label != ignore_index
-    pred = pred + 1
-    label = label + 1
-    pred = pred * mask
-    label = label * mask
-    pred = F.one_hot(pred, num_classes + 1)
-    label = F.one_hot(label, num_classes + 1)
-    pred = pred[:, :, :, 1:]
-    label = label[:, :, :, 1:]
-
-    pred_area = []
-    label_area = []
-    intersect_area = []
-
-    for i in range(num_classes):
-        pred_i = pred[:, :, :, i]
-        label_i = label[:, :, :, i]
-        pred_area_i = paddle.sum(pred_i)
-        label_area_i = paddle.sum(label_i)
-        intersect_area_i = paddle.sum(pred_i * label_i)
-        pred_area.append(pred_area_i)
-        label_area.append(label_area_i)
-        intersect_area.append(intersect_area_i)
-    pred_area = paddle.concat(pred_area)
-    label_area = paddle.concat(label_area)
-    intersect_area = paddle.concat(intersect_area)
-    return intersect_area, pred_area, label_area
-
-
-def mean_iou(intersect_area, pred_area, label_area):
-    """
-    Calculate iou.
-
-    Args:
-        intersect_area (Tensor): The intersection area of prediction and ground truth on all classes.
-        pred_area (Tensor): The prediction area on all classes.
-        label_area (Tensor): The ground truth area on all classes.
-
-    Returns:
-        np.ndarray: iou on all classes.
-        float: mean iou of all classes.
-    """
-    intersect_area = intersect_area.numpy()
-    pred_area = pred_area.numpy()
-    label_area = label_area.numpy()
-    union = pred_area + label_area - intersect_area
-    class_iou = []
-    for i in range(len(intersect_area)):
-        if union[i] == 0:
-            iou = 0
-        else:
-            iou = intersect_area[i] / union[i]
-        class_iou.append(iou)
-    miou = np.mean(class_iou)
-    return np.array(class_iou), miou
-
-
-def accuracy(intersect_area, pred_area):
-    """
-    Calculate accuracy
-
-    Args:
-        intersect_area (Tensor): The intersection area of prediction and ground truth on all classes..
-        pred_area (Tensor): The prediction area on all classes.
-
-    Returns:
-        np.ndarray: accuracy on all classes.
-        float: mean accuracy.
-    """
-    intersect_area = intersect_area.numpy()
-    pred_area = pred_area.numpy()
-    class_acc = []
-    for i in range(len(intersect_area)):
-        if pred_area[i] == 0:
-            acc = 0
-        else:
-            acc = intersect_area[i] / pred_area[i]
-        class_acc.append(acc)
-    macc = np.sum(intersect_area) / np.sum(pred_area)
-    return np.array(class_acc), macc
-
-
-def kappa(intersect_area, pred_area, label_area):
-    """
-    Calculate kappa coefficient
-
-    Args:
-        intersect_area (Tensor): The intersection area of prediction and ground truth on all classes..
-        pred_area (Tensor): The prediction area on all classes.
-        label_area (Tensor): The ground truth area on all classes.
-
-    Returns:
-        float: kappa coefficient.
-    """
-    intersect_area = intersect_area.numpy()
-    pred_area = pred_area.numpy()
-    label_area = label_area.numpy()
-    total_area = np.sum(label_area)
-    po = np.sum(intersect_area) / total_area
-    pe = np.sum(pred_area * label_area) / (total_area * total_area)
-    kappa = (po - pe) / (1 - pe)
-    return kappa
-
-
 def f1_score(intersect_area, pred_area, label_area):
     intersect_area = intersect_area.numpy()
     pred_area = pred_area.numpy()
@@ -172,3 +42,21 @@ def f1_score(intersect_area, pred_area, label_area):
             f1_sco = 2 * prec * rec / (prec + rec)
         class_f1_sco.append(f1_sco)
     return np.array(class_f1_sco)
+
+
+def confusion_matrix(pred, label, num_classes, ignore_index=255):
+    label = paddle.transpose(label, (0, 2, 3, 1))
+    pred = paddle.transpose(pred, (0, 2, 3, 1))
+    mask = label != ignore_index
+
+    label = paddle.masked_select(label, mask)
+    pred = paddle.masked_select(pred, mask)
+
+    cat_matrix = num_classes * label + pred
+    conf_mat = paddle.histogram(
+        cat_matrix,
+        bins=num_classes * num_classes,
+        min=0,
+        max=num_classes * num_classes - 1).reshape([num_classes, num_classes])
+
+    return conf_mat

+ 1 - 1
dygraph/requirements.txt

@@ -1,4 +1,4 @@
--r PaddleClas/requirements.txt
+-r ./PaddleClas/requirements.txt
 -r ./PaddleSeg/requirements.txt
 ./PaddleSeg
 -r ./PaddleDetection/requirements.txt

+ 1 - 1
dygraph/tutorials/slim/prune/image_classification/mobilenetv2_train.py

@@ -32,7 +32,7 @@ eval_dataset = pdx.datasets.ImageNet(
 # 初始化模型,并进行训练
 # 可使用VisualDL查看训练指标,参考https://github.com/PaddlePaddle/PaddleX/tree/release/2.0-rc/tutorials/train#visualdl可视化训练指标
 num_classes = len(train_dataset.labels)
-model = pdx.models.MobileNetV3_large(num_classes=num_classes)
+model = pdx.models.MobileNetV2(num_classes=num_classes)
 
 # API说明:https://github.com/PaddlePaddle/PaddleX/blob/95c53dec89ab0f3769330fa445c6d9213986ca5f/paddlex/cv/models/classifier.py#L153
 # 各参数介绍与调整说明:https://paddlex.readthedocs.io/zh_CN/develop/appendix/parameters.html

+ 54 - 0
dygraph/tutorials/train/instance_segmentation/mask_rcnn_r50_fpn.py

@@ -0,0 +1,54 @@
+import paddlex as pdx
+from paddlex import transforms as T
+
+# 下载和解压小度熊分拣数据集
+dataset = 'https://bj.bcebos.com/paddlex/datasets/xiaoduxiong_ins_det.tar.gz'
+pdx.utils.download_and_decompress(dataset, path='./')
+
+# 定义训练和验证时的transforms
+# API说明:https://github.com/PaddlePaddle/PaddleX/blob/release/2.0-rc/paddlex/cv/transforms/operators.py
+train_transforms = T.Compose([
+    T.RandomResizeByShort(
+        short_sizes=[640, 672, 704, 736, 768, 800],
+        max_size=1333,
+        interp='CUBIC'), T.RandomHorizontalFlip(), T.Normalize(
+            mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
+])
+
+eval_transforms = T.Compose([
+    T.ResizeByShort(
+        short_size=800, max_size=1333, interp='CUBIC'), T.Normalize(
+            mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
+])
+
+# 定义训练和验证所用的数据集
+# API说明:https://github.com/PaddlePaddle/PaddleX/blob/develop/dygraph/paddlex/cv/datasets/coco.py#L26
+train_dataset = pdx.datasets.CocoDetection(
+    data_dir='xiaoduxiong_ins_det/JPEGImages',
+    ann_file='xiaoduxiong_ins_det/train.json',
+    transforms=train_transforms,
+    shuffle=True)
+eval_dataset = pdx.datasets.CocoDetection(
+    data_dir='xiaoduxiong_ins_det/JPEGImages',
+    ann_file='xiaoduxiong_ins_det/val.json',
+    transforms=eval_transforms)
+
+# 初始化模型,并进行训练
+# 可使用VisualDL查看训练指标,参考https://github.com/PaddlePaddle/PaddleX/tree/release/2.0-rc/tutorials/train#visualdl可视化训练指标
+num_classes = len(train_dataset.labels)
+model = pdx.models.MaskRCNN(
+    num_classes=num_classes, backbone='ResNet50', with_fpn=True)
+
+# API说明:https://github.com/PaddlePaddle/PaddleX/blob/release/2.0-rc/paddlex/cv/models/detector.py#L155
+# 各参数介绍与调整说明:https://paddlex.readthedocs.io/zh_CN/develop/appendix/parameters.html
+model.train(
+    num_epochs=12,
+    train_dataset=train_dataset,
+    train_batch_size=1,
+    eval_dataset=eval_dataset,
+    learning_rate=0.00125,
+    lr_decay_epochs=[8, 11],
+    warmup_steps=10,
+    warmup_start_lr=0.0,
+    save_dir='output/mask_rcnn_r50_fpn',
+    use_vdl=True)

+ 1 - 1
dygraph/tutorials/train/object_detection/faster_rcnn_hrnet_w18.py

@@ -42,7 +42,7 @@ eval_dataset = pdx.datasets.VOCDetection(
 num_classes = len(train_dataset.labels)
 model = pdx.models.FasterRCNN(num_classes=num_classes, backbone='HRNet_W18')
 
-# API说明:https://github.com/PaddlePaddle/PaddleX/blob/release/2.0-rc/paddlex/cv/models/detector.py#L154
+# API说明:https://github.com/PaddlePaddle/PaddleX/blob/release/2.0-rc/paddlex/cv/models/detector.py#L155
 # 各参数介绍与调整说明:https://paddlex.readthedocs.io/zh_CN/develop/appendix/parameters.html
 model.train(
     num_epochs=24,

+ 1 - 1
dygraph/tutorials/train/object_detection/faster_rcnn_r50_fpn.py

@@ -43,7 +43,7 @@ num_classes = len(train_dataset.labels)
 model = pdx.models.FasterRCNN(
     num_classes=num_classes, backbone='ResNet50', with_fpn=True)
 
-# API说明:https://github.com/PaddlePaddle/PaddleX/blob/release/2.0-rc/paddlex/cv/models/detector.py#L154
+# API说明:https://github.com/PaddlePaddle/PaddleX/blob/release/2.0-rc/paddlex/cv/models/detector.py#L155
 # 各参数介绍与调整说明:https://paddlex.readthedocs.io/zh_CN/develop/appendix/parameters.html
 model.train(
     num_epochs=12,

+ 1 - 1
dygraph/tutorials/train/object_detection/ppyolo.py

@@ -43,7 +43,7 @@ eval_dataset = pdx.datasets.VOCDetection(
 num_classes = len(train_dataset.labels)
 model = pdx.models.PPYOLO(num_classes=num_classes, backbone='ResNet50_vd_dcn')
 
-# API说明:https://github.com/PaddlePaddle/PaddleX/blob/release/2.0-rc/paddlex/cv/models/detector.py#L154
+# API说明:https://github.com/PaddlePaddle/PaddleX/blob/release/2.0-rc/paddlex/cv/models/detector.py#L155
 # 各参数介绍与调整说明:https://paddlex.readthedocs.io/zh_CN/develop/appendix/parameters.html
 model.train(
     num_epochs=405,

+ 1 - 1
dygraph/tutorials/train/object_detection/ppyolotiny.py

@@ -43,7 +43,7 @@ eval_dataset = pdx.datasets.VOCDetection(
 num_classes = len(train_dataset.labels)
 model = pdx.models.PPYOLOTiny(num_classes=num_classes)
 
-# API说明:https://github.com/PaddlePaddle/PaddleX/blob/release/2.0-rc/paddlex/cv/models/detector.py#L154
+# API说明:https://github.com/PaddlePaddle/PaddleX/blob/release/2.0-rc/paddlex/cv/models/detector.py#L155
 # 各参数介绍与调整说明:https://paddlex.readthedocs.io/zh_CN/develop/appendix/parameters.html
 model.train(
     num_epochs=650,

+ 1 - 1
dygraph/tutorials/train/object_detection/ppyolov2.py

@@ -47,7 +47,7 @@ num_classes = len(train_dataset.labels)
 model = pdx.models.PPYOLOv2(
     num_classes=num_classes, backbone='ResNet50_vd_dcn')
 
-# API说明:https://github.com/PaddlePaddle/PaddleX/blob/release/2.0-rc/paddlex/cv/models/detector.py#L154
+# API说明:https://github.com/PaddlePaddle/PaddleX/blob/release/2.0-rc/paddlex/cv/models/detector.py#L155
 # 各参数介绍与调整说明:https://paddlex.readthedocs.io/zh_CN/develop/appendix/parameters.html
 model.train(
     num_epochs=365,

+ 1 - 1
dygraph/tutorials/train/object_detection/yolov3_darknet53.py

@@ -43,7 +43,7 @@ eval_dataset = pdx.datasets.VOCDetection(
 num_classes = len(train_dataset.labels)
 model = pdx.models.YOLOv3(num_classes=num_classes, backbone='DarkNet53')
 
-# API说明:https://github.com/PaddlePaddle/PaddleX/blob/release/2.0-rc/paddlex/cv/models/detector.py#L154
+# API说明:https://github.com/PaddlePaddle/PaddleX/blob/release/2.0-rc/paddlex/cv/models/detector.py#L155
 # 各参数介绍与调整说明:https://paddlex.readthedocs.io/zh_CN/develop/appendix/parameters.html
 model.train(
     num_epochs=270,