飞桨常规赛:PALM眼底彩照中黄斑中央凹定位
飞桨常规赛:PALM眼底彩照中黄斑中央凹定位
aistudio地址:
https://aistudio.baidu.com/aistudio/projectdetail/2190500?contributionType=1
0. 赛题介绍
常规赛:PALM眼底彩照中黄斑中央凹定位由ISBI2019 PALM眼科挑战赛赛题再现,其中黄斑中央凹定位的任务旨在对眼科图像进行判断是否存在黄斑中央凹,并对其进行定位。
数据集由中山大学中山眼科中心提供800张带黄斑中央凹坐标标注的眼底彩照供选手训练模型,另提供400张带标注数据供平台进行模型测试。图像分辨率为1444×1444,或2124×2056。黄斑中央凹坐标信息存储在xlsx文件中,名为“Fovea_Location_train”,第一列对应眼底图像的文件名(包括扩展名“.jpg”),第二列包含x坐标,第三列包含y坐标。
评价指标为平均欧式距离,计算每个测试样本预测的黄斑中央凹坐标与金标准的差距,最终计算平均的欧式距离。 最终评分为平均欧式距离的倒数。
比赛链接: 常规赛:PALM眼底彩照中黄斑中央凹定位
1. 包准备
用resnet50加载一个训练更多的预训练模型。
import os
import pandas as pd
import numpy as np
import paddle
import paddle.vision.transforms as T
from paddle.io import Dataset
from PIL import Image
/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/paddle/fluid/layers/utils.py:26: DeprecationWarning: `np.int` is a deprecated alias for the builtin `int`. To silence this warning, use `int` by itself. Doing this will not modify any behavior and is safe. When replacing `np.int`, you may wish to use e.g. `np.int64` or `np.int32` to specify the precision. If you wish to review your current use, check the release note link for additional information.
Deprecated in NumPy 1.20; for more details and guidance: https://numpy.org/devdocs/release/1.20.0-notes.html#deprecationsdef convert_to_list(value, n, name, dtype=np.int):
2. 数据准备
2.1解压数据集
! unzip -oq data/data100179/常规赛:PALM眼底彩照中黄斑中央凹定位.zip
! rm -rf __MACOSX
! mv 常规赛:PALM眼底彩照中黄斑中央凹定位 PLAM
2.2 配置数据集
- 因为数据中本身就有了这个图像名和标签,我们就不用生成数据列表了。直接继承io中的Dataset,用于读取数据。因为与开始说数据的大小有两种分辨率为1444×1444,或2124×2056。这里都放到了1440+360=1800,基本上取二者之间吧。
- 划分的比列为0.85,图像增强只有简单的色彩和水平翻转。
- 这里主要注意当图像大小发生了变化,对应的中央凹的坐标也需要进行相应的变化。数据增广就没有写几何变换有关的了,因为这样就要自己写一下对应增强。
- 最开始数据我还加上了是否存在的类别标签,但是后期效果不明显,可能是自己的原因,这里给注释了,大佬们觉得有道理也可是试试。
2.3 数据说明
import warnings
warnings.filterwarnings("ignore") #拒绝烦人的警告信息
from paddle.io import DataLoadertpsize = 256
split = 0.9
batch_size = 16class PLAMDatas(Dataset):def __init__(self, data_path, class_xls, mode='train', transforms=None, re_size=tpsize):super(PLAMDatas, self).__init__()self.data_path = data_pathself.name_label = (pd.read_excel(class_xls)).valueslens = len(self.name_label)if mode == 'train':self.name_label = self.name_label[:int(split*lens)]else:self.name_label = self.name_label[int(split*lens):]self.transforms = transformsself.re_size = re_sizedef __getitem__(self, index):name, x, y = self.name_label[index] # 得到的数据赋值一下data_path = os.path.join(self.data_path, name) # 文件系统路径+图片的name=图片的路径data = np.asarray(Image.open(data_path).convert('RGB'))H, W, _ = data.shapeif self.transforms is not None:data = self.transforms(data)data = data.astype('float32')label = np.array([x * self.re_size / W, y * self.re_size / H]).astype('float32') # 图片大小变了,对应的坐标自然也要改变return data, labeldef __len__(self):return len(self.name_label)# 配置数据增广
train_transforms = T.Compose([T.Resize((tpsize, tpsize), interpolation='bicubic'), #都调整到1800 选用bicubic,放大不至于太失真T.ToTensor()
])val_transforms = T.Compose([T.Resize((tpsize, tpsize), interpolation='bicubic'),T.ToTensor()
])# 配置数据集
train_dataset = PLAMDatas(data_path='PLAM/Train/fundus_image', class_xls='PLAM/Train/Fovea_Location_train.xlsx', mode='train', transforms=train_transforms)
val_dataset = PLAMDatas(data_path='PLAM/Train/fundus_image', class_xls='PLAM/Train/Fovea_Location_train.xlsx', mode='test', transforms=val_transforms)train_dataloader = DataLoader(dataset=train_dataset, batch_size=batch_size,shuffle=True, drop_last=False)
dev_dataloader = DataLoader( dataset=val_dataset, batch_size=batch_size,shuffle=True, drop_last=False)
这里也可以输出测试一下,看看数据读取有没有什么问题。避免后面报一堆错不知道哪儿去找问题。顺便看看点是不是点到位了。
print(len(train_dataset), len(val_dataset))
print(len(train_dataloader), len(dev_dataloader))
720 80
45 5
3. 模型训练
3.1 模型准备
import paddle
import paddle.nn as nn
from paddle.vision.models import resnet50# 模型定义
# pre_params = paddle.load('resnet_50_save_models/final.pdparams')
# model.set_state_dict(pre_params)
model = nn.Sequential(resnet50(pretrained=True),nn.LeakyReLU(),nn.Linear(1000, 2) # 坐标定位
)
paddle.summary(model, (1, 3, tpsize, tpsize))
model = paddle.Model(model)
2021-07-23 14:09:08,122 - INFO - unique_endpoints {''}
2021-07-23 14:09:08,124 - INFO - Downloading resnet50.pdparams from https://paddle-hapi.bj.bcebos.com/models/resnet50.pdparams
100%|██████████| 151272/151272 [00:02<00:00, 61339.97it/s]
2021-07-23 14:09:10,837 - INFO - File /home/aistudio/.cache/paddle/hapi/weights/resnet50.pdparams md5 checking...-------------------------------------------------------------------------------Layer (type) Input Shape Output Shape Param #
===============================================================================Conv2D-1 [[1, 3, 256, 256]] [1, 64, 128, 128] 9,408 BatchNorm2D-1 [[1, 64, 128, 128]] [1, 64, 128, 128] 256 ReLU-1 [[1, 64, 128, 128]] [1, 64, 128, 128] 0 MaxPool2D-1 [[1, 64, 128, 128]] [1, 64, 64, 64] 0 Conv2D-3 [[1, 64, 64, 64]] [1, 64, 64, 64] 4,096 BatchNorm2D-3 [[1, 64, 64, 64]] [1, 64, 64, 64] 256 ReLU-2 [[1, 256, 64, 64]] [1, 256, 64, 64] 0 Conv2D-4 [[1, 64, 64, 64]] [1, 64, 64, 64] 36,864 BatchNorm2D-4 [[1, 64, 64, 64]] [1, 64, 64, 64] 256 Conv2D-5 [[1, 64, 64, 64]] [1, 256, 64, 64] 16,384 BatchNorm2D-5 [[1, 256, 64, 64]] [1, 256, 64, 64] 1,024 Conv2D-2 [[1, 64, 64, 64]] [1, 256, 64, 64] 16,384 BatchNorm2D-2 [[1, 256, 64, 64]] [1, 256, 64, 64] 1,024 BottleneckBlock-1 [[1, 64, 64, 64]] [1, 256, 64, 64] 0 Conv2D-6 [[1, 256, 64, 64]] [1, 64, 64, 64] 16,384 BatchNorm2D-6 [[1, 64, 64, 64]] [1, 64, 64, 64] 256 ReLU-3 [[1, 256, 64, 64]] [1, 256, 64, 64] 0 Conv2D-7 [[1, 64, 64, 64]] [1, 64, 64, 64] 36,864 BatchNorm2D-7 [[1, 64, 64, 64]] [1, 64, 64, 64] 256 Conv2D-8 [[1, 64, 64, 64]] [1, 256, 64, 64] 16,384 BatchNorm2D-8 [[1, 256, 64, 64]] [1, 256, 64, 64] 1,024 BottleneckBlock-2 [[1, 256, 64, 64]] [1, 256, 64, 64] 0 Conv2D-9 [[1, 256, 64, 64]] [1, 64, 64, 64] 16,384 BatchNorm2D-9 [[1, 64, 64, 64]] [1, 64, 64, 64] 256 ReLU-4 [[1, 256, 64, 64]] [1, 256, 64, 64] 0 Conv2D-10 [[1, 64, 64, 64]] [1, 64, 64, 64] 36,864 BatchNorm2D-10 [[1, 64, 64, 64]] [1, 64, 64, 64] 256 Conv2D-11 [[1, 64, 64, 64]] [1, 256, 64, 64] 16,384 BatchNorm2D-11 [[1, 256, 64, 64]] [1, 256, 64, 64] 1,024 BottleneckBlock-3 [[1, 256, 64, 64]] [1, 256, 64, 64] 0 Conv2D-13 [[1, 256, 64, 64]] [1, 128, 64, 64] 32,768 BatchNorm2D-13 [[1, 128, 64, 64]] [1, 128, 64, 64] 512 ReLU-5 [[1, 512, 32, 32]] [1, 512, 32, 32] 0 Conv2D-14 [[1, 128, 64, 64]] [1, 128, 32, 32] 147,456 BatchNorm2D-14 [[1, 128, 32, 32]] [1, 128, 32, 32] 512 Conv2D-15 [[1, 128, 32, 32]] [1, 512, 32, 32] 65,536 BatchNorm2D-15 [[1, 512, 32, 32]] [1, 512, 32, 32] 2,048 Conv2D-12 [[1, 256, 64, 64]] [1, 512, 32, 32] 131,072 BatchNorm2D-12 [[1, 512, 32, 32]] [1, 512, 32, 32] 2,048 BottleneckBlock-4 [[1, 256, 64, 64]] [1, 512, 32, 32] 0 Conv2D-16 [[1, 512, 32, 32]] [1, 128, 32, 32] 65,536 BatchNorm2D-16 [[1, 128, 32, 32]] [1, 128, 32, 32] 512 ReLU-6 [[1, 512, 32, 32]] [1, 512, 32, 32] 0 Conv2D-17 [[1, 128, 32, 32]] [1, 128, 32, 32] 147,456 BatchNorm2D-17 [[1, 128, 32, 32]] [1, 128, 32, 32] 512 Conv2D-18 [[1, 128, 32, 32]] [1, 512, 32, 32] 65,536 BatchNorm2D-18 [[1, 512, 32, 32]] [1, 512, 32, 32] 2,048 BottleneckBlock-5 [[1, 512, 32, 32]] [1, 512, 32, 32] 0 Conv2D-19 [[1, 512, 32, 32]] [1, 128, 32, 32] 65,536 BatchNorm2D-19 [[1, 128, 32, 32]] [1, 128, 32, 32] 512 ReLU-7 [[1, 512, 32, 32]] [1, 512, 32, 32] 0 Conv2D-20 [[1, 128, 32, 32]] [1, 128, 32, 32] 147,456 BatchNorm2D-20 [[1, 128, 32, 32]] [1, 128, 32, 32] 512 Conv2D-21 [[1, 128, 32, 32]] [1, 512, 32, 32] 65,536 BatchNorm2D-21 [[1, 512, 32, 32]] [1, 512, 32, 32] 2,048 BottleneckBlock-6 [[1, 512, 32, 32]] [1, 512, 32, 32] 0 Conv2D-22 [[1, 512, 32, 32]] [1, 128, 32, 32] 65,536 BatchNorm2D-22 [[1, 128, 32, 32]] [1, 128, 32, 32] 512 ReLU-8 [[1, 512, 32, 32]] [1, 512, 32, 32] 0 Conv2D-23 [[1, 128, 32, 32]] [1, 128, 32, 32] 147,456 BatchNorm2D-23 [[1, 128, 32, 32]] [1, 128, 32, 32] 512 Conv2D-24 [[1, 128, 32, 32]] [1, 512, 32, 32] 65,536 BatchNorm2D-24 [[1, 512, 32, 32]] [1, 512, 32, 32] 2,048 BottleneckBlock-7 [[1, 512, 32, 32]] [1, 512, 32, 32] 0 Conv2D-26 [[1, 512, 32, 32]] [1, 256, 32, 32] 131,072 BatchNorm2D-26 [[1, 256, 32, 32]] [1, 256, 32, 32] 1,024 ReLU-9 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 0 Conv2D-27 [[1, 256, 32, 32]] [1, 256, 16, 16] 589,824 BatchNorm2D-27 [[1, 256, 16, 16]] [1, 256, 16, 16] 1,024 Conv2D-28 [[1, 256, 16, 16]] [1, 1024, 16, 16] 262,144 BatchNorm2D-28 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 4,096 Conv2D-25 [[1, 512, 32, 32]] [1, 1024, 16, 16] 524,288 BatchNorm2D-25 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 4,096 BottleneckBlock-8 [[1, 512, 32, 32]] [1, 1024, 16, 16] 0 Conv2D-29 [[1, 1024, 16, 16]] [1, 256, 16, 16] 262,144 BatchNorm2D-29 [[1, 256, 16, 16]] [1, 256, 16, 16] 1,024 ReLU-10 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 0 Conv2D-30 [[1, 256, 16, 16]] [1, 256, 16, 16] 589,824 BatchNorm2D-30 [[1, 256, 16, 16]] [1, 256, 16, 16] 1,024 Conv2D-31 [[1, 256, 16, 16]] [1, 1024, 16, 16] 262,144 BatchNorm2D-31 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 4,096 BottleneckBlock-9 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 0 Conv2D-32 [[1, 1024, 16, 16]] [1, 256, 16, 16] 262,144 BatchNorm2D-32 [[1, 256, 16, 16]] [1, 256, 16, 16] 1,024 ReLU-11 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 0 Conv2D-33 [[1, 256, 16, 16]] [1, 256, 16, 16] 589,824 BatchNorm2D-33 [[1, 256, 16, 16]] [1, 256, 16, 16] 1,024 Conv2D-34 [[1, 256, 16, 16]] [1, 1024, 16, 16] 262,144 BatchNorm2D-34 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 4,096
BottleneckBlock-10 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 0 Conv2D-35 [[1, 1024, 16, 16]] [1, 256, 16, 16] 262,144 BatchNorm2D-35 [[1, 256, 16, 16]] [1, 256, 16, 16] 1,024 ReLU-12 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 0 Conv2D-36 [[1, 256, 16, 16]] [1, 256, 16, 16] 589,824 BatchNorm2D-36 [[1, 256, 16, 16]] [1, 256, 16, 16] 1,024 Conv2D-37 [[1, 256, 16, 16]] [1, 1024, 16, 16] 262,144 BatchNorm2D-37 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 4,096
BottleneckBlock-11 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 0 Conv2D-38 [[1, 1024, 16, 16]] [1, 256, 16, 16] 262,144 BatchNorm2D-38 [[1, 256, 16, 16]] [1, 256, 16, 16] 1,024 ReLU-13 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 0 Conv2D-39 [[1, 256, 16, 16]] [1, 256, 16, 16] 589,824 BatchNorm2D-39 [[1, 256, 16, 16]] [1, 256, 16, 16] 1,024 Conv2D-40 [[1, 256, 16, 16]] [1, 1024, 16, 16] 262,144 BatchNorm2D-40 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 4,096
BottleneckBlock-12 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 0 Conv2D-41 [[1, 1024, 16, 16]] [1, 256, 16, 16] 262,144 BatchNorm2D-41 [[1, 256, 16, 16]] [1, 256, 16, 16] 1,024 ReLU-14 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 0 Conv2D-42 [[1, 256, 16, 16]] [1, 256, 16, 16] 589,824 BatchNorm2D-42 [[1, 256, 16, 16]] [1, 256, 16, 16] 1,024 Conv2D-43 [[1, 256, 16, 16]] [1, 1024, 16, 16] 262,144 BatchNorm2D-43 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 4,096
BottleneckBlock-13 [[1, 1024, 16, 16]] [1, 1024, 16, 16] 0 Conv2D-45 [[1, 1024, 16, 16]] [1, 512, 16, 16] 524,288 BatchNorm2D-45 [[1, 512, 16, 16]] [1, 512, 16, 16] 2,048 ReLU-15 [[1, 2048, 8, 8]] [1, 2048, 8, 8] 0 Conv2D-46 [[1, 512, 16, 16]] [1, 512, 8, 8] 2,359,296 BatchNorm2D-46 [[1, 512, 8, 8]] [1, 512, 8, 8] 2,048 Conv2D-47 [[1, 512, 8, 8]] [1, 2048, 8, 8] 1,048,576 BatchNorm2D-47 [[1, 2048, 8, 8]] [1, 2048, 8, 8] 8,192 Conv2D-44 [[1, 1024, 16, 16]] [1, 2048, 8, 8] 2,097,152 BatchNorm2D-44 [[1, 2048, 8, 8]] [1, 2048, 8, 8] 8,192
BottleneckBlock-14 [[1, 1024, 16, 16]] [1, 2048, 8, 8] 0 Conv2D-48 [[1, 2048, 8, 8]] [1, 512, 8, 8] 1,048,576 BatchNorm2D-48 [[1, 512, 8, 8]] [1, 512, 8, 8] 2,048 ReLU-16 [[1, 2048, 8, 8]] [1, 2048, 8, 8] 0 Conv2D-49 [[1, 512, 8, 8]] [1, 512, 8, 8] 2,359,296 BatchNorm2D-49 [[1, 512, 8, 8]] [1, 512, 8, 8] 2,048 Conv2D-50 [[1, 512, 8, 8]] [1, 2048, 8, 8] 1,048,576 BatchNorm2D-50 [[1, 2048, 8, 8]] [1, 2048, 8, 8] 8,192
BottleneckBlock-15 [[1, 2048, 8, 8]] [1, 2048, 8, 8] 0 Conv2D-51 [[1, 2048, 8, 8]] [1, 512, 8, 8] 1,048,576 BatchNorm2D-51 [[1, 512, 8, 8]] [1, 512, 8, 8] 2,048 ReLU-17 [[1, 2048, 8, 8]] [1, 2048, 8, 8] 0 Conv2D-52 [[1, 512, 8, 8]] [1, 512, 8, 8] 2,359,296 BatchNorm2D-52 [[1, 512, 8, 8]] [1, 512, 8, 8] 2,048 Conv2D-53 [[1, 512, 8, 8]] [1, 2048, 8, 8] 1,048,576 BatchNorm2D-53 [[1, 2048, 8, 8]] [1, 2048, 8, 8] 8,192
BottleneckBlock-16 [[1, 2048, 8, 8]] [1, 2048, 8, 8] 0
AdaptiveAvgPool2D-1 [[1, 2048, 8, 8]] [1, 2048, 1, 1] 0 Linear-1 [[1, 2048]] [1, 1000] 2,049,000 ResNet-1 [[1, 3, 256, 256]] [1, 1000] 0 LeakyReLU-1 [[1, 1000]] [1, 1000] 0 Linear-2 [[1, 1000]] [1, 2] 2,002
===============================================================================
Total params: 25,612,154
Trainable params: 25,505,914
Non-trainable params: 106,240
-------------------------------------------------------------------------------
Input size (MB): 0.75
Forward/backward pass size (MB): 341.54
Params size (MB): 97.70
Estimated Total Size (MB): 439.99
-------------------------------------------------------------------------------
3.2 损失设定
这里采用红白黑大佬设计了loss损失函数,计算MSE和欧氏距离加权后的平均损失值
# 自定义损失
import paddle
import paddle.nn as nn
import paddle.nn.functional as Fclass FocusBCELoss(nn.Layer):'''本赛题的任务损失函数'''def __init__(self, weights=[0.5, 0.5]):super(FocusBCELoss, self).__init__()self.weights = weights # 损失权重def forward(self, predict, label):# MSE均方误差mse_loss_x = paddle.nn.functional.mse_loss(predict[:, 0], label[:, 0], reduction='mean')mse_loss_y = paddle.nn.functional.mse_loss(predict[:, 1], label[:, 1], reduction='mean')mse_loss = 0.5 * mse_loss_x + 0.5 * mse_loss_y# 欧氏距离distance_loss = paddle.subtract(predict, label)distance_loss = paddle.square(distance_loss)distance_loss = paddle.sum(distance_loss, axis=-1)distance_loss = paddle.sqrt(distance_loss)distance_loss = paddle.sum(distance_loss, axis=0) / predict.shape[0] # predict.shape[0] == batch_sizealpha1, alpha2 = self.weightsall_loss = alpha1*mse_loss + alpha2*distance_lossreturn all_loss, mse_loss, distance_loss
3.3 开始训练
这里采用PolynomialDecay跑
运行时长: 16小时20分钟53秒401毫秒
# 模型准备
epochs = 1000lr = paddle.optimizer.lr.PolynomialDecay(learning_rate=2e-3, decay_steps=int(800*tpsize))
opt = paddle.optimizer.Adam(learning_rate=lr, parameters=model.parameters(), weight_decay=paddle.regularizer.L2Decay(5e-6))
loss = FocusBCELoss(weights=[0.4, 0.6]) # weights,不同类别的损失权重model.prepare(optimizer = opt, loss = loss)
visualdl=paddle.callbacks.VisualDL(log_dir='visual_log')#在使用GPU机器时,可以将use_gpu变量设置成True
use_gpu = True
paddle.set_device('gpu:0') if use_gpu else paddle.set_device('cpu')# 模型微调
model.fit(train_data=train_dataset, eval_data=val_dataset, batch_size=batch_size, epochs=epochs, eval_freq=10, log_freq=1, save_dir='resnet_50_save_models_256_0.9_16', save_freq=10, verbose=1, drop_last=False, shuffle=True, num_workers=0,callbacks=[visualdl]
)
跑完1000轮之后的loss基本上稳定在了0.5-2.5
Epoch 993/1000
step 45/45 [==============================] - loss: 0.7025 0.5066 0.8331 - 1s/step
Epoch 994/1000
step 45/45 [==============================] - loss: 1.8822 2.1647 1.6938 - 1s/step
Epoch 995/1000
step 45/45 [==============================] - loss: 2.5339 2.9379 2.2646 - 1s/step
Epoch 996/1000
step 45/45 [==============================] - loss: 1.3060 1.2016 1.3756 - 1s/step
Epoch 997/1000
step 45/45 [==============================] - loss: 0.7337 0.4999 0.8896 - 1s/step
Epoch 998/1000
step 45/45 [==============================] - loss: 1.0640 0.8573 1.2018 - 1s/step
Epoch 999/1000
step 45/45 [==============================] - loss: 1.3307 1.1557 1.4473 - 1s/step
Epoch 1000/1000
step 45/45 [==============================] - loss: 0.5870 0.3452 0.7482 - 1s/step
可视化情况如下:
4. 模型预测
预测这里就是因为图像的大小变了,所以预测得到的坐标还需要进行一次计算还原到原来的大小,感觉这也是误差的一个来源。
import os
import numpy as np
import pandas as pd
from PIL import Image
import paddle.vision.transforms as T
import paddle
import paddle.nn as nn
import paddle.nn.functional as F
from paddle.vision.models import resnet50save_path = 'Fovea_Localization_Results.csv'
file_path = 'PLAM/PALM-Testing400-Images'
imgs_name = os.listdir(file_path)model = nn.Sequential(resnet50(pretrained=False),nn.LeakyReLU(),nn.Linear(1000, 2)
)
params = paddle.load('resnet_50_save_models_256_0.9_16/final.pdparams')
model.set_state_dict(params)
model.eval()inf_transforms = T.Compose([T.Resize((tpsize, tpsize), interpolation='bicubic'),T.ToTensor()
])pre_data = []
for img_name in imgs_name:data_path = os.path.join(file_path, img_name)data = np.asarray(Image.open(data_path).convert('RGB'))H, W, _ = data.shapedata = inf_transforms(data)data = data.astype('float32').reshape([1, 3, tpsize, tpsize])pred = model(data)pre = [None] * 2# 还原坐标pre[0] = pred.numpy()[0][0] * W / tpsizepre[1] = pred.numpy()[0][1] * H / tpsizeprint(img_name, pre)pre_data.append([img_name, pre[0], pre[1]])df = pd.DataFrame(pre_data, columns=['FileName', 'Fovea_X', 'Fovea_Y'])
df.sort_values(by="FileName",inplace=True,ascending=True) #千万记得排序!
df.to_csv(save_path, index=None)
df
FileName | Fovea_X | Fovea_Y | |
---|---|---|---|
180 | T0001.jpg | 1293.032825 | 990.452656 |
319 | T0002.jpg | 1078.404609 | 1054.413445 |
114 | T0003.jpg | 1038.041723 | 1045.842130 |
381 | T0004.jpg | 1195.271087 | 1053.216895 |
124 | T0005.jpg | 1229.637487 | 723.297234 |
... | ... | ... | ... |
246 | T0396.jpg | 1208.370796 | 972.463286 |
127 | T0397.jpg | 1245.531012 | 1054.355603 |
207 | T0398.jpg | 1310.560631 | 998.023864 |
335 | T0399.jpg | 1031.135805 | 1115.349770 |
330 | T0400.jpg | 1141.022628 | 719.758717 |
400 rows × 3 columns
5. 查看结果
import os
import numpy as np
import pandas as pd
from PIL import Image
import matplotlib.pyplot as plt%matplotlib inlinepath = 'PLAM/PALM-Testing400-Images'
flrs = np.array(pd.read_csv('Fovea_Localization_Results.csv'))
for flr in flrs:img = np.array(Image.open(os.path.join(path, flr[0])))x, y = flr[1:]plt.imshow(img.astype('uint8'))plt.plot(x, y, 'or')plt.show()break
2021-07-24 11:51:37,562 - INFO - font search path ['/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/matplotlib/mpl-data/fonts/ttf', '/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/matplotlib/mpl-data/fonts/afm', '/opt/conda/envs/python35-paddle120-env/lib/python3.7/site-packages/matplotlib/mpl-data/fonts/pdfcorefonts']
2021-07-24 11:51:38,128 - INFO - generated new fontManager
总结
- 数据要选择一个合适的标准进行处理,两种图片大小1440和2124,如果缩放的太小或者太大都会导致其中一个效果不好,上个版本我尝试了大尺寸的图片,由于尺寸太大,网络模型太大,导致训练特别慢,这次改用小尺寸换用更多轮次看看效果如何。
- 网络选取要合理QAQ,我这个应该就很不合理
- 通过近几个版本的尝试,还是采用了resnet50,然后使用256的分辨率最为合适,太大了并没有降低误差反而得分很低。
- 记得去进行排序,本来质量就不高,再不排序,分值就更低了。
参考资料
眼底彩照中黄斑中央凹定位相关论文
- Macula segmentation and fovea localization employing image processing and heuristic based clustering for automated retinal screening
- A Pixel-Wise Distance Regression Approach for Joint Retinal Optic Disc and Fovea Detection
- 飞桨常规赛:PALM眼底彩照中黄斑中央凹定位 - 5月第一名方案
- 常规赛:PALM眼底彩照中黄斑中央凹定位比赛Baseline
个人总结
全网同名: iterhui
我在AI Studio上获得钻石等级,点亮9个徽章,来互关呀~
https://aistudio.baidu.com/aistudio/personalcenter/thirdview/643467
飞桨常规赛:PALM眼底彩照中黄斑中央凹定位相关推荐
- 飞桨常规赛:PALM眼底彩照中黄斑中央凹定位-9月第1名方案
飞桨常规赛:PALM眼底彩照中黄斑中央凹定位-9月第1名方案 (1)比赛介绍 赛题介绍 榜首个人主页,戳此处查看 PALM黄斑定位常规赛的重点是研究和发展与患者眼底照片黄斑结构定位相关的算法.该常规赛 ...
- 飞桨常规赛:PALM眼底彩照中黄斑中央凹定位-11月第1名方案
飞桨常规赛:PALM眼底彩照中黄斑中央凹定位-11月第1名方案 (1)比赛介绍 赛题介绍 榜首个人主页,戳此处查看 PALM黄斑定位常规赛的重点是研究和发展与患者眼底照片黄斑结构定位相关的算法.该常规 ...
- [常规赛] PALM眼底彩照视盘探测与分割 - 10月第1名方案
1 赛题背景 本赛题原型为ISBI2019PALM眼科大赛. 近视已成为全球公共卫生负担.在近视患者中,约35%为高度近视.近视导致眼轴长度的延长,可能引起视网膜和脉络膜的病理改变.随着近视屈光度的增 ...
- [常规赛] PALM眼底彩照视盘探测与分割 - 9月第1名方案
赛题链接:常规赛:PALM眼底彩照视盘探测与分割 赛题简介:ISBI2019 PALM眼科挑战赛赛题再现,提供800张眼底彩照训练数据集, 要求选手训练模型完成眼底视盘结构的探测和分割任务. 赛题背景 ...
- 飞桨常规赛:黄斑中央凹定位(GAMMA挑战赛任务二) - 11月第3名方案
飞桨常规赛:黄斑中央凹定位(GAMMA挑战赛任务二) 11月第3名方案--鸣沙山下.伽利略 比赛地址:https://aistudio.baidu.com/aistudio/competition/d ...
- 常规赛:PALM眼底彩照视盘探测与分割202105-202205全时段第二名(得分0.97123)方案
这只是百度的aistudio平台的一个常规赛(可以一直刷榜的比赛,每月计算一次排名),虽然在整个时段是榜2,但只是一个普通的语义分割比赛,参数人数不过百.有意参赛刷榜的的朋友可以访问以下链接 常规赛: ...
- 飞桨常规赛:PALM病理性近视病灶检测与分割 - 10月第1名方案
常规赛:PALM病理性近视病灶检测与分割 具体介绍 赛题介绍 PALM病理性近视病灶检测与分割常规赛的重点是研究和发展与病理性近视诊断和患者眼底照片病变分割相关的算法.该常规赛的目标是评估和比较在一个 ...
- 常规赛:PALM眼底彩照视盘探测与分割202105-202205全时段第一名(得分0.97387)方案
本来该比赛博主卷到写上一篇博文0.97123是不打算在卷了的,事情的起因也很简单博主4月底参加这个比赛的时候提交了结果,成绩0.96689很一般.不料,到了月底有人(202105-202204月份的全 ...
- PaddleOCR实践之飞桨常规赛:中文场景文字识别
简介 本项目是参加飞桨常规赛:中文场景文字识别(已结束)的项目,项目score为85.87141. 生成的预测文件为work/PaddleOCR中的test2.txt文件 项目任务为识别包含中文文字的 ...
最新文章
- Redis 官方可视化工具,功能强大、干净又卫生!
- 创建支持nginx服务的docker镜像
- java.security.key jar_异常: java.security.InvalidKeyException: Illegal key size
- VSCode 设置 Tab 空格
- ec6108v9a精简刷机包_华为悦盒无安装限制固件下载|华为悦盒EC6108V9A第三方精简流畅无安装限制固件 下载 - 巴士下载站...
- 加拿大卡尔顿大学两个月进阶java—2
- Phyton学习笔记
- 物联网操作系统碎片化问题分析与思考
- 高薪!OPPO研究院招高级算法工程师、专家(可直接内推)
- html文字显示为单行,双行
- Elasticsearch 查询时 判断不为null或不为空字符串
- 1304 佳佳的斐波那契(矩阵乘法)
- mac apache加载php,Apache 2.4没有加载php5.5与Mac OS 10.8
- spring-data-elasticsearch使用Sort排序时Please use a keyword field instead. ……异常解决
- JAVA学习路01——使用JAVA将“.txt”文件拆分、合并
- 腾讯云服务器突然远程连不上
- 什么是swagger以及swagger注解详解
- 一.无线电能传输技术简介
- 10个最佳iOS Map App模板
- Radmin通过互联网联机