前言

上一节学习了以TensorFlow为底端的keras接口最简单的使用,这里就继续学习怎么写卷积分类模型和各种保存方法(仅保存权重、权重和网络结构同时保存)

国际惯例,参考博客:
官方教程
【注】其实不用看博客,直接翻到文末看我的colab就行,里面涵盖了学习方法,包括自己的出错内容和一些简单笔记,下面为了展示方便,每次都重新定义了网络结构,对Python熟悉的大佬可以直接def create_model():函数,把模型结构保存起来,后面直接调用就行

构建卷积模型分类

回顾一下上篇博客介绍的构建模型方法,有两种写法:

model = keras.models.Sequential([keras.layers.Flatten(...),keras.layers.Dense(...),...
])
model = keras.models.Sequential()
model.add(keras.layers.Flatten(...))
model.add(keras.layers.Dense(...))
])

第一种简单,第二种舒服,本博文采用第二种写法构建一个简单的卷积网络

引入相关包

保存模型需要路径(引入os),数据归一化处理(引入numpy),此外注意,虽然我们学习keras,但是不仅要引入keras,还得引入tensorflow,具体原因后续再说

import tensorflow as tf
from tensorflow import keras
import numpy as np
import matplotlib.pyplot as plt
import os

构建数据集

还是用mnist吧,后续根据需要出一个训练本地图片数据的教程,看看是不是还得数据流操作
注意,要把标签改为单热度编码格式,数据也得归一化

mnist_dataset = keras.datasets.mnist
(train_x,train_y),(test_x,test_y)= mnist_dataset.load_data()
train_y = keras.utils.to_categorical(train_y,10)
test_y = keras.utils.to_categorical(test_y,10)
train_x = train_x / 255.0
test_x = test_x / 255.0

还得注意就是keras的卷积操作接受的数据是一个思维矩阵,需要指定是channels_first即<样本,通道,行,列>, 还是channels_last即<样本,行,列,通道>,默认最后的维度是通道(channels_last)

train_x = train_x[ ..., np.newaxis ]
test_x = test_x[..., np.newaxis ]
print(train_x.shape)#(60000, 28, 28, 1)

构建模型

构建简单的AlexNet,但是直接用这个结构可能有问题,因为输入图片总共28\*28,经过多次卷积池化会越变越小,最后可能都不够做卷积池化了,为稍微改了改

model = keras.models.Sequential()
model.add( keras.layers.Conv2D( filters = 64, kernel_size=(11,11),strides = (1,1), padding='valid', activation= tf.keras.activations.relu) )
model.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))
model.add( keras.layers.Conv2D( filters = 192, kernel_size=(5,5),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))
model.add( keras.layers.Conv2D( filters = 384, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model.add( keras.layers.Conv2D( filters = 384, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model.add( keras.layers.Conv2D( filters = 256, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))model.add( keras.layers.Flatten() )
model.add( keras.layers.Dense( units=4096, activation= keras.activations.relu ) )
model.add( keras.layers.Dropout(rate=0.5) )
model.add( keras.layers.Dense( units=4096, activation= keras.activations.relu ) )
model.add( keras.layers.Dropout(rate=0.5) )model.add( keras.layers.Dense(units=10 , activation= keras.activations.softmax ) )

编译和训练模型

keras中关于交叉熵分类有两个函数sparse_categorical_crossentropycategorical_crossentropy,这里就出现了第一个坑,如果将标签[batch_size,10]输入到编译器,使用sparse_...的时候回报错

logits and labels must have the same first dimension,got logits shape [200,10] and labels shape [2000]

好像是默认把他拉长拼起来了,所以我们要使用后者

model.compile( optimizer= keras.optimizers.Adam(),loss= keras.losses.categorical_crossentropy, metrics=['accuracy'] )

然后就可以训练模型了

model.fit(train_x,train_y,epochs=2, batch_size=200)
'''
Epoch 1/2
60000/60000 [==============================] - 26s 435us/step - loss: 0.2646 - acc: 0.9110
Epoch 2/2
60000/60000 [==============================] - 24s 407us/step - loss: 0.0510 - acc: 0.9855
<tensorflow.python.keras.callbacks.History at 0x7f65fe2de940>
'''

还能看网络结构和参数,使用summary()函数

model.summary()
'''
_________________________________________________________________
Layer (type)                 Output Shape              Param #
=================================================================
conv2d (Conv2D)              multiple                  7808
_________________________________________________________________
max_pooling2d (MaxPooling2D) multiple                  0
_________________________________________________________________
conv2d_1 (Conv2D)            multiple                  307392
_________________________________________________________________
max_pooling2d_1 (MaxPooling2 multiple                  0
_________________________________________________________________
conv2d_2 (Conv2D)            multiple                  663936
_________________________________________________________________
conv2d_3 (Conv2D)            multiple                  1327488
_________________________________________________________________
conv2d_4 (Conv2D)            multiple                  884992
_________________________________________________________________
max_pooling2d_2 (MaxPooling2 multiple                  0
_________________________________________________________________
flatten (Flatten)            multiple                  0
_________________________________________________________________
dense (Dense)                multiple                  4198400
_________________________________________________________________
dropout (Dropout)            multiple                  0
_________________________________________________________________
dense_1 (Dense)              multiple                  16781312
_________________________________________________________________
dropout_1 (Dropout)          multiple                  0
_________________________________________________________________
dense_2 (Dense)              multiple                  40970
=================================================================
Total params: 24,212,298
Trainable params: 24,212,298
Non-trainable params: 0
_________________________________________________________________
'''

可以用测试集评估模型

print(test_x.shape, test_y.shape)
model.evaluate( test_x ,test_y)
'''
(10000, 28, 28, 1) (10000, 10)
10000/10000 [==============================] - 3s 283us/step
[0.03784323987539392, 0.9897]
'''

还能预测单张图片,但是要注意输入的第一个维度是样本数, 记得增加一个维度

test_img_idx = 1000
test_img = test_x[test_img_idx,...]
test_img= test_img[np.newaxis,...]
img_prob = model.predict( test_img )plt.figure()
plt.imshow( np.squeeze(test_img) )
plt.title(img_prob.argmax())

保存模型

训练过程中保存检查点

函数为keras.callbacks.ModelCheckpoint

checkpoint_path='./train_save/mnist.ckpt'
checkpoint_dir= os.path.dirname(checkpoint_path)
# 创建检查回调点
cp_callback= keras.callbacks.ModelCheckpoint( checkpoint_path, save_weights_only= True, verbose=1 )model.fit(train_x, train_y,epochs=2, validation_data=(test_x,test_y), callbacks=[cp_callback] )
'''
Train on 60000 samples, validate on 10000 samples
Epoch 1/2
59968/60000 [============================>.] - ETA: 0s - loss: 0.1442 - acc: 0.9681
Epoch 00001: saving model to ./train_save/mnist.ckpt
WARNING:tensorflow:This model was compiled with a Keras optimizer (<tensorflow.python.keras.optimizers.Adam object at 0x7faff48fae80>) but is being saved in TensorFlow format with `save_weights`. The model's weights will be saved, but unlike with TensorFlow optimizers in the TensorFlow format the optimizer's state will not be saved.Consider using a TensorFlow optimizer from `tf.train`.
60000/60000 [==============================] - 93s 2ms/step - loss: 0.1442 - acc: 0.9681 - val_loss: 0.0693 - val_acc: 0.9811
Epoch 2/2
59968/60000 [============================>.] - ETA: 0s - loss: 0.0757 - acc: 0.9840
Epoch 00002: saving model to ./train_save/mnist.ckpt
WARNING:tensorflow:This model was compiled with a Keras optimizer (<tensorflow.python.keras.optimizers.Adam object at 0x7faff48fae80>) but is being saved in TensorFlow format with `save_weights`. The model's weights will be saved, but unlike with TensorFlow optimizers in the TensorFlow format the optimizer's state will not be saved.Consider using a TensorFlow optimizer from `tf.train`.
60000/60000 [==============================] - 92s 2ms/step - loss: 0.0757 - acc: 0.9839 - val_loss: 0.0489 - val_acc: 0.9876
<tensorflow.python.keras.callbacks.History at 0x7fafead25f98>
'''

发现有个warnning,意思是说模型使用的是keras的优化器,保存以后不是tensorflow能直接使用的模型格式,好像少了个状态,需要使用tensorflow自带的优化器,好吧,调整代码

import os
model.compile(optimizer = tf.train.AdamOptimizer(),loss = keras.losses.categorical_crossentropy, metrics=['accuracy'] )checkpoint_path='./train_save2/mnist.ckpt'
checkpoint_dir= os.path.dirname(checkpoint_path)
# 创建检查回调点
cp_callback= keras.callbacks.ModelCheckpoint( checkpoint_path, save_weights_only= True, verbose=1 )model.fit(train_x, train_y,epochs=2, validation_data=(test_x,test_y), callbacks=[cp_callback] )
'''
Train on 60000 samples, validate on 10000 samples
Epoch 1/2
59936/60000 [============================>.] - ETA: 0s - loss: 0.2241 - acc: 0.9325
Epoch 00001: saving model to ./train_save2/mnist.ckpt
60000/60000 [==============================] - 60s 1ms/step - loss: 0.2239 - acc: 0.9326 - val_loss: 0.1009 - val_acc: 0.9765
Epoch 2/2
59936/60000 [============================>.] - ETA: 0s - loss: 0.0866 - acc: 0.9801
Epoch 00002: saving model to ./train_save2/mnist.ckpt
60000/60000 [==============================] - 56s 930us/step - loss: 0.0867 - acc: 0.9800 - val_loss: 0.0591 - val_acc: 0.9855
<tensorflow.python.keras.callbacks.History at 0x7fad1a90d7b8>
'''

这回没出错了,尝试构建一个没训练的模型,将参数载入进来

model_test = keras.models.Sequential()
model_test.add( keras.layers.Conv2D( filters = 64, kernel_size=(11,11),strides = (1,1), padding='valid', activation= tf.keras.activations.relu) )
model_test.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))
model_test.add( keras.layers.Conv2D( filters = 192, kernel_size=(5,5),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))
model_test.add( keras.layers.Conv2D( filters = 384, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test.add( keras.layers.Conv2D( filters = 384, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test.add( keras.layers.Conv2D( filters = 256, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))model_test.add( keras.layers.Flatten() )
model_test.add( keras.layers.Dense( units=4096, activation= keras.activations.relu ) )
model_test.add( keras.layers.Dropout(rate=0.5) )
model_test.add( keras.layers.Dense( units=4096, activation= keras.activations.relu ) )
model_test.add( keras.layers.Dropout(rate=0.5) )model_test.add( keras.layers.Dense(units=10 , activation= keras.activations.softmax ) )model_test.compile(optimizer = tf.train.RMSPropOptimizer(learning_rate=0.01),loss = keras.losses.categorical_crossentropy, metrics=['accuracy'] )

载入最近的检查点

! ls train_save
latest = tf.train.latest_checkpoint('train_save2')#checkpoint  mnist.ckpt.data-00000-of-00001  mnist.ckpt.index
type(latest)#str
loss,acc = model_test.evaluate(test_x,test_y)
print("未载入权重时:准确率{:5.2f}%".format(100*acc))
model_test.load_weights(latest)
loss,acc = model_test.evaluate(test_x,test_y)
print("载入权重时:准确率{:5.2f}%".format(100*acc))
'''
10000/10000 [==============================] - 3s 308us/step
未载入权重时:准确率 9.60%
10000/10000 [==============================] - 3s 264us/step
载入权重时:准确率98.60%
'''

间隔保存

也可以指定多少次训练保存一次检查点,这样能够有效防止过拟合,以后自己可以挑选比较好的训练参数

checkpoint_path="train_save3/cp-{epoch:04d}.ckpt"
checkpoint_dir = os.path.dirname(checkpoint_path)
cp_callback = keras.callbacks.ModelCheckpoint(checkpoint_path,verbose=1, save_weights_only=True, period=1)
model.fit(train_x,train_y,epochs=2, callbacks=[cp_callback], validation_data=[test_x,test_y],verbose=1)
'''
Train on 60000 samples, validate on 10000 samples
Epoch 1/2
59936/60000 [============================>.] - ETA: 0s - loss: 0.0447 - acc: 0.9897
Epoch 00001: saving model to train_save3/cp-0001.ckpt
60000/60000 [==============================] - 61s 1ms/step - loss: 0.0446 - acc: 0.9897 - val_loss: 0.0421 - val_acc: 0.9920
Epoch 2/2
59936/60000 [============================>.] - ETA: 0s - loss: 0.0478 - acc: 0.9885
Epoch 00002: saving model to train_save3/cp-0002.ckpt
60000/60000 [==============================] - 61s 1ms/step - loss: 0.0478 - acc: 0.9884 - val_loss: 0.0590 - val_acc: 0.9859
<tensorflow.python.keras.callbacks.History at 0x7fafea497dd8>
'''

重新构建一个未训练的模型,调用第一次的训练结果

model_test1 = keras.models.Sequential()
model_test1.add( keras.layers.Conv2D( filters = 64, kernel_size=(11,11),strides = (1,1), padding='valid', activation= tf.keras.activations.relu) )
model_test1.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))
model_test1.add( keras.layers.Conv2D( filters = 192, kernel_size=(5,5),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test1.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))
model_test1.add( keras.layers.Conv2D( filters = 384, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test1.add( keras.layers.Conv2D( filters = 384, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test1.add( keras.layers.Conv2D( filters = 256, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test1.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))model_test1.add( keras.layers.Flatten() )
model_test1.add( keras.layers.Dense( units=4096, activation= keras.activations.relu ) )
model_test1.add( keras.layers.Dropout(rate=0.5) )
model_test1.add( keras.layers.Dense( units=4096, activation= keras.activations.relu ) )
model_test1.add( keras.layers.Dropout(rate=0.5) )model_test1.add( keras.layers.Dense(units=10 , activation= keras.activations.softmax ) )model_test1.compile(optimizer = tf.train.AdamOptimizer(learning_rate=0.01),loss = keras.losses.categorical_crossentropy, metrics=['accuracy'] )

选择第一个检查点载入

loss,acc=model_test1.evaluate(test_x,test_y)
print("未载入权重时:准确率{:5.2f}%".format(100*acc))
model_test1.load_weights("train_save3/cp-0001.ckpt")
loss,acc=model_test1.evaluate(test_x,test_y)
print("载入权重时:准确率{:5.2f}%".format(100*acc))
'''
10000/10000 [==============================] - 3s 260us/step
未载入权重时:准确率10.28%
10000/10000 [==============================] - 2s 244us/step
载入权重时:准确率98.30%
'''

手动保存模型

在训练完毕以后,也可以自行调用save_weights函数保存权重

model.save_weights('./train_save3/mnist_checkpoint')

构建未训练模型

model_test2 = keras.models.Sequential()
model_test2.add( keras.layers.Conv2D( filters = 64, kernel_size=(11,11),strides = (1,1), padding='valid', activation= tf.keras.activations.relu) )
model_test2.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))
model_test2.add( keras.layers.Conv2D( filters = 192, kernel_size=(5,5),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test2.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))
model_test2.add( keras.layers.Conv2D( filters = 384, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test2.add( keras.layers.Conv2D( filters = 384, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test2.add( keras.layers.Conv2D( filters = 256, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test2.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))model_test2.add( keras.layers.Flatten() )
model_test2.add( keras.layers.Dense( units=4096, activation= keras.activations.relu ) )
model_test2.add( keras.layers.Dropout(rate=0.5) )
model_test2.add( keras.layers.Dense( units=4096, activation= keras.activations.relu ) )
model_test2.add( keras.layers.Dropout(rate=0.5) )model_test2.add( keras.layers.Dense(units=10 , activation= keras.activations.softmax ) )model_test2.compile(optimizer = tf.train.AdamOptimizer(learning_rate=0.01),loss = keras.losses.categorical_crossentropy, metrics=['accuracy'] )

读取权重以及评估模型

loss,acc = model_test2.evaluate(test_x,test_y)
print("未载入权重时:准确率{:5.2f}%".format(100*acc))
model_test2.load_weights('./train_save3/mnist_checkpoint')
loss,acc = model_test2.evaluate(test_x,test_y)
print("载入权重时:准确率{:5.2f}%".format(100*acc))
'''
10000/10000 [==============================] - 3s 303us/step
未载入权重时:准确率12.15%
10000/10000 [==============================] - 3s 260us/step
载入权重时:准确率98.59%
'''

全部保存

同时保存模型与参数
构建未训练模型

model_test3 = keras.models.Sequential()
model_test3.add( keras.layers.Conv2D( filters = 64, kernel_size=(11,11),strides = (1,1), padding='valid', activation= tf.keras.activations.relu) )
model_test3.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))
model_test3.add( keras.layers.Conv2D( filters = 192, kernel_size=(5,5),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test3.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))
model_test3.add( keras.layers.Conv2D( filters = 384, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test3.add( keras.layers.Conv2D( filters = 384, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test3.add( keras.layers.Conv2D( filters = 256, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test3.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))model_test3.add( keras.layers.Flatten() )
model_test3.add( keras.layers.Dense( units=4096, activation= keras.activations.relu ) )
model_test3.add( keras.layers.Dropout(rate=0.5) )
model_test3.add( keras.layers.Dense( units=4096, activation= keras.activations.relu ) )
model_test3.add( keras.layers.Dropout(rate=0.5) )model_test3.add( keras.layers.Dense(units=10 , activation= keras.activations.softmax ) )model_test3.compile(optimizer = tf.train.AdamOptimizer(),loss = keras.losses.categorical_crossentropy, metrics=['accuracy'] )
model_test3.fit(train_x,train_y,batch_size=200,epochs=2)

保存

model_test3.save('my_model.h5')
'''
Currently `save` requires model to be a graph network. Consider using `save_weights`, in order to save the weights of the model.
'''

出现错误,意思是需要定义的模型是一个图网络结构,只能保存权重。其实错误原因在于我们的第一层没有定义输入的大小,尝试定义一波

model_test4 = keras.models.Sequential()
model_test4.add( keras.layers.Conv2D( filters = 64, kernel_size=(11,11),strides = (1,1), padding='valid', activation= tf.keras.activations.relu,input_shape=(28,28,1)) )
model_test4.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))
model_test4.add( keras.layers.Conv2D( filters = 192, kernel_size=(5,5),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test4.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))
model_test4.add( keras.layers.Conv2D( filters = 384, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test4.add( keras.layers.Conv2D( filters = 384, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test4.add( keras.layers.Conv2D( filters = 256, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test4.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))model_test4.add( keras.layers.Flatten() )
model_test4.add( keras.layers.Dense( units=4096, activation= keras.activations.relu ) )
model_test4.add( keras.layers.Dropout(rate=0.5) )
model_test4.add( keras.layers.Dense( units=4096, activation= keras.activations.relu ) )
model_test4.add( keras.layers.Dropout(rate=0.5) )model_test4.add( keras.layers.Dense(units=10 , activation= keras.activations.softmax ) )model_test4.compile(optimizer = tf.train.AdamOptimizer(),loss = keras.losses.categorical_crossentropy, metrics=['accuracy'] )
model_test4.fit(train_x,train_y,batch_size=200,epochs=2)
'''
Epoch 1/2
60000/60000 [==============================] - 22s 364us/step - loss: 0.4112 - acc: 0.8556
Epoch 2/2
60000/60000 [==============================] - 21s 342us/step - loss: 0.0584 - acc: 0.9838
<tensorflow.python.keras.callbacks.History at 0x7f137972f4a8>
'''

尝试保存

model_test4.save('my_model.h5')
'''
WARNING:tensorflow:TensorFlow optimizers do not make it possible to access optimizer attributes or optimizer state after instantiation. As a result, we cannot save the optimizer as part of the model save file.You will have to compile your model again after loading it. Prefer using a Keras optimizer instead (see keras.io/optimizers).
'''

又出warning,说是不能使用tensorflow的优化器,要使用keras自带的优化器,好吧,改

model_test5 = keras.models.Sequential()
model_test5.add( keras.layers.Conv2D( filters = 64, kernel_size=(11,11),strides = (1,1), padding='valid', activation= tf.keras.activations.relu,input_shape=(28,28,1)) )
model_test5.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))
model_test5.add( keras.layers.Conv2D( filters = 192, kernel_size=(5,5),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test5.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))
model_test5.add( keras.layers.Conv2D( filters = 384, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test5.add( keras.layers.Conv2D( filters = 384, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test5.add( keras.layers.Conv2D( filters = 256, kernel_size=(3,3),strides = (1,1), padding='same', activation= tf.keras.activations.relu) )
model_test5.add( keras.layers.MaxPool2D( pool_size=(2,2),strides=(2,2) ))model_test5.add( keras.layers.Flatten() )
model_test5.add( keras.layers.Dense( units=4096, activation= keras.activations.relu ) )
model_test5.add( keras.layers.Dropout(rate=0.5) )
model_test5.add( keras.layers.Dense( units=4096, activation= keras.activations.relu ) )
model_test5.add( keras.layers.Dropout(rate=0.5) )model_test5.add( keras.layers.Dense(units=10 , activation= keras.activations.softmax ) )model_test5.compile(optimizer = tf.keras.optimizers.Adam(),loss = keras.losses.categorical_crossentropy, metrics=['accuracy'] )
model_test5.fit(train_x,train_y,batch_size=200,epochs=2)
model_test5.save("my_model2.h5")
'''
Epoch 1/2
60000/60000 [==============================] - 26s 434us/step - loss: 0.2850 - acc: 0.9043
Epoch 2/2
60000/60000 [==============================] - 25s 409us/step - loss: 0.0555 - acc: 0.9847
<tensorflow.python.keras.callbacks.History at 0x7f661033a9e8>
'''

这样就不出错了,尝试调用模型和参数,因为保存了模型结构和参数,所以不需要重新定义网络结构

model_test6= keras.models.load_model("my_model2.h5")
model_test6.summary()
'''
_________________________________________________________________
Layer (type)                 Output Shape              Param #
=================================================================
conv2d_10 (Conv2D)           (None, 18, 18, 64)        7808
_________________________________________________________________
max_pooling2d_6 (MaxPooling2 (None, 9, 9, 64)          0
_________________________________________________________________
conv2d_11 (Conv2D)           (None, 9, 9, 192)         307392
_________________________________________________________________
max_pooling2d_7 (MaxPooling2 (None, 4, 4, 192)         0
_________________________________________________________________
conv2d_12 (Conv2D)           (None, 4, 4, 384)         663936
_________________________________________________________________
conv2d_13 (Conv2D)           (None, 4, 4, 384)         1327488
_________________________________________________________________
conv2d_14 (Conv2D)           (None, 4, 4, 256)         884992
_________________________________________________________________
max_pooling2d_8 (MaxPooling2 (None, 2, 2, 256)         0
_________________________________________________________________
flatten_2 (Flatten)          (None, 1024)              0
_________________________________________________________________
dense_6 (Dense)              (None, 4096)              4198400
_________________________________________________________________
dropout_4 (Dropout)          (None, 4096)              0
_________________________________________________________________
dense_7 (Dense)              (None, 4096)              16781312
_________________________________________________________________
dropout_5 (Dropout)          (None, 4096)              0
_________________________________________________________________
dense_8 (Dense)              (None, 10)                40970
=================================================================
Total params: 24,212,298
Trainable params: 24,212,298
Non-trainable params: 0
_________________________________________________________________
'''

稳如狗,做一下测试

  • 测试集上的测试
model_test6.evaluate(test_x,test_y)
'''
10000/10000 [==============================] - 3s 302us/step
[0.05095283883444499, 0.9868]
'''
  • 单张图片的测试
test_img= test_x[5000,...]
test_img=test_img[ np.newaxis,...]
pred_label = model_test6.predict_classes(test_img)
plt.figure()
plt.imshow( np.squeeze(test_img))
plt.title(pred_label)

后记

这一章主要学习了如何搭建简单的卷积网络,以及集中保存方法:仅权重以及权重和模型结构。
主要记住的就是如果仅保存权重,注意用tensorflow自带的优化器,而保存网络和权重的时候要用keras的优化器
下一章针对深度学习的几个理论做一下理解以及实验,包括BatchNormResNet等。
博客代码链接:戳这里

【TensorFlow-windows】keras接口——卷积手写数字识别,模型保存和调用相关推荐

  1. TensorFlow高阶 API: keras教程-使用tf.keras搭建mnist手写数字识别网络

    TensorFlow高阶 API:keras教程-使用tf.keras搭建mnist手写数字识别网络 目录 TensorFlow高阶 API:keras教程-使用tf.keras搭建mnist手写数字 ...

  2. 使用tf.keras搭建mnist手写数字识别网络

    使用tf.keras搭建mnist手写数字识别网络 目录 使用tf.keras搭建mnist手写数字识别网络 1.使用tf.keras.Sequential搭建序列模型 1.1 tf.keras.Se ...

  3. python cnn代码详解图解_基于TensorFlow的CNN实现Mnist手写数字识别

    本文实例为大家分享了基于TensorFlow的CNN实现Mnist手写数字识别的具体代码,供大家参考,具体内容如下 一.CNN模型结构 输入层:Mnist数据集(28*28) 第一层卷积:感受视野5* ...

  4. 基于TensorFlow和mnist数据集的手写数字识别系统 ,可识别电话号码,识别准确率高,有对比实验,两组模型,可讲解代码

    基于TensorFlow和mnist数据集的手写数字识别系统 ,可识别电话号码,识别准确率高,有对比实验,两组模型,可讲解代码

  5. TensorFlow笔记(3)——利用TensorFlow和MNIST数据集训练一个最简单的手写数字识别模型...

    前言 当我们开始学习编程的时候,第一件事往往是学习打印"Hello World".就好比编程入门有Hello World,机器学习入门有MNIST. MNIST是一个入门级的计算机 ...

  6. 持久化的基于L2正则化和平均滑动模型的MNIST手写数字识别模型

    持久化的基于L2正则化和平均滑动模型的MNIST手写数字识别模型 觉得有用的话,欢迎一起讨论相互学习~Follow Me 参考文献Tensorflow实战Google深度学习框架 实验平台: Tens ...

  7. 在MNIST数据集上训练一个手写数字识别模型

    使用Pytorch在MNIST数据集上训练一个手写数字识别模型, 代码和参数文件 可下载 1.1 数据下载 import torchvision as tvtraining_sets = tv.dat ...

  8. 使用飞桨完成手写数字识别模型

    手写数字识别任务 数字识别是计算机从纸质文档.照片或其他来源接收.理解并识别可读的数字的能力,目前比较受关注的是手写数字识别.手写数字识别是一个典型的图像分类问题,已经被广泛应用于汇款单号识别.手写邮 ...

  9. 深度学习入门实例——基于keras的mnist手写数字识别

    本文介绍了利用keras做mnist数据集的手写数字识别. 参考网址 http://www.cnblogs.com/lc1217/p/7132364.html mnist数据集中的图片为28*28的单 ...

最新文章

  1. HTTP.sys 远程执行代码验证工具
  2. sleep 与 wait 区别
  3. [原]NYOJ-开灯问题-77
  4. 一台服务器装两个sql server 如何访问吗_服务器多网卡多路由策略
  5. 用纸筒做机器人_365天,每天都可以玩出新花样,卷纸筒就是这么牛的神器!
  6. pagehelper的使用_SpringBoot项目中,如何更规范的使用PageHelper分页?
  7. freemaker中小数展示为整数的问题
  8. 【mysql】显式加锁
  9. 通过skimmer发动供应链攻击,苏富比地产100多个网站受影响
  10. 绘图神器 —— Graphviz 绘制数据结构相关图形
  11. mt4 指标 涨跌幅 颜色k线_精品主图 精准K线买卖点提示通达信指标公式源码
  12. 时间戳转化为YYYY-MM-DD-XX-MM-UU日期格式
  13. 调用java webservice 报错_java调用webservice方法报错
  14. Win10 搭建自带的ftp站点
  15. 投票系统(投票问题可多选、带柱状图)
  16. module not specified如何解决
  17. 解决微擎人人商城分销中心无法生成推广二维码、商品二维码的问题
  18. 《R语言数据分析》作业答案
  19. 『NLP学习笔记』Cross Entropy Loss 的硬截断、软化到 Focal Loss
  20. 谷鸟人工智能选股系统技术文档

热门文章

  1. 树莓派蓝屏_树莓派刷写Windows 10 ARM版后成功运行桌面程序
  2. 16年蓝桥杯省赛-方格填数(dfs+回溯)
  3. 华为鸿蒙分布式系统2020,鸿蒙2.0来了!华为开发者大会HDC 2020宣布
  4. 数据结构--栈(附上STL栈)
  5. Codeforce 322E Ciel the Commander (点分治)
  6. C++--浅谈开发系统的经验
  7. FSAF 让网络自己决定实例输出层
  8. 钽电容极性 钽电容极性判断图解
  9. python logging之multi-module
  10. 互联网产品的灰度发布