栏目分类:
子分类:
返回
名师互学网用户登录
快速导航关闭
当前搜索
当前分类
子分类
实用工具
热门搜索
名师互学网 > IT > 软件开发 > 后端开发 > Python

TensorFlow 从入门到精通(3)—— 手写数字识别

Python 更新时间: 发布时间: IT归档 最新发布 模块sitemap 名妆网 法律咨询 聚返吧 英语巴士网 伯小乐 网商动力

TensorFlow 从入门到精通(3)—— 手写数字识别

import tensorflow as tfprint(tf.__version__)
2.6.0

一、数据集

mnist = tf.keras.datasets.mnist(train_images,train_labels),(test_images,test_labels) = mnist.load_data()
train_images.shape,train_labels.shape
((60000, 28, 28), (60000,))
# 可视化imageimport matplotlib.pyplot as pltdef plot_image(image):    plt.imshow(image.reshape(28,28),cmap='binary')    plt.show()
plot_image(train_images[0])

# 处理数据集:将数据集拉成二维的x_data = tf.cast(train_images.reshape(-1,28*28),dtype=tf.float32)# 划分数据集valid_percent = 0.2train_num = int(x_data.shape[0] * (1-valid_percent))x_train = x_data[:train_num] # 训练集y_train = train_labels[:train_num]x_valid = x_data[train_num:] # 验证集y_valid = train_labels[train_num:]x_test = tf.cast(test_images.reshape(-1,28*28),dtype=tf.float32) # 测试集y_test = test_labels
x_data.shape[0] * 0.8
48000.0
# 归一化x_train = x_train / 255.0x_valid = x_valid / 255.0x_test = x_test / 255.0
# 对标签进行独热编码y_train = tf.one_hot(y_train,depth=10)y_valid = tf.one_hot(y_valid,depth=10)y_test = tf.one_hot(y_test,depth=10)

二、模型

W = tf.Variable(tf.random.normal(shape=(28*28,10)))B = tf.Variable(tf.zeros(10))
def model(x,w,b):    model_ = tf.matmul(x,w) + b    return tf.nn.softmax(model_)
def loss(x,y,w,b):    pred = model(x,w,b)    loss_ = tf.keras.losses.categorical_crossentropy(y_true=y,y_pred=pred)    return tf.reduce_mean(loss_)
def grad(x,y,w,b):    with tf.GradientTape() as tape:        loss_ = loss(x,y,w,b)    return tape.gradient(loss_,[w,b])
# 准确率def accuracy(x,y,w,b):  pred = model(x,w,b)  # print(pred,y)  corrent_prediction = tf.equal(tf.argmax(pred,1),tf.argmax(y,1))  # print(corrent_prediction,corrent_prediction.shape)  return tf.reduce_mean(tf.cast(corrent_prediction,tf.float32))

三、训练

training_epochs = 20learining_rate = 0.001batch_size = 50
total_step = train_num // batch_sizeoptimizer = tf.keras.optimizers.Adam(learining_rate)train_loss_list = []valid_loss_list = []train_acc_list = []valid_acc_list = []for epoch in range(training_epochs):    for step in range(total_step):        xs = x_train[step*batch_size:(step+1)*batch_size]        ys = y_train[step*batch_size:(step+1)*batch_size]#         print(xs.shape,ys.shape)        grads = grad(xs,ys,W,B)        optimizer.apply_gradients(zip(grads,[W,B]))    train_loss = loss(x_train,y_train,W,B).numpy()    valid_loss = loss(x_valid,y_valid,W,B).numpy()    train_acc = accuracy(x_train,y_train,W,B).numpy()    valid_acc = accuracy(x_valid,y_valid,W,B).numpy()    train_acc_list.append(train_acc)    valid_acc_list.append(valid_acc)    train_loss_list.append(train_loss)    valid_loss_list.append(valid_loss)    print(f"{epoch+1}:train_loss:{train_loss}train_acc:{train_acc}valid_loss:{valid_loss}valid_acc:{valid_acc}")
1:train_loss:1.5718567371368408train_acc:0.706458330154419valid_loss:1.4870891571044922valid_acc:0.71974998712539672:train_loss:0.9688711762428284train_acc:0.799958348274231valid_loss:0.9067797660827637valid_acc:0.81441664695739753:train_loss:0.7616745829582214train_acc:0.8380833268165588valid_loss:0.7226295471191406valid_acc:0.84616667032241824:train_loss:0.6529313921928406train_acc:0.8577708601951599valid_loss:0.6307686567306519valid_acc:0.86374998092651375:train_loss:0.5832834839820862train_acc:0.8693541884422302valid_loss:0.5737946629524231valid_acc:0.87291663885116586:train_loss:0.5349117517471313train_acc:0.8781874775886536valid_loss:0.5341132283210754valid_acc:0.87949997186660777:train_loss:0.49877774715423584train_acc:0.8850208520889282valid_loss:0.504259467124939valid_acc:0.88458335399627698:train_loss:0.4704902470111847train_acc:0.8897500038146973valid_loss:0.4805634021759033valid_acc:0.88824999332427989:train_loss:0.4476514160633087train_acc:0.8929374814033508valid_loss:0.4613181948661804valid_acc:0.8904166817665110:train_loss:0.42855820059776306train_acc:0.8967708349227905valid_loss:0.44507089257240295valid_acc:0.893333315849304211:train_loss:0.41240566968917847train_acc:0.8991041779518127valid_loss:0.4314454197883606valid_acc:0.895416676998138412:train_loss:0.39855533838272095train_acc:0.9014999866485596valid_loss:0.4198707342147827valid_acc:0.897333323955535913:train_loss:0.38649922609329224train_acc:0.9040833115577698valid_loss:0.40982159972190857valid_acc:0.899083316326141414:train_loss:0.3758573532104492train_acc:0.9057083129882812valid_loss:0.40100958943367004valid_acc:0.901333332061767615:train_loss:0.3664068877696991train_acc:0.9076458215713501valid_loss:0.3931334316730499valid_acc:0.902416646480560316:train_loss:0.35793840885162354train_acc:0.9091874957084656valid_loss:0.3861088752746582valid_acc:0.903750002384185817:train_loss:0.3502780497074127train_acc:0.9105416536331177valid_loss:0.37972745299339294valid_acc:0.904999971389770518:train_loss:0.34334996342658997train_acc:0.9118541479110718valid_loss:0.3739684522151947valid_acc:0.90591669082641619:train_loss:0.3370097279548645train_acc:0.9130833148956299valid_loss:0.3686791658401489valid_acc:0.90616667270660420:train_loss:0.33120784163475037train_acc:0.91427081823349valid_loss:0.3638934791088104valid_acc:0.9070833325386047
# 可视化损失import matplotlib.pyplot as pltplt.plot(train_loss_list,'r',label='train')plt.plot(valid_loss_list,'g',label='valid')plt.legend(loc=1)

# 可视化准确率import matplotlib.pyplot as pltplt.plot(train_acc_list,'r',label='trainacc')plt.plot(valid_acc_list,'g',label='validacc')plt.legend(loc=1)

四、预测

# 在测试集上评估模型准确率acc_test = accuracy(x_test,y_test,W,B).numpy()acc_test
0.9087
# 定义预测函数def predict(x,w,b):  pred = model(x,w,b)  res = tf.argmax(pred,1).numpy()  return res
pred_test = predict(x_test,W,B)
# 定义可视化函数import numpy as npimport matplotlib.pyplot as pltdef plot_image_labels_prediction(images, # 图像列表                 labels, # 标签列表                 preds, # 预测值列表                 index=0, # 从第index个开始                 num=10 # 缺省一次显示10幅                 ):  fig = plt.gcf() # 获取当前图表  fig.set_size_inches(10,4) # 1英寸等于2.54cm  if num > 10:    num = 10 # 最多显示10个子图  for i in range(0,num):    ax = plt.subplot(2,5,i+1)    ax.imshow(np.reshape(images[index],(28,28)),cmap='binary') # 显示第index个图像    title = 'label='+str(labels[index]) # 构建该图上要显示的title信息    if len(preds) > 0:      title += ',predict='+str(preds[index])    ax.set_title(title,fontsize=10) # 显示图上的title信息    ax.set_xticks([]) # 不显示坐标轴    ax.set_yticks([])    index = index + 1    plt.show()
plot_image_labels_prediction(test_images,test_labels,pred_test,9000,20)









转载请注明:文章转载自 www.mshxw.com
本文地址:https://www.mshxw.com/it/268484.html
我们一直用心在做
关于我们 文章归档 网站地图 联系我们

版权所有 (c)2021-2022 MSHXW.COM

ICP备案号:晋ICP备2021003244-6号