日韩性视频-久久久蜜桃-www中文字幕-在线中文字幕av-亚洲欧美一区二区三区四区-撸久久-香蕉视频一区-久久无码精品丰满人妻-国产高潮av-激情福利社-日韩av网址大全-国产精品久久999-日本五十路在线-性欧美在线-久久99精品波多结衣一区-男女午夜免费视频-黑人极品ⅴideos精品欧美棵-人人妻人人澡人人爽精品欧美一区-日韩一区在线看-欧美a级在线免费观看

歡迎訪問 生活随笔!

生活随笔

當前位置: 首頁 > 编程资源 > 编程问答 >内容正文

编程问答

02.改善深层神经网络:超参数调试、正则化以及优化 W3. 超参数调试、Batch Norm和程序框架(作业:TensorFlow教程+数字手势预测)

發布時間:2024/7/5 编程问答 34 豆豆

文章目錄

    • 1. 探索TensorFlow庫
      • 1.1 線性函數
      • 1.2 計算 sigmoid
      • 1.3 計算損失函數
      • 1.4 One_Hot 編碼
      • 1.5 用0,1初始化
    • 2. 用TensorFlow建立你的第一個神經網絡
      • 2.0 數字手勢識別
      • 2.1 創建 placeholder
      • 2.2 初始化參數
      • 2.3 前向傳播
      • 2.4 計算損失
      • 2.5 后向傳播、更新參數
      • 2.6 建立完整的TF模型
      • 2.7 用自己的照片測試
    • 總結

測試題:參考博文

筆記:02.改善深層神經網絡:超參數調試、正則化以及優化 W3. 超參數調試、Batch Norm和程序框架

  • 像TensorFlow、Paddle、Torch、Caffe、Keras等機器學習框架可以顯著加快機器學習的發展
  • 神經網絡編程框架 不僅可以縮短編碼時間,有時還可以執行優化加速你的代碼

本作業TensorFlow內容:

  • 初始化變量
  • 定義 session
  • 訓練算法
  • 實現一個神經網絡

1. 探索TensorFlow庫

  • 導入一些庫
import math import numpy as np import h5py import matplotlib.pyplot as plt import tensorflow as tf from tensorflow.python.framework import ops from tf_utils import load_dataset, random_mini_batches, convert_to_one_hot, predict%matplotlib inline np.random.seed(1) import sys sys.path.append(r"/path/file")

比如要計算損失函數:
loss=L(y^,y)=(y^(i)?y(i))2loss = \mathcal{L}(\hat{y}, y) = (\hat y^{(i)} - y^{(i)})^2loss=L(y^?,y)=(y^?(i)?y(i))2

y_hat = tf.constant(36, name='y_hat') # Define y_hat constant. Set to 36.y = tf.constant(39, name='y') # Define y. Set to 39loss = tf.Variable((y - y_hat)**2, name='loss') # Create a variable for the lossinit = tf.global_variables_initializer() # When init is run later (session.run(init)), # the loss variable will be initialized and ready to be computedwith tf.Session() as session: # Create a session and print the outputsession.run(init) # Initializes the variablesprint(session.run(loss)) # Prints the loss

輸出:(36?39)2=9(36-39)^2 = 9(36?39)2=9

9

TensorFlow編程步驟:

  • 創建Tensors(變量)(尚未執行的)
  • 寫出操作方法(訓練之類的)
  • 初始化Tensors
  • 創建Session
  • 運行Session(運行上面的操作方法)
a = tf.constant(2) b = tf.constant(10) c = tf.multiply(a,b) print(c)

輸出:

Tensor("Mul:0", shape=(), dtype=int32)
  • 我們沒有看見 20,看見輸出了一個 Tensor,沒有shape,int32類型
  • 我只是把變量放入了計算圖(computation graph),但是沒有運行
sess = tf.Session() print(sess.run(c))

輸出:20


placeholder:placeholder(dtype, shape=None, name=None)

  • placeholder 是一個值只能稍后給定的對象
  • 當我們運行一個Session時,采用一個字典 (feed_dict ) 給placeholder傳遞數值
# Change the value of x in the feed_dictx = tf.placeholder(tf.int64, name = 'xa') # 定義變量x,稍后才能賦值 print(sess.run(2 * x, feed_dict = {x: 3})) # 運行(計算圖,給placeholder變量喂數據) sess.close()

輸出:6

1.1 線性函數

計算 WX+bWX + bWX+bWWW 是矩陣,bbb 是向量

  • X = tf.constant(np.random.randn(3,1), name = "X")
  • tf.matmul(..., ...) 矩陣乘法
  • tf.add(..., ...) 加法
  • np.random.randn(...) 隨機初始化(n 正態分布)
# GRADED FUNCTION: linear_functiondef linear_function():"""Implements a linear function: Initializes W to be a random tensor of shape (4,3)Initializes X to be a random tensor of shape (3,1)Initializes b to be a random tensor of shape (4,1)Returns: result -- runs the session for Y = WX + b """np.random.seed(1)### START CODE HERE ### (4 lines of code)X = tf.constant(np.random.randn(3,1), name='X')W = tf.constant(np.random.randn(4,3), name='W')b = tf.constant(np.random.randn(4,1), name='b')Y = tf.matmul(W,X)+b# Y = tf.add(tf.matmul(W,X), b) # 也可以# Y = W*X+b # 錯誤寫法!!!!!!!!!!### END CODE HERE ### # Create the session using tf.Session() and run it with sess.run(...) on the variable you want to calculate### START CODE HERE ###sess = tf.Session()result = sess.run(Y)### END CODE HERE ### # close the session sess.close()return result

1.2 計算 sigmoid

  • 定義 placeholder變量 x
  • 定義操作tf.sigmoid(),計算sigmoid值
  • 在session里運行

Method 1:

sess = tf.Session() # Run the variables initialization (if needed), run the operations result = sess.run(..., feed_dict = {...}) sess.close() # Close the session

Method 2: with不用自己寫close,比較方便

with tf.Session() as sess: # run the variables initialization (if needed), run the operationsresult = sess.run(..., feed_dict = {...})# This takes care of closing the session for you :) # GRADED FUNCTION: sigmoiddef sigmoid(z):"""Computes the sigmoid of zArguments:z -- input value, scalar or vectorReturns: results -- the sigmoid of z"""### START CODE HERE ### ( approx. 4 lines of code)# Create a placeholder for x. Name it 'x'.x = tf.placeholder(tf.float32, name='x')# compute sigmoid(x)sigmoid = tf.sigmoid(x)# Create a session, and run it. # Please use the method 2 explained above. # You should use a feed_dict to pass z's value to x. with tf.Session() as sess:# Run session and call the output "result"result = sess.run(sigmoid,feed_dict={x:z})### END CODE HERE ###return result

1.3 計算損失函數

計算損失函數:
J=?1m∑i=1m(y(i)log?a[2](i)+(1?y(i))log?(1?a[2](i)))J = - \frac{1}{m} \sum_{i = 1}^m \bigg (y^{(i)} \log a^{ [2] (i)} + (1-y^{(i)})\log (1-a^{ [2] (i)} )\bigg )J=?m1?i=1m?(y(i)loga[2](i)+(1?y(i))log(1?a[2](i)))

  • TF 內置 tf.nn.sigmoid_cross_entropy_with_logits(logits = ..., labels = ...),它計算以下式子:
    ?1m∑i=1m(y(i)log?σ(z[2](i))+(1?y(i))log?(1?σ(z[2](i)))- \frac{1}{m} \sum_{i = 1}^m \bigg ( y^{(i)} \log \sigma(z^{[2](i)}) + (1-y^{(i)})\log (1-\sigma(z^{[2](i)})\bigg )?m1?i=1m?(y(i)logσ(z[2](i))+(1?y(i))log(1?σ(z[2](i)))
    代碼需要輸入 z,計算 sigmoid 值(得到 a),然后計算交叉熵損失(tf 一行搞定!)
# GRADED FUNCTION: costdef cost(logits, labels):"""Computes the cost using the sigmoid cross entropyArguments:logits -- vector containing z, output of the last linear unit (before the final sigmoid activation)labels -- vector of labels y (1 or 0) Note: What we've been calling "z" and "y" in this class are respectively called "logits" and "labels" in the TensorFlow documentation. So logits will feed into z, and labels into y. Returns:cost -- runs the session of the cost (formula (2))"""### START CODE HERE ### # Create the placeholders for "logits" (z) and "labels" (y) (approx. 2 lines)z = tf.placeholder(tf.float32, name='z')y = tf.placeholder(tf.float32, name='y')# Use the loss function (approx. 1 line)cost = tf.nn.sigmoid_cross_entropy_with_logits(logits=z, labels=y)# Create a session (approx. 1 line). See method 1 above.sess = tf.Session()# Run the session (approx. 1 line).cost = sess.run(cost, feed_dict={z:logits, y:labels})# Close the session (approx. 1 line). See method 1 above.sess.close()### END CODE HERE ###return cost

1.4 One_Hot 編碼

比如有標簽 y 向量,有4種標簽值,要進行編碼:

  • tf.one_hot(labels, depth, axis)
# GRADED FUNCTION: one_hot_matrixdef one_hot_matrix(labels, C):"""Creates a matrix where the i-th row corresponds to the ith class number and the jth columncorresponds to the jth training example. So if example j had a label i. Then entry (i,j) will be 1. Arguments:labels -- vector containing the labels C -- number of classes, the depth of the one hot dimensionReturns: one_hot -- one hot matrix"""### START CODE HERE #### Create a tf.constant equal to C (depth), name it 'C'. (approx. 1 line)C = tf.constant(C, name='C')# Use tf.one_hot, be careful with the axis (approx. 1 line)one_hot_matrix = tf.one_hot(labels, C, axis=0) # 不寫 axis默認為 1, 0 是列向# Create the session (approx. 1 line)sess = tf.Session()# Run the session (approx. 1 line)one_hot = sess.run(one_hot_matrix)# Close the session (approx. 1 line). See method 1 above.sess.close()### END CODE HERE ###return one_hot

1.5 用0,1初始化

  • tf.ones(shape),tf.zeros(shape)
# GRADED FUNCTION: onesdef ones(shape):"""Creates an array of ones of dimension shapeArguments:shape -- shape of the array you want to createReturns: ones -- array containing only ones"""### START CODE HERE #### Create "ones" tensor using tf.ones(...). (approx. 1 line)ones = tf.ones(shape)# Create the session (approx. 1 line)sess = tf.Session()# Run the session to compute 'ones' (approx. 1 line)ones = sess.run(ones)# Close the session (approx. 1 line). See method 1 above.sess.close()### END CODE HERE ###return ones

2. 用TensorFlow建立你的第一個神經網絡

實現TF模型步驟:

  • 創建計算圖
  • 運行圖

2.0 數字手勢識別

  • 訓練集:1080張圖片(64*64像素)(0-5手勢,每種180張)
  • 測試集:120張圖片(64*64像素)(0-5手勢,每種20張)

加載數據集,看一看

# Loading the dataset X_train_orig, Y_train_orig, X_test_orig, Y_test_orig, classes = load_dataset()# Example of a picture index = 1 plt.imshow(X_train_orig[index]) print ("y = " + str(np.squeeze(Y_train_orig[:, index])))

  • 歸一化像素值(除以255),對標簽進行 one_hot 編碼
# Flatten the training and test images X_train_flatten = X_train_orig.reshape(X_train_orig.shape[0], -1).T X_test_flatten = X_test_orig.reshape(X_test_orig.shape[0], -1).T # Normalize image vectors X_train = X_train_flatten/255. X_test = X_test_flatten/255. # Convert training and test labels to one hot matrices Y_train = convert_to_one_hot(Y_train_orig, 6) Y_test = convert_to_one_hot(Y_test_orig, 6)print ("number of training examples = " + str(X_train.shape[1])) print ("number of test examples = " + str(X_test.shape[1])) print ("X_train shape: " + str(X_train.shape)) print ("Y_train shape: " + str(Y_train.shape)) print ("X_test shape: " + str(X_test.shape)) print ("Y_test shape: " + str(Y_test.shape))

輸出:(注意維度正確與否)

number of training examples = 1080 number of test examples = 120 X_train shape: (12288, 1080) # 64*64*3 = 12288 Y_train shape: (6, 1080) X_test shape: (12288, 120) Y_test shape: (6, 120)

下面建立的模型結構為:LINEAR -> RELU -> LINEAR -> RELU -> LINEAR -> SOFTMAX
(多于2種輸出,將 sigmoid 輸出改成 softmax 輸出)

2.1 創建 placeholder

創建 X,Y 的 placeholder,一會運行的時候,給他 feed 訓練數據

# GRADED FUNCTION: create_placeholdersdef create_placeholders(n_x, n_y):"""Creates the placeholders for the tensorflow session.Arguments:n_x -- scalar, size of an image vector (num_px * num_px = 64 * 64 * 3 = 12288)n_y -- scalar, number of classes (from 0 to 5, so -> 6)Returns:X -- placeholder for the data input, of shape [n_x, None] and dtype "float"Y -- placeholder for the input labels, of shape [n_y, None] and dtype "float"Tips:- You will use None because it let's us be flexible on the number of examples you will for the placeholders.In fact, the number of examples during test/train is different."""### START CODE HERE ### (approx. 2 lines)X = tf.placeholder(tf.float32, shape=(n_x, None), name='X')Y = tf.placeholder(tf.float32, shape=(n_y, None), name='Y')### END CODE HERE ###return X, Y

2.2 初始化參數

用 Xavier 初始化權重,0初始化偏置
參考:深度學習中Xavier初始化

  • W1 = tf.get_variable("W1", [25,12288], initializer = tf.contrib.layers.xavier_initializer(seed = 1))
  • b1 = tf.get_variable("b1", [25,1], initializer = tf.zeros_initializer())
# GRADED FUNCTION: initialize_parametersdef initialize_parameters():"""Initializes parameters to build a neural network with tensorflow. The shapes are:W1 : [25, 12288]b1 : [25, 1]W2 : [12, 25]b2 : [12, 1]W3 : [6, 12]b3 : [6, 1]Returns:parameters -- a dictionary of tensors containing W1, b1, W2, b2, W3, b3"""tf.set_random_seed(1) # so that your "random" numbers match ours### START CODE HERE ### (approx. 6 lines of code)W1 = tf.get_variable('W1',[25,12288],initializer=tf.contrib.layers.xavier_initializer(seed=1))b1 = tf.get_variable('b1',[25,1],initializer=tf.zeros_initializer())W2 = tf.get_variable('W2',[12,25],initializer=tf.contrib.layers.xavier_initializer(seed=1))b2 = tf.get_variable('b2',[12,1],initializer=tf.zeros_initializer())W3 = tf.get_variable('W3',[6,12],initializer=tf.contrib.layers.xavier_initializer(seed=1))b3 = tf.get_variable('b3',[6,1],initializer=tf.zeros_initializer())### END CODE HERE ###parameters = {"W1": W1,"b1": b1,"W2": W2,"b2": b2,"W3": W3,"b3": b3}return parameters

2.3 前向傳播

  • tf.add(...,...) 加
  • tf.matmul(...,...) 矩陣乘法
  • tf.nn.relu(...) ReLU 激活函數

注意,前向傳播在 z3 處停止
原因是在tensorflow中最后一個線性層的輸出作為計算損失的函數的輸入
所以,不需要 a3

# GRADED FUNCTION: forward_propagationdef forward_propagation(X, parameters):"""Implements the forward propagation for the model: LINEAR -> RELU -> LINEAR -> RELU -> LINEAR -> SOFTMAXArguments:X -- input dataset placeholder, of shape (input size, number of examples)parameters -- python dictionary containing your parameters "W1", "b1", "W2", "b2", "W3", "b3"the shapes are given in initialize_parametersReturns:Z3 -- the output of the last LINEAR unit"""# Retrieve the parameters from the dictionary "parameters" W1 = parameters['W1']b1 = parameters['b1']W2 = parameters['W2']b2 = parameters['b2']W3 = parameters['W3']b3 = parameters['b3']### START CODE HERE ### (approx. 5 lines) # Numpy Equivalents:Z1 = tf.matmul(W1, X) + b1 # Z1 = np.dot(W1, X) + b1A1 = tf.nn.relu(Z1) # A1 = relu(Z1)Z2 = tf.matmul(W2, A1) + b2 # Z2 = np.dot(W2, a1) + b2A2 = tf.nn.relu(Z2) # A2 = relu(Z2)Z3 = tf.matmul(W3, A2) + b3 # Z3 = np.dot(W3,Z2) + b3### END CODE HERE ###return Z3 # 測試 tf.reset_default_graph()with tf.Session() as sess:X, Y = create_placeholders(12288, 6)parameters = initialize_parameters()Z3 = forward_propagation(X, parameters)print("Z3 = " + str(Z3))# Z3 = Tensor("add_2:0", shape=(6, ?), dtype=float32)

2.4 計算損失

  • tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(logits = ..., labels = ...))
  • reduce_mean 求平均
  • tf.nn.softmax_cross_entropy_with_logits 的輸入 logits 和labels 的形狀必須是(樣本個數, 分類個數classes)
# GRADED FUNCTION: compute_cost def compute_cost(Z3, Y):"""Computes the costArguments:Z3 -- output of forward propagation (output of the last LINEAR unit), of shape (6, number of examples)Y -- "true" labels vector placeholder, same shape as Z3Returns:cost - Tensor of the cost function"""# to fit the tensorflow requirement for tf.nn.softmax_cross_entropy_with_logits(...,...)logits = tf.transpose(Z3) # 形狀不對,先轉置labels = tf.transpose(Y)### START CODE HERE ### (1 line of code)cost = tf.reduce_mean(tf.nn.softmax_cross_entropy_with_logits(logits=logits, labels=labels))### END CODE HERE ###return cost

2.5 后向傳播、更新參數

此步,框架會幫你完成,你需要建立 optimizer 優化器對象

調用這個優化器(傳入cost),在 Session 中運行它

例如:

創建梯度下降優化器

  • optimizer = tf.train.GradientDescentOptimizer(learning_rate = learning_rate).minimize(cost)

運行優化

  • _ , c = sess.run([optimizer, cost], feed_dict={X: minibatch_X, Y: minibatch_Y})

2.6 建立完整的TF模型

  • 使用上面的函數
  • 使用 Adam 優化器 optimizer = tf.train.AdamOptimizer(learning_rate=learning_rate).minimize(cost)
def model(X_train, Y_train, X_test, Y_test, learning_rate = 0.0001,num_epochs = 1500, minibatch_size = 32, print_cost = True):"""Implements a three-layer tensorflow neural network: LINEAR->RELU->LINEAR->RELU->LINEAR->SOFTMAX.Arguments:X_train -- training set, of shape (input size = 12288, number of training examples = 1080)Y_train -- test set, of shape (output size = 6, number of training examples = 1080)X_test -- training set, of shape (input size = 12288, number of training examples = 120)Y_test -- test set, of shape (output size = 6, number of test examples = 120)learning_rate -- learning rate of the optimizationnum_epochs -- number of epochs of the optimization loopminibatch_size -- size of a minibatchprint_cost -- True to print the cost every 100 epochsReturns:parameters -- parameters learnt by the model. They can then be used to predict."""ops.reset_default_graph() # to be able to rerun the model without overwriting tf variablestf.set_random_seed(1) # to keep consistent resultsseed = 3 # to keep consistent results(n_x, m) = X_train.shape # (n_x: input size, m : number of examples in the train set)n_y = Y_train.shape[0] # n_y : output sizecosts = [] # To keep track of the cost# Create Placeholders of shape (n_x, n_y)### START CODE HERE ### (1 line)X, Y = create_placeholders(n_x, n_y)### END CODE HERE #### Initialize parameters### START CODE HERE ### (1 line)parameters = initialize_parameters()### END CODE HERE #### Forward propagation: Build the forward propagation in the tensorflow graph### START CODE HERE ### (1 line)Z3 = forward_propagation(X, parameters)### END CODE HERE #### Cost function: Add cost function to tensorflow graph### START CODE HERE ### (1 line)cost = compute_cost(Z3, Y)### END CODE HERE #### Backpropagation: Define the tensorflow optimizer. Use an AdamOptimizer.### START CODE HERE ### (1 line)optimizer = tf.train.AdamOptimizer(learning_rate=learning_rate).minimize(cost)### END CODE HERE #### 初始化所有變量init = tf.global_variables_initializer()# Start the session to compute the tensorflow graphwith tf.Session() as sess:# Run the initializationsess.run(init)# Do the training loopfor epoch in range(num_epochs):epoch_cost = 0. # Defines a cost related to an epochnum_minibatches = int(m / minibatch_size) # 多少個minibatch子集seed = seed + 1minibatches = random_mini_batches(X_train, Y_train, minibatch_size, seed)for minibatch in minibatches:#對每個隨機的子集# Select a minibatch(minibatch_X, minibatch_Y) = minibatch# IMPORTANT: The line that runs the graph on a minibatch.# Run the session to execute the "optimizer" and the "cost", # the feedict should contain a minibatch for (X,Y).### START CODE HERE ### (1 line)_ , minibatch_cost = sess.run([optimizer, cost], feed_dict={X:minibatch_X, Y:minibatch_Y})### END CODE HERE ###epoch_cost += minibatch_cost / num_minibatches# Print the cost every epochif print_cost == True and epoch % 100 == 0:print ("Cost after epoch %i: %f" % (epoch, epoch_cost))if print_cost == True and epoch % 5 == 0:costs.append(epoch_cost)# plot the costplt.plot(np.squeeze(costs))plt.ylabel('cost')plt.xlabel('iterations (per tens)')plt.title("Learning rate =" + str(learning_rate))plt.show()# lets save the parameters in a variableparameters = sess.run(parameters)print ("Parameters have been trained!")# Calculate the correct predictionscorrect_prediction = tf.equal(tf.argmax(Z3), tf.argmax(Y))# Calculate accuracy on the test setaccuracy = tf.reduce_mean(tf.cast(correct_prediction, "float"))print ("Train Accuracy:", accuracy.eval({X: X_train, Y: Y_train}))print ("Test Accuracy:", accuracy.eval({X: X_test, Y: Y_test}))return parameters
  • 運行模型進行訓練
parameters = model(X_train, Y_train, X_test, Y_test) Cost after epoch 0: 1.855702 Cost after epoch 100: 1.017255 Cost after epoch 200: 0.733184 Cost after epoch 300: 0.573071 Cost after epoch 400: 0.468573 Cost after epoch 500: 0.381228 Cost after epoch 600: 0.313815 Cost after epoch 700: 0.253708 Cost after epoch 800: 0.203900 Cost after epoch 900: 0.166453 Cost after epoch 1000: 0.146636 Cost after epoch 1100: 0.107279 Cost after epoch 1200: 0.086699 Cost after epoch 1300: 0.059341 Cost after epoch 1400: 0.052289

Parameters have been trained! Train Accuracy: 0.9990741 Test Accuracy: 0.725
  • 可以看出模型在訓練集上擬合的很好,在測試集上效果一般,存在過擬合,可以嘗試添加 L2 正則化、dropout 正則化,降低過擬合

2.7 用自己的照片測試

import scipy from PIL import Image from scipy import ndimage## START CODE HERE ## (PUT YOUR IMAGE NAME) my_image = "thumbs_up.jpg" ## END CODE HERE ### We preprocess your image to fit your algorithm. fname = "images/" + my_image image = np.array(ndimage.imread(fname, flatten=False)) my_image = scipy.misc.imresize(image, size=(64,64)).reshape((1, 64*64*3)).T my_image_prediction = predict(my_image, parameters)plt.imshow(image) print("Your algorithm predicts: y = " + str(np.squeeze(my_image_prediction))) Your algorithm predicts: y = 3


程序預測出錯了,因為訓練集里沒有豎著大拇指表示的1

Your algorithm predicts: y = 5

總結

  • TensorFlow是一個深度學習編程框架
  • TensorFlow中的兩個主要對象是 TensorsOperators
  • code 步驟:
  • 創建包含Tensors (Variables, Placeholders …)和 Operators (tf.matmul, tf.add, …)
  • 創建一個 Session
  • 初始化 Session
  • 運行 Session 執行
  • 可以多次執行
    optimizer 優化器對象上運行 Session 時,反向傳播和優化將自動完成

  • 我的CSDN博客地址 https://michael.blog.csdn.net/

    長按或掃碼關注我的公眾號(Michael阿明),一起加油、一起學習進步!

    總結

    以上是生活随笔為你收集整理的02.改善深层神经网络:超参数调试、正则化以及优化 W3. 超参数调试、Batch Norm和程序框架(作业:TensorFlow教程+数字手势预测)的全部內容,希望文章能夠幫你解決所遇到的問題。

    如果覺得生活随笔網站內容還不錯,歡迎將生活随笔推薦給好友。