From f57925947d62091da9e4b9aa3316c72cd81a6ff4 Mon Sep 17 00:00:00 2001 From: Pietro Monticone <38562595+pitmonticone@users.noreply.github.com> Date: Mon, 30 Oct 2023 17:57:39 +0100 Subject: [PATCH] Update 7_2_Backpropagation.ipynb --- Notebooks/Chap07/7_2_Backpropagation.ipynb | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/Notebooks/Chap07/7_2_Backpropagation.ipynb b/Notebooks/Chap07/7_2_Backpropagation.ipynb index e79a340..c52cf46 100644 --- a/Notebooks/Chap07/7_2_Backpropagation.ipynb +++ b/Notebooks/Chap07/7_2_Backpropagation.ipynb @@ -143,7 +143,7 @@ " # Run through the layers, calculating all_f[0...K-1] and all_h[1...K]\n", " for layer in range(K):\n", " # Update preactivations and activations at this layer according to eqn 7.16\n", - " # Remmember to use np.matmul for matrrix multiplications\n", + " # Remember to use np.matmul for matrix multiplications\n", " # TODO -- Replace the lines below\n", " all_f[layer] = all_h[layer]\n", " all_h[layer+1] = all_f[layer]\n", @@ -299,7 +299,7 @@ "# Let's test if we have the derivatives right using finite differences\n", "delta_fd = 0.000001\n", "\n", - "# Test the dervatives of the bias vectors\n", + "# Test the derivatives of the bias vectors\n", "for layer in range(K):\n", " dl_dbias = np.zeros_like(all_dl_dbiases[layer])\n", " # For every element in the bias\n",