Change deprecated 'lr' argument to 'learning_rate'
This gets rid of the UserWarning: `The 'lr' argument is deprecated, use 'learning_rate' instead.`
This commit is contained in:
@@ -173,7 +173,7 @@
|
|||||||
"source": [
|
"source": [
|
||||||
"# Loss function\n",
|
"# Loss function\n",
|
||||||
"loss_sv = tf.keras.losses.MeanSquaredError()\n",
|
"loss_sv = tf.keras.losses.MeanSquaredError()\n",
|
||||||
"optimizer_sv = tf.keras.optimizers.Adam(lr=0.001)\n",
|
"optimizer_sv = tf.keras.optimizers.Adam(learning_rate=0.001)\n",
|
||||||
"nn_sv.compile(optimizer=optimizer_sv, loss=loss_sv)\n",
|
"nn_sv.compile(optimizer=optimizer_sv, loss=loss_sv)\n",
|
||||||
"\n",
|
"\n",
|
||||||
"# Training\n",
|
"# Training\n",
|
||||||
@@ -291,7 +291,7 @@
|
|||||||
"def loss_dp(y_true, y_pred):\n",
|
"def loss_dp(y_true, y_pred):\n",
|
||||||
" return mse(y_true,y_pred**2)\n",
|
" return mse(y_true,y_pred**2)\n",
|
||||||
"\n",
|
"\n",
|
||||||
"optimizer_dp = tf.keras.optimizers.Adam(lr=0.001)\n",
|
"optimizer_dp = tf.keras.optimizers.Adam(learning_rate=0.001)\n",
|
||||||
"nn_dp.compile(optimizer=optimizer_dp, loss=loss_dp)"
|
"nn_dp.compile(optimizer=optimizer_dp, loss=loss_dp)"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
|
|||||||
Reference in New Issue
Block a user