diff --git a/19_learner.ipynb b/19_learner.ipynb index e53b42a..8944fe9 100644 --- a/19_learner.ipynb +++ b/19_learner.ipynb @@ -1326,7 +1326,7 @@ "outputs": [], "source": [ "class SGD:\n", - " def __init__(self, params, lr, wd=0.): store_attr(self, 'params,lr,wd')\n", + " def __init__(self, params, lr, wd=0.): store_attr()\n", " def step(self):\n", " for p in self.params:\n", " p.data -= (p.grad.data + p.data*self.wd) * self.lr\n", @@ -1367,7 +1367,7 @@ "source": [ "class Learner:\n", " def __init__(self, model, dls, loss_func, lr, cbs, opt_func=SGD):\n", - " store_attr(self, 'model,dls,loss_func,lr,cbs,opt_func')\n", + " store_attr()\n", " for cb in cbs: cb.learner = self\n", "\n", " def one_batch(self):\n", diff --git a/clean/19_learner.ipynb b/clean/19_learner.ipynb index e9cf3a0..7012fc0 100644 --- a/clean/19_learner.ipynb +++ b/clean/19_learner.ipynb @@ -606,7 +606,7 @@ "outputs": [], "source": [ "class SGD:\n", - " def __init__(self, params, lr, wd=0.): store_attr(self, 'params,lr,wd')\n", + " def __init__(self, params, lr, wd=0.): store_attr()\n", " def step(self):\n", " for p in self.params:\n", " p.data -= (p.grad.data + p.data*self.wd) * self.lr\n", @@ -633,7 +633,7 @@ "source": [ "class Learner:\n", " def __init__(self, model, dls, loss_func, lr, cbs, opt_func=SGD):\n", - " store_attr(self, 'model,dls,loss_func,lr,cbs,opt_func')\n", + " store_attr()\n", " for cb in cbs: cb.learner = self\n", "\n", " def one_batch(self):\n",