Skip to content

Commit 88c945f

Browse files
committed
revision
1 parent 1bef414 commit 88c945f

10 files changed

+487
-610
lines changed

Chapter01.ipynb

+235-351
Large diffs are not rendered by default.

Chapter02.ipynb

+41-41
Original file line numberDiff line numberDiff line change
@@ -123,7 +123,7 @@
123123
"name": "stdout",
124124
"output_type": "stream",
125125
"text": [
126-
"OrderedDict([('0.weight', tensor([[1.9690]], device='cuda:0')), ('0.bias', tensor([1.0235], device='cuda:0'))])\n"
126+
"OrderedDict([('0.weight', tensor([[1.9690]])), ('0.bias', tensor([1.0235]))])\n"
127127
]
128128
}
129129
],
@@ -327,7 +327,7 @@
327327
"name": "stdout",
328328
"output_type": "stream",
329329
"text": [
330-
"Writing model_configuration/v1.py\n"
330+
"Overwriting model_configuration/v1.py\n"
331331
]
332332
}
333333
],
@@ -370,7 +370,7 @@
370370
{
371371
"data": {
372372
"text/plain": [
373-
"<function __main__.make_train_step.<locals>.perform_train_step(x, y)>"
373+
"<function __main__.make_train_step_fn.<locals>.perform_train_step_fn(x, y)>"
374374
]
375375
},
376376
"execution_count": 18,
@@ -398,7 +398,7 @@
398398
"name": "stdout",
399399
"output_type": "stream",
400400
"text": [
401-
"Writing model_training/v1.py\n"
401+
"Overwriting model_training/v1.py\n"
402402
]
403403
}
404404
],
@@ -435,7 +435,7 @@
435435
"name": "stdout",
436436
"output_type": "stream",
437437
"text": [
438-
"OrderedDict([('0.weight', tensor([[1.9690]], device='cuda:0')), ('0.bias', tensor([1.0235], device='cuda:0'))])\n"
438+
"OrderedDict([('0.weight', tensor([[1.9690]])), ('0.bias', tensor([1.0235]))])\n"
439439
]
440440
}
441441
],
@@ -556,37 +556,37 @@
556556
{
557557
"data": {
558558
"text/plain": [
559-
"[tensor([[0.1196],\n",
560-
" [0.1395],\n",
561-
" [0.2809],\n",
562-
" [0.1834],\n",
563-
" [0.3585],\n",
559+
"[tensor([[0.9507],\n",
564560
" [0.5427],\n",
565-
" [0.0885],\n",
566-
" [0.9489],\n",
567-
" [0.9699],\n",
568-
" [0.7751],\n",
569-
" [0.9696],\n",
570-
" [0.7320],\n",
571-
" [0.0055],\n",
572-
" [0.7069],\n",
573-
" [0.8155],\n",
574-
" [0.5979]]), tensor([[1.3214],\n",
575-
" [1.3051],\n",
576-
" [1.5846],\n",
577-
" [1.4637],\n",
578-
" [1.7462],\n",
561+
" [0.1409],\n",
562+
" [0.3745],\n",
563+
" [0.1987],\n",
564+
" [0.8948],\n",
565+
" [0.7722],\n",
566+
" [0.7852],\n",
567+
" [0.5248],\n",
568+
" [0.2809],\n",
569+
" [0.1159],\n",
570+
" [0.0740],\n",
571+
" [0.1849],\n",
572+
" [0.4561],\n",
573+
" [0.7608],\n",
574+
" [0.1560]]), tensor([[2.8715],\n",
579575
" [2.2161],\n",
580-
" [1.0708],\n",
581-
" [2.8903],\n",
582-
" [2.9727],\n",
583-
" [2.4936],\n",
584-
" [2.8401],\n",
585-
" [2.4732],\n",
586-
" [1.0632],\n",
587-
" [2.4388],\n",
588-
" [2.6606],\n",
589-
" [2.0407]])]"
576+
" [1.1211],\n",
577+
" [1.7578],\n",
578+
" [1.2654],\n",
579+
" [2.7393],\n",
580+
" [2.4208],\n",
581+
" [2.5283],\n",
582+
" [2.0167],\n",
583+
" [1.5846],\n",
584+
" [1.1603],\n",
585+
" [1.1713],\n",
586+
" [1.5888],\n",
587+
" [1.7706],\n",
588+
" [2.4970],\n",
589+
" [1.2901]])]"
590590
]
591591
},
592592
"execution_count": 25,
@@ -614,7 +614,7 @@
614614
"name": "stdout",
615615
"output_type": "stream",
616616
"text": [
617-
"Writing data_preparation/v1.py\n"
617+
"Overwriting data_preparation/v1.py\n"
618618
]
619619
}
620620
],
@@ -666,7 +666,7 @@
666666
"name": "stdout",
667667
"output_type": "stream",
668668
"text": [
669-
"Writing model_training/v2.py\n"
669+
"Overwriting model_training/v2.py\n"
670670
]
671671
}
672672
],
@@ -718,7 +718,7 @@
718718
"name": "stdout",
719719
"output_type": "stream",
720720
"text": [
721-
"OrderedDict([('0.weight', tensor([[1.9684]], device='cuda:0')), ('0.bias', tensor([1.0235], device='cuda:0'))])\n"
721+
"OrderedDict([('0.weight', tensor([[1.9694]])), ('0.bias', tensor([1.0234]))])\n"
722722
]
723723
}
724724
],
@@ -1013,7 +1013,7 @@
10131013
"\n",
10141014
"for epoch in range(n_epochs):\n",
10151015
" # inner loop\n",
1016-
" loss = mini_batch(device, train_loader, train_step)\n",
1016+
" loss = mini_batch(device, train_loader, train_step_fn)\n",
10171017
" losses.append(loss)\n",
10181018
" \n",
10191019
" # VALIDATION\n",
@@ -1267,7 +1267,7 @@
12671267
"loss_fn = nn.MSELoss(reduction='mean')\n",
12681268
"\n",
12691269
"# Creates the train_step function for our model, loss function and optimizer\n",
1270-
"train_step = make_train_step(model, loss_fn, optimizer)\n",
1270+
"train_step_fn = make_train_step_fn(model, loss_fn, optimizer)\n",
12711271
"\n",
12721272
"# Creates the val_step function for our model and loss function\n",
12731273
"val_step_fn = make_val_step_fn(model, loss_fn)\n",
@@ -1320,7 +1320,7 @@
13201320
"\n",
13211321
"for epoch in range(n_epochs):\n",
13221322
" # inner loop\n",
1323-
" loss = mini_batch(device, train_loader, train_step)\n",
1323+
" loss = mini_batch(device, train_loader, train_step_fn)\n",
13241324
" losses.append(loss)\n",
13251325
" \n",
13261326
" # VALIDATION\n",
@@ -1812,7 +1812,7 @@
18121812
"name": "python",
18131813
"nbconvert_exporter": "python",
18141814
"pygments_lexer": "ipython3",
1815-
"version": "3.8.5"
1815+
"version": "3.7.5"
18161816
}
18171817
},
18181818
"nbformat": 4,

Chapter03.ipynb

+2-2
Large diffs are not rendered by default.

0 commit comments

Comments
 (0)