-
Notifications
You must be signed in to change notification settings - Fork 1
/
Copy pathnohup.out
4270 lines (4176 loc) · 219 KB
/
nohup.out
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
# Images found: 404453
# Images found: 404453
Variable containing:
1.00000e-02 *
4.1956
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
6.4286
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
3.7729
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
0.1276
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
5.9644
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.3008
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
5.0794
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
3.1106
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
8.2889
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.1910
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.3767
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.0466
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
8.9705
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.7688
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.5644
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.5258
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.9618
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.5092
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.5991
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.4554
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.5416
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.9055
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.6358
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
0.1041
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
4.0962
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
5.0271
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.3061
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.4667
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.4573
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.5410
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.9205
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
4.2521
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.9729
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
5.2815
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.8396
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
4.4841
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
9.4381
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
5.0017
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
0.1753
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
3.5776
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
5.0538
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
3.6028
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
5.2065
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
3.6140
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
8.6690
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
9.6661
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
5.5657
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
7.3313
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
3.8217
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
6.3838
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.8765
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
4.4088
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.3754
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.4363
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
5.3165
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.0257
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.9142
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
5.7607
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.1584
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
4.4553
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
0.1361
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
8.0622
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
6.1891
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
6.4175
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-03 *
7.0857
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
0.1695
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.2459
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.9666
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
3.9347
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.8216
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.3377
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
5.2132
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
0.2109
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
3.1832
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
8.6407
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
5.6431
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
2.7574
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
3.6550
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
4.4422
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
0.1532
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.9811
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
0.1508
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
3.4127
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
5.0776
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
4.8172
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
0.1584
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
7.0360
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
3.6791
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
6.0354
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
5.0118
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
4.0037
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
3.4959
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.1690
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
1.4407
[torch.cuda.FloatTensor of size 1 (GPU 0)]
Variable containing:
1.00000e-02 *
3.6148
[torch.cuda.FloatTensor of size 1 (GPU 0)]
THCudaCheck FAIL file=/opt/conda/conda-bld/pytorch_1503970438496/work/torch/lib/THC/generic/THCStorage.cu line=66 error=2 : out of memory
# Images found: 404453
# Images found: 404453
Traceback (most recent call last):
File "main.py", line 114, in <module>
inputs, labels= Variable(inputs.cuda()), Variable(labels.cuda())
File "/home/pwz/miniconda3/lib/python3.6/site-packages/torch/_utils.py", line 66, in _cuda
return new_type(self.size()).copy_(self, async)
RuntimeError: cuda runtime error (2) : out of memory at /opt/conda/conda-bld/pytorch_1503970438496/work/torch/lib/THC/generic/THCStorage.cu:66
THCudaCheck FAIL file=/opt/conda/conda-bld/pytorch_1503970438496/work/torch/lib/THC/generic/THCStorage.cu line=66 error=2 : out of memory
# Images found: 404453
# Images found: 404453
Traceback (most recent call last):
File "main.py", line 114, in <module>
inputs, labels= Variable(inputs.cuda()), Variable(labels.cuda())
File "/home/pwz/miniconda3/lib/python3.6/site-packages/torch/_utils.py", line 66, in _cuda
return new_type(self.size()).copy_(self, async)
RuntimeError: cuda runtime error (2) : out of memory at /opt/conda/conda-bld/pytorch_1503970438496/work/torch/lib/THC/generic/THCStorage.cu:66
THCudaCheck FAIL file=/opt/conda/conda-bld/pytorch_1503970438496/work/torch/lib/THC/generic/THCStorage.cu line=66 error=2 : out of memory
# Images found: 404453
# Images found: 404453
Traceback (most recent call last):
File "main.py", line 121, in <module>
output = net(inputs) # places output
File "/home/pwz/miniconda3/lib/python3.6/site-packages/torch/nn/modules/module.py", line 224, in __call__
result = self.forward(*input, **kwargs)
File "/home/pwz/sambavol/3dreconstruction/model/reconNet.py", line 113, in forward
x = self.decoding(x) # convert to 3D voxel distribution
File "/home/pwz/miniconda3/lib/python3.6/site-packages/torch/nn/modules/module.py", line 224, in __call__
result = self.forward(*input, **kwargs)
File "/home/pwz/miniconda3/lib/python3.6/site-packages/torch/nn/modules/container.py", line 67, in forward
input = module(input)
File "/home/pwz/miniconda3/lib/python3.6/site-packages/torch/nn/modules/module.py", line 224, in __call__
result = self.forward(*input, **kwargs)
File "/home/pwz/miniconda3/lib/python3.6/site-packages/torch/nn/modules/conv.py", line 663, in forward
output_padding, self.groups, self.dilation)
File "/home/pwz/miniconda3/lib/python3.6/site-packages/torch/nn/functional.py", line 203, in conv_transpose3d
return f(input, weight, bias)
RuntimeError: cuda runtime error (2) : out of memory at /opt/conda/conda-bld/pytorch_1503970438496/work/torch/lib/THC/generic/THCStorage.cu:66
# Images found: 404453
# Images found: 404453
PLACES TRAINING Epoch: 1 10 loss: 4.7886302948
PLACES TRAINING Epoch: 1 20 loss: 4.7885618687
PLACES TRAINING Epoch: 1 30 loss: 4.7884809494
PLACES TRAINING Epoch: 1 40 loss: 4.7884079456
PLACES TRAINING Epoch: 1 50 loss: 4.7883368492
PLACES TRAINING Epoch: 1 60 loss: 4.7882752419
PLACES TRAINING Epoch: 1 70 loss: 4.7882125854
PLACES TRAINING Epoch: 1 80 loss: 4.7881651878
PLACES TRAINING Epoch: 1 90 loss: 4.7881205082
PLACES TRAINING Epoch: 1 100 loss: 4.7880945683
PLACES TRAINING Epoch: 1 110 loss: 4.7880633831
PLACES TRAINING Epoch: 1 120 loss: 4.7880323887
PLACES TRAINING Epoch: 1 130 loss: 4.7880110264
PLACES TRAINING Epoch: 1 140 loss: 4.7879999161
PLACES TRAINING Epoch: 1 150 loss: 4.7879752159
PLACES TRAINING Epoch: 1 160 loss: 4.7879776478
PLACES TRAINING Epoch: 1 170 loss: 4.7879537106
PLACES TRAINING Epoch: 1 180 loss: 4.7879507542
PLACES TRAINING Epoch: 1 190 loss: 4.7879428864
PLACES TRAINING Epoch: 1 200 loss: 4.7879436970
PLACES TRAINING Epoch: 1 210 loss: 4.7879308701
PLACES TRAINING Epoch: 1 220 loss: 4.7879217625
PLACES TRAINING Epoch: 1 230 loss: 4.7879154205
PLACES TRAINING Epoch: 1 240 loss: 4.7879158974
PLACES TRAINING Epoch: 1 250 loss: 4.7879190922
PLACES TRAINING Epoch: 1 260 loss: 4.7879096031
PLACES TRAINING Epoch: 1 270 loss: 4.7879016876
PLACES TRAINING Epoch: 1 280 loss: 4.7879105568
PLACES TRAINING Epoch: 1 290 loss: 4.7879012585
PLACES TRAINING Epoch: 1 300 loss: 4.7878999233
PLACES TRAINING Epoch: 1 310 loss: 4.7879010201
PLACES TRAINING Epoch: 1 320 loss: 4.7878974438
PLACES TRAINING Epoch: 1 330 loss: 4.7878970623
PLACES TRAINING Epoch: 1 340 loss: 4.7878887177
PLACES TRAINING Epoch: 1 350 loss: 4.7878927708
PLACES TRAINING Epoch: 1 360 loss: 4.7878808498
PLACES TRAINING Epoch: 1 370 loss: 4.7878970146
PLACES TRAINING Epoch: 1 380 loss: 4.7878888607
PLACES TRAINING Epoch: 1 390 loss: 4.7879046917
PLACES TRAINING Epoch: 1 400 loss: 4.7878905773
PLACES TRAINING Epoch: 1 410 loss: 4.7878850937
PLACES TRAINING Epoch: 1 420 loss: 4.7878954411
PLACES TRAINING Epoch: 1 430 loss: 4.7878825665
PLACES TRAINING Epoch: 1 440 loss: 4.7878924847
PLACES TRAINING Epoch: 1 450 loss: 4.7879029751
PLACES TRAINING Epoch: 1 460 loss: 4.7878980637
PLACES TRAINING Epoch: 1 470 loss: 4.7878848076
PLACES TRAINING Epoch: 1 480 loss: 4.7879002571
PLACES TRAINING Epoch: 1 490 loss: 4.7878771305
PLACES TRAINING Epoch: 1 500 loss: 4.7878816128
PLACES TRAINING Epoch: 1 510 loss: 4.7878776073
PLACES TRAINING Epoch: 1 520 loss: 4.7878930092
PLACES TRAINING Epoch: 1 530 loss: 4.7878915310
PLACES TRAINING Epoch: 1 540 loss: 4.7878642082
PLACES TRAINING Epoch: 1 550 loss: 4.7878627300
PLACES TRAINING Epoch: 1 560 loss: 4.7878909111
PLACES TRAINING Epoch: 1 570 loss: 4.7878579617
PLACES TRAINING Epoch: 1 580 loss: 4.7878612995
PLACES TRAINING Epoch: 1 590 loss: 4.7878369331
PLACES TRAINING Epoch: 1 600 loss: 4.7878631592
PLACES TRAINING Epoch: 1 610 loss: 4.7878300190
PLACES TRAINING Epoch: 1 620 loss: 4.7878521442
PLACES TRAINING Epoch: 1 630 loss: 4.7878450394
PLACES TRAINING Epoch: 1 640 loss: 4.7878368855
PLACES TRAINING Epoch: 1 650 loss: 4.7878104687
PLACES TRAINING Epoch: 1 660 loss: 4.7878005028
PLACES TRAINING Epoch: 1 670 loss: 4.7878145218
PLACES TRAINING Epoch: 1 680 loss: 4.7878066063
PLACES TRAINING Epoch: 1 690 loss: 4.7878122807
PLACES TRAINING Epoch: 1 700 loss: 4.7877920151
PLACES TRAINING Epoch: 1 710 loss: 4.7877779961
PLACES TRAINING Epoch: 1 720 loss: 4.7877859116
PLACES TRAINING Epoch: 1 730 loss: 4.7877809048
PLACES TRAINING Epoch: 1 740 loss: 4.7877804279
PLACES TRAINING Epoch: 1 750 loss: 4.7877768517
PLACES TRAINING Epoch: 1 760 loss: 4.7877883434
PLACES TRAINING Epoch: 1 770 loss: 4.7877455235
PLACES TRAINING Epoch: 1 780 loss: 4.7877438068
PLACES TRAINING Epoch: 1 790 loss: 4.7877759457
PLACES TRAINING Epoch: 1 800 loss: 4.7877202988
PLACES TRAINING Epoch: 1 810 loss: 4.7877257347
PLACES TRAINING Epoch: 1 820 loss: 4.7877417088
PLACES TRAINING Epoch: 1 830 loss: 4.7877042770
PLACES TRAINING Epoch: 1 840 loss: 4.7877052307
PLACES TRAINING Epoch: 1 850 loss: 4.7877338886
PLACES TRAINING Epoch: 1 860 loss: 4.7877107143
PLACES TRAINING Epoch: 1 870 loss: 4.7877125740
PLACES TRAINING Epoch: 1 880 loss: 4.7876343250
PLACES TRAINING Epoch: 1 890 loss: 4.7876293182
PLACES TRAINING Epoch: 1 900 loss: 4.7876598358
PLACES TRAINING Epoch: 1 910 loss: 4.7875970364
PLACES TRAINING Epoch: 1 920 loss: 4.7876192093
PLACES TRAINING Epoch: 1 930 loss: 4.7875816345
PLACES TRAINING Epoch: 1 940 loss: 4.7875976562
PLACES TRAINING Epoch: 1 950 loss: 4.7876012802
PLACES TRAINING Epoch: 1 960 loss: 4.7876300812
PLACES TRAINING Epoch: 1 970 loss: 4.7874696732
PLACES TRAINING Epoch: 1 980 loss: 4.7874656200
PLACES TRAINING Epoch: 1 990 loss: 4.7875513077
PLACES TRAINING Epoch: 1 1000 loss: 4.7875005245
PLACES TRAINING Epoch: 1 1010 loss: 4.7875313759
PLACES TRAINING Epoch: 1 1020 loss: 4.7874162197
PLACES TRAINING Epoch: 1 1030 loss: 4.7875265121
PLACES TRAINING Epoch: 1 1040 loss: 4.7874111652
PLACES TRAINING Epoch: 1 1050 loss: 4.7873582363
PLACES TRAINING Epoch: 1 1060 loss: 4.7872899055
PLACES TRAINING Epoch: 1 1070 loss: 4.7873237133
PLACES TRAINING Epoch: 1 1080 loss: 4.7872685432
PLACES TRAINING Epoch: 1 1090 loss: 4.7871816635
PLACES TRAINING Epoch: 1 1100 loss: 4.7872426987
PLACES TRAINING Epoch: 1 1110 loss: 4.7871777058
PLACES TRAINING Epoch: 1 1120 loss: 4.7870805264
PLACES TRAINING Epoch: 1 1130 loss: 4.7871105194
PLACES TRAINING Epoch: 1 1140 loss: 4.7870764732
PLACES TRAINING Epoch: 1 1150 loss: 4.7869080544
PLACES TRAINING Epoch: 1 1160 loss: 4.7867296696
PLACES TRAINING Epoch: 1 1170 loss: 4.7868642807
PLACES TRAINING Epoch: 1 1180 loss: 4.7865507126
PLACES TRAINING Epoch: 1 1190 loss: 4.7866690636
PLACES TRAINING Epoch: 1 1200 loss: 4.7862783432
PLACES TRAINING Epoch: 1 1210 loss: 4.7863171577
PLACES TRAINING Epoch: 1 1220 loss: 4.7860982895
PLACES TRAINING Epoch: 1 1230 loss: 4.7857401371
PLACES TRAINING Epoch: 1 1240 loss: 4.7855258942
PLACES TRAINING Epoch: 1 1250 loss: 4.7851826191
PLACES TRAINING Epoch: 1 1260 loss: 4.7844129086
PLACES TRAINING Epoch: 1 1270 loss: 4.7837672234
PLACES TRAINING Epoch: 1 1280 loss: 4.7834339142
PLACES TRAINING Epoch: 1 1290 loss: 4.7830818176
PLACES TRAINING Epoch: 1 1300 loss: 4.7828629971
PLACES TRAINING Epoch: 1 1310 loss: 4.7820512772
PLACES TRAINING Epoch: 1 1320 loss: 4.7828031540
PLACES TRAINING Epoch: 1 1330 loss: 4.7819653034
PLACES TRAINING Epoch: 1 1340 loss: 4.7814518452
PLACES TRAINING Epoch: 1 1350 loss: 4.7816332817
PLACES TRAINING Epoch: 1 1360 loss: 4.7806031704
PLACES TRAINING Epoch: 1 1370 loss: 4.7813429832
PLACES TRAINING Epoch: 1 1380 loss: 4.7813388824
PLACES TRAINING Epoch: 1 1390 loss: 4.7799996853
PLACES TRAINING Epoch: 1 1400 loss: 4.7798505306
PLACES TRAINING Epoch: 1 1410 loss: 4.7806661129
PLACES TRAINING Epoch: 1 1420 loss: 4.7793375969
PLACES TRAINING Epoch: 1 1430 loss: 4.7794232845
PLACES TRAINING Epoch: 1 1440 loss: 4.7789787292
PLACES TRAINING Epoch: 1 1450 loss: 4.7783875942
PLACES TRAINING Epoch: 1 1460 loss: 4.7789820194
PLACES TRAINING Epoch: 1 1470 loss: 4.7791722298
PLACES TRAINING Epoch: 1 1480 loss: 4.7778816700
PLACES TRAINING Epoch: 1 1490 loss: 4.7778403759
PLACES TRAINING Epoch: 1 1500 loss: 4.7770515442
PLACES TRAINING Epoch: 1 1510 loss: 4.7780581474
PLACES TRAINING Epoch: 1 1520 loss: 4.7775148392
PLACES TRAINING Epoch: 1 1530 loss: 4.7775978088
PLACES TRAINING Epoch: 1 1540 loss: 4.7770130634
PLACES TRAINING Epoch: 1 1550 loss: 4.7767552853
PLACES TRAINING Epoch: 1 1560 loss: 4.7765508652
PLACES TRAINING Epoch: 1 1570 loss: 4.7769782543
PLACES TRAINING Epoch: 1 1580 loss: 4.7762802601
PLACES TRAINING Epoch: 1 1590 loss: 4.7746053696
PLACES TRAINING Epoch: 1 1600 loss: 4.7763378143
PLACES TRAINING Epoch: 1 1610 loss: 4.7764671326
PLACES TRAINING Epoch: 1 1620 loss: 4.7766948223
PLACES TRAINING Epoch: 1 1630 loss: 4.7756355286
PLACES TRAINING Epoch: 1 1640 loss: 4.7755176067
PLACES TRAINING Epoch: 1 1650 loss: 4.7761536121
PLACES TRAINING Epoch: 1 1660 loss: 4.7762480736
PLACES TRAINING Epoch: 1 1670 loss: 4.7753727436
PLACES TRAINING Epoch: 1 1680 loss: 4.7752280235
PLACES TRAINING Epoch: 1 1690 loss: 4.7764227390
PLACES TRAINING Epoch: 1 1700 loss: 4.7756756783
PLACES TRAINING Epoch: 1 1710 loss: 4.7753003120
PLACES TRAINING Epoch: 1 1720 loss: 4.7748317719
PLACES TRAINING Epoch: 1 1730 loss: 4.7742067337
PLACES TRAINING Epoch: 1 1740 loss: 4.7753345490
PLACES TRAINING Epoch: 1 1750 loss: 4.7743945599
PLACES TRAINING Epoch: 1 1760 loss: 4.7757220745
PLACES TRAINING Epoch: 1 1770 loss: 4.7724943161
PLACES TRAINING Epoch: 1 1780 loss: 4.7742180347
PLACES TRAINING Epoch: 1 1790 loss: 4.7737714291
PLACES TRAINING Epoch: 1 1800 loss: 4.7735708237
PLACES TRAINING Epoch: 1 1810 loss: 4.7746742725
PLACES TRAINING Epoch: 1 1820 loss: 4.7735352516
PLACES TRAINING Epoch: 1 1830 loss: 4.7739082336
PLACES TRAINING Epoch: 1 1840 loss: 4.7719822407
PLACES TRAINING Epoch: 1 1850 loss: 4.7701690197
PLACES TRAINING Epoch: 1 1860 loss: 4.7733465195
PLACES TRAINING Epoch: 1 1870 loss: 4.7725806236
PLACES TRAINING Epoch: 1 1880 loss: 4.7757245541
PLACES TRAINING Epoch: 1 1890 loss: 4.7734151840
PLACES TRAINING Epoch: 1 1900 loss: 4.7715197086
PLACES TRAINING Epoch: 1 1910 loss: 4.7733643055
PLACES TRAINING Epoch: 1 1920 loss: 4.7716982365
PLACES TRAINING Epoch: 1 1930 loss: 4.7724658012
PLACES TRAINING Epoch: 1 1940 loss: 4.7735514164
PLACES TRAINING Epoch: 1 1950 loss: 4.7718641758
PLACES TRAINING Epoch: 1 1960 loss: 4.7712352276
PLACES TRAINING Epoch: 1 1970 loss: 4.7708001614
PLACES TRAINING Epoch: 1 1980 loss: 4.7699496269
PLACES TRAINING Epoch: 1 1990 loss: 4.7713542938
PLACES TRAINING Epoch: 1 2000 loss: 4.7735596180
PLACES TRAINING Epoch: 1 2010 loss: 4.7693915367
PLACES TRAINING Epoch: 1 2020 loss: 4.7727363586
PLACES TRAINING Epoch: 1 2030 loss: 4.7694354057
PLACES TRAINING Epoch: 1 2040 loss: 4.7715166569
PLACES TRAINING Epoch: 1 2050 loss: 4.7718371868
PLACES TRAINING Epoch: 1 2060 loss: 4.7732407093
PLACES TRAINING Epoch: 1 2070 loss: 4.7712067604
PLACES TRAINING Epoch: 1 2080 loss: 4.7726894855
PLACES TRAINING Epoch: 1 2090 loss: 4.7689301491
PLACES TRAINING Epoch: 1 2100 loss: 4.7697597504
PLACES TRAINING Epoch: 1 2110 loss: 4.7717631817
PLACES TRAINING Epoch: 1 2120 loss: 4.7707133293
PLACES TRAINING Epoch: 1 2130 loss: 4.7682919025
PLACES TRAINING Epoch: 1 2140 loss: 4.7702991962
PLACES TRAINING Epoch: 1 2150 loss: 4.7698646545
PLACES TRAINING Epoch: 1 2160 loss: 4.7697356701
PLACES TRAINING Epoch: 1 2170 loss: 4.7681602478
PLACES TRAINING Epoch: 1 2180 loss: 4.7709449768
PLACES TRAINING Epoch: 1 2190 loss: 4.7694787502
PLACES TRAINING Epoch: 1 2200 loss: 4.7694143295
PLACES TRAINING Epoch: 1 2210 loss: 4.7695420742
PLACES TRAINING Epoch: 1 2220 loss: 4.7695052624
PLACES TRAINING Epoch: 1 2230 loss: 4.7693904877
PLACES TRAINING Epoch: 1 2240 loss: 4.7688739777
PLACES TRAINING Epoch: 1 2250 loss: 4.7691915989
PLACES TRAINING Epoch: 1 2260 loss: 4.7677333355
PLACES TRAINING Epoch: 1 2270 loss: 4.7689488888
PLACES TRAINING Epoch: 1 2280 loss: 4.7674462795
PLACES TRAINING Epoch: 1 2290 loss: 4.7674194813
PLACES TRAINING Epoch: 1 2300 loss: 4.7675898075
PLACES TRAINING Epoch: 1 2310 loss: 4.7670148849
PLACES TRAINING Epoch: 1 2320 loss: 4.7691652775
PLACES TRAINING Epoch: 1 2330 loss: 4.7668782234
PLACES TRAINING Epoch: 1 2340 loss: 4.7670744896
PLACES TRAINING Epoch: 1 2350 loss: 4.7668692112
PLACES TRAINING Epoch: 1 2360 loss: 4.7653708935
PLACES TRAINING Epoch: 1 2370 loss: 4.7674992561
PLACES TRAINING Epoch: 1 2380 loss: 4.7675780296
PLACES TRAINING Epoch: 1 2390 loss: 4.7678493977
PLACES TRAINING Epoch: 1 2400 loss: 4.7666809559
PLACES TRAINING Epoch: 1 2410 loss: 4.7681175709
PLACES TRAINING Epoch: 1 2420 loss: 4.7650554180
PLACES TRAINING Epoch: 1 2430 loss: 4.7660272598
PLACES TRAINING Epoch: 1 2440 loss: 4.7661413670
PLACES TRAINING Epoch: 1 2450 loss: 4.7670907497
PLACES TRAINING Epoch: 1 2460 loss: 4.7665151596
PLACES TRAINING Epoch: 1 2470 loss: 4.7670853138
PLACES TRAINING Epoch: 1 2480 loss: 4.7645226479
PLACES TRAINING Epoch: 1 2490 loss: 4.7655515194
PLACES TRAINING Epoch: 1 2500 loss: 4.7678144455
PLACES TRAINING Epoch: 1 2510 loss: 4.7666143894
PLACES TRAINING Epoch: 1 2520 loss: 4.7664645672
PLACES TRAINING Epoch: 1 2530 loss: 4.7661966801
PLACES TRAINING Epoch: 1 2540 loss: 4.7652577400
PLACES TRAINING Epoch: 1 2550 loss: 4.7674498081
PLACES TRAINING Epoch: 1 2560 loss: 4.7633671284
PLACES TRAINING Epoch: 1 2570 loss: 4.7672943115
PLACES TRAINING Epoch: 1 2580 loss: 4.7675203323
PLACES TRAINING Epoch: 1 2590 loss: 4.7662276745
PLACES TRAINING Epoch: 1 2600 loss: 4.7644129276
PLACES TRAINING Epoch: 1 2610 loss: 4.7656124592
PLACES TRAINING Epoch: 1 2620 loss: 4.7650300980
PLACES TRAINING Epoch: 1 2630 loss: 4.7660876751
PLACES TRAINING Epoch: 1 2640 loss: 4.7649938583
PLACES TRAINING Epoch: 1 2650 loss: 4.7626018047
PLACES TRAINING Epoch: 1 2660 loss: 4.7653009892
PLACES TRAINING Epoch: 1 2670 loss: 4.7664123058
PLACES TRAINING Epoch: 1 2680 loss: 4.7608278275
PLACES TRAINING Epoch: 1 2690 loss: 4.7644752502
PLACES TRAINING Epoch: 1 2700 loss: 4.7640473843
PLACES TRAINING Epoch: 1 2710 loss: 4.7657165527
PLACES TRAINING Epoch: 1 2720 loss: 4.7659591675
PLACES TRAINING Epoch: 1 2730 loss: 4.7633041859
PLACES TRAINING Epoch: 1 2740 loss: 4.7669604301
PLACES TRAINING Epoch: 1 2750 loss: 4.7647109509
PLACES TRAINING Epoch: 1 2760 loss: 4.7632184505
PLACES TRAINING Epoch: 1 2770 loss: 4.7632282734
PLACES TRAINING Epoch: 1 2780 loss: 4.7640161037
PLACES TRAINING Epoch: 1 2790 loss: 4.7654199600
PLACES TRAINING Epoch: 1 2800 loss: 4.7647404671
PLACES TRAINING Epoch: 1 2810 loss: 4.7645141602
PLACES TRAINING Epoch: 1 2820 loss: 4.7641579151
PLACES TRAINING Epoch: 1 2830 loss: 4.7626901150
PLACES TRAINING Epoch: 1 2840 loss: 4.7634956360
PLACES TRAINING Epoch: 1 2850 loss: 4.7651075363
PLACES TRAINING Epoch: 1 2860 loss: 4.7614078522
PLACES TRAINING Epoch: 1 2870 loss: 4.7647854805
PLACES TRAINING Epoch: 1 2880 loss: 4.7642693043
PLACES TRAINING Epoch: 1 2890 loss: 4.7615381718
PLACES TRAINING Epoch: 1 2900 loss: 4.7623332977
PLACES TRAINING Epoch: 1 2910 loss: 4.7622663021
PLACES TRAINING Epoch: 1 2920 loss: 4.7622841358
PLACES TRAINING Epoch: 1 2930 loss: 4.7611585140
PLACES TRAINING Epoch: 1 2940 loss: 4.7631280899
PLACES TRAINING Epoch: 1 2950 loss: 4.7613943577
PLACES TRAINING Epoch: 1 2960 loss: 4.7628435612
PLACES TRAINING Epoch: 1 2970 loss: 4.7632819653
PLACES TRAINING Epoch: 1 2980 loss: 4.7642022133
PLACES TRAINING Epoch: 1 2990 loss: 4.7610198021
PLACES TRAINING Epoch: 1 3000 loss: 4.7622976303
PLACES TRAINING Epoch: 1 3010 loss: 4.7622348785
PLACES TRAINING Epoch: 1 3020 loss: 4.7583690643
PLACES TRAINING Epoch: 1 3030 loss: 4.7610699177
PLACES TRAINING Epoch: 1 3040 loss: 4.7620903015
PLACES TRAINING Epoch: 1 3050 loss: 4.7614843845
PLACES TRAINING Epoch: 1 3060 loss: 4.7594771862
PLACES TRAINING Epoch: 1 3070 loss: 4.7621128559
PLACES TRAINING Epoch: 1 3080 loss: 4.7615098953
PLACES TRAINING Epoch: 1 3090 loss: 4.7607894421
PLACES TRAINING Epoch: 1 3100 loss: 4.7609599113
PLACES TRAINING Epoch: 1 3110 loss: 4.7641319275
PLACES TRAINING Epoch: 1 3120 loss: 4.7615234852
PLACES TRAINING Epoch: 1 3130 loss: 4.7600917339
PLACES TRAINING Epoch: 1 3140 loss: 4.7615292549
PLACES TRAINING Epoch: 1 3150 loss: 4.7597286701
PLACES TRAINING Epoch: 1 3160 loss: 4.7615473270
PLACES TRAINING Epoch: 1 3170 loss: 4.7593470097
PLACES TRAINING Epoch: 1 3180 loss: 4.7593730927
PLACES TRAINING Epoch: 1 3190 loss: 4.7615660667
PLACES TRAINING Epoch: 1 3200 loss: 4.7615429401
PLACES TRAINING Epoch: 1 3210 loss: 4.7606844902
PLACES TRAINING Epoch: 1 3220 loss: 4.7623759747
PLACES TRAINING Epoch: 1 3230 loss: 4.7615379810
PLACES TRAINING Epoch: 1 3240 loss: 4.7573424816
PLACES TRAINING Epoch: 1 3250 loss: 4.7627648830
PLACES TRAINING Epoch: 1 3260 loss: 4.7605018139
PLACES TRAINING Epoch: 1 3270 loss: 4.7602666378
PLACES TRAINING Epoch: 1 3280 loss: 4.7610440254
PLACES TRAINING Epoch: 1 3290 loss: 4.7616984367
PLACES TRAINING Epoch: 1 3300 loss: 4.7592377663
PLACES TRAINING Epoch: 1 3310 loss: 4.7628463745
PLACES TRAINING Epoch: 1 3320 loss: 4.7615147591
PLACES TRAINING Epoch: 1 3330 loss: 4.7586565018
PLACES TRAINING Epoch: 1 3340 loss: 4.7576406479
PLACES TRAINING Epoch: 1 3350 loss: 4.7616669655
PLACES TRAINING Epoch: 1 3360 loss: 4.7585811138
PLACES TRAINING Epoch: 1 3370 loss: 4.7614290714
PLACES TRAINING Epoch: 1 3380 loss: 4.7609188557
PLACES TRAINING Epoch: 1 3390 loss: 4.7617190838
PLACES TRAINING Epoch: 1 3400 loss: 4.7583067417
PLACES TRAINING Epoch: 1 3410 loss: 4.7603348255
PLACES TRAINING Epoch: 1 3420 loss: 4.7624415874
PLACES TRAINING Epoch: 1 3430 loss: 4.7592326641
PLACES TRAINING Epoch: 1 3440 loss: 4.7591650009
PLACES TRAINING Epoch: 1 3450 loss: 4.7584899902
PLACES TRAINING Epoch: 1 3460 loss: 4.7595151901
PLACES TRAINING Epoch: 1 3470 loss: 4.7586859226
PLACES TRAINING Epoch: 1 3480 loss: 4.7585825443
PLACES TRAINING Epoch: 1 3490 loss: 4.7593651295
PLACES TRAINING Epoch: 1 3500 loss: 4.7583021641
PLACES TRAINING Epoch: 1 3510 loss: 4.7565919876
PLACES TRAINING Epoch: 1 3520 loss: 4.7597816467
PLACES TRAINING Epoch: 1 3530 loss: 4.7631779194
PLACES TRAINING Epoch: 1 3540 loss: 4.7565361500
PLACES TRAINING Epoch: 1 3550 loss: 4.7597218990
PLACES TRAINING Epoch: 1 3560 loss: 4.7611124992
PLACES TRAINING Epoch: 1 3570 loss: 4.7585959911
PLACES TRAINING Epoch: 1 3580 loss: 4.7591169357
PLACES TRAINING Epoch: 1 3590 loss: 4.7559693336
PLACES TRAINING Epoch: 1 3600 loss: 4.7566501141
PLACES TRAINING Epoch: 1 3610 loss: 4.7598305225
PLACES TRAINING Epoch: 1 3620 loss: 4.7627681732
PLACES TRAINING Epoch: 1 3630 loss: 4.7552387238
PLACES TRAINING Epoch: 1 3640 loss: 4.7578782558
PLACES TRAINING Epoch: 1 3650 loss: 4.7600572586
PLACES TRAINING Epoch: 1 3660 loss: 4.7598036766
PLACES TRAINING Epoch: 1 3670 loss: 4.7572637081
PLACES TRAINING Epoch: 1 3680 loss: 4.7584522724
PLACES TRAINING Epoch: 1 3690 loss: 4.7569078922
PLACES TRAINING Epoch: 1 3700 loss: 4.7572388649
PLACES TRAINING Epoch: 1 3710 loss: 4.7584235191
PLACES TRAINING Epoch: 1 3720 loss: 4.7593946457
PLACES TRAINING Epoch: 1 3730 loss: 4.7603552818
PLACES TRAINING Epoch: 1 3740 loss: 4.7595046997
PLACES TRAINING Epoch: 1 3750 loss: 4.7563752174
PLACES TRAINING Epoch: 1 3760 loss: 4.7589564800
PLACES TRAINING Epoch: 1 3770 loss: 4.7591719151
PLACES TRAINING Epoch: 1 3780 loss: 4.7508859634
PLACES TRAINING Epoch: 1 3790 loss: 4.7530536652
PLACES TRAINING Epoch: 1 3800 loss: 4.7541408539
PLACES TRAINING Epoch: 1 3810 loss: 4.7532649040
PLACES TRAINING Epoch: 1 3820 loss: 4.7560633183
PLACES TRAINING Epoch: 1 3830 loss: 4.7577080727
PLACES TRAINING Epoch: 1 3840 loss: 4.7561780453
PLACES TRAINING Epoch: 1 3850 loss: 4.7586032867
PLACES TRAINING Epoch: 1 3860 loss: 4.7530214787
PLACES TRAINING Epoch: 1 3870 loss: 4.7572885990
PLACES TRAINING Epoch: 1 3880 loss: 4.7565866947
PLACES TRAINING Epoch: 1 3890 loss: 4.7596075535
PLACES TRAINING Epoch: 1 3900 loss: 4.7587506294
PLACES TRAINING Epoch: 1 3910 loss: 4.7552214622
PLACES TRAINING Epoch: 1 3920 loss: 4.7570096016
PLACES TRAINING Epoch: 1 3930 loss: 4.7575015545
PLACES TRAINING Epoch: 1 3940 loss: 4.7606736660
PLACES TRAINING Epoch: 1 3950 loss: 4.7542372704
PLACES TRAINING Epoch: 1 3960 loss: 4.7596268177
PLACES TRAINING Epoch: 1 3970 loss: 4.7534315109
PLACES TRAINING Epoch: 1 3980 loss: 4.7568914413
PLACES TRAINING Epoch: 1 3990 loss: 4.7578136921
PLACES TRAINING Epoch: 1 4000 loss: 4.7542815208
PLACES TRAINING Epoch: 2 10 loss: 4.7531573296
PLACES TRAINING Epoch: 2 20 loss: 4.7578805447
PLACES TRAINING Epoch: 2 30 loss: 4.7578695297
PLACES TRAINING Epoch: 2 40 loss: 4.7534953594
PLACES TRAINING Epoch: 2 50 loss: 4.7575272083
PLACES TRAINING Epoch: 2 60 loss: 4.7553108215
PLACES TRAINING Epoch: 2 70 loss: 4.7534821987
PLACES TRAINING Epoch: 2 80 loss: 4.7569152832
PLACES TRAINING Epoch: 2 90 loss: 4.7565157413
PLACES TRAINING Epoch: 2 100 loss: 4.7557299137
PLACES TRAINING Epoch: 2 110 loss: 4.7509200096
PLACES TRAINING Epoch: 2 120 loss: 4.7572397232
PLACES TRAINING Epoch: 2 130 loss: 4.7560464859
PLACES TRAINING Epoch: 2 140 loss: 4.7550502777
PLACES TRAINING Epoch: 2 150 loss: 4.7550207138
PLACES TRAINING Epoch: 2 160 loss: 4.7573751926
PLACES TRAINING Epoch: 2 170 loss: 4.7586455345
PLACES TRAINING Epoch: 2 180 loss: 4.7548754215
PLACES TRAINING Epoch: 2 190 loss: 4.7561770439
PLACES TRAINING Epoch: 2 200 loss: 4.7527674675
PLACES TRAINING Epoch: 2 210 loss: 4.7560284615
PLACES TRAINING Epoch: 2 220 loss: 4.7567427635
PLACES TRAINING Epoch: 2 230 loss: 4.7550588131
PLACES TRAINING Epoch: 2 240 loss: 4.7539480209
PLACES TRAINING Epoch: 2 250 loss: 4.7530015945
PLACES TRAINING Epoch: 2 260 loss: 4.7523174286
PLACES TRAINING Epoch: 2 270 loss: 4.7578066349
PLACES TRAINING Epoch: 2 280 loss: 4.7556485176
PLACES TRAINING Epoch: 2 290 loss: 4.7566751480
PLACES TRAINING Epoch: 2 300 loss: 4.7546515942
PLACES TRAINING Epoch: 2 310 loss: 4.7535258770
PLACES TRAINING Epoch: 2 320 loss: 4.7574964523
PLACES TRAINING Epoch: 2 330 loss: 4.7567722797
PLACES TRAINING Epoch: 2 340 loss: 4.7555072784
PLACES TRAINING Epoch: 2 350 loss: 4.7554291248
PLACES TRAINING Epoch: 2 360 loss: 4.7583531857
PLACES TRAINING Epoch: 2 370 loss: 4.7533902168
PLACES TRAINING Epoch: 2 380 loss: 4.7568219662
PLACES TRAINING Epoch: 2 390 loss: 4.7556842327
PLACES TRAINING Epoch: 2 400 loss: 4.7563450813
PLACES TRAINING Epoch: 2 410 loss: 4.7545895100
PLACES TRAINING Epoch: 2 420 loss: 4.7547745705
PLACES TRAINING Epoch: 2 430 loss: 4.7530380726
PLACES TRAINING Epoch: 2 440 loss: 4.7568811893
PLACES TRAINING Epoch: 2 450 loss: 4.7553707123
PLACES TRAINING Epoch: 2 460 loss: 4.7525396347
PLACES TRAINING Epoch: 2 470 loss: 4.7521994114
PLACES TRAINING Epoch: 2 480 loss: 4.7568358421
PLACES TRAINING Epoch: 2 490 loss: 4.7516401291
PLACES TRAINING Epoch: 2 500 loss: 4.7521582127
PLACES TRAINING Epoch: 2 510 loss: 4.7516724586
PLACES TRAINING Epoch: 2 520 loss: 4.7518911362
PLACES TRAINING Epoch: 2 530 loss: 4.7547305107
PLACES TRAINING Epoch: 2 540 loss: 4.7561289310
PLACES TRAINING Epoch: 2 550 loss: 4.7531617165
PLACES TRAINING Epoch: 2 560 loss: 4.7566081047
PLACES TRAINING Epoch: 2 570 loss: 4.7520429134
PLACES TRAINING Epoch: 2 580 loss: 4.7536694050
PLACES TRAINING Epoch: 2 590 loss: 4.7571280956
PLACES TRAINING Epoch: 2 600 loss: 4.7553897381
PLACES TRAINING Epoch: 2 610 loss: 4.7559665680
PLACES TRAINING Epoch: 2 620 loss: 4.7533555984
PLACES TRAINING Epoch: 2 630 loss: 4.7521225452
PLACES TRAINING Epoch: 2 640 loss: 4.7518477917
PLACES TRAINING Epoch: 2 650 loss: 4.7522311211
PLACES TRAINING Epoch: 2 660 loss: 4.7530757427
PLACES TRAINING Epoch: 2 670 loss: 4.7538430691
PLACES TRAINING Epoch: 2 680 loss: 4.7527695179
PLACES TRAINING Epoch: 2 690 loss: 4.7579547882
PLACES TRAINING Epoch: 2 700 loss: 4.7539823055
PLACES TRAINING Epoch: 2 710 loss: 4.7543943405
PLACES TRAINING Epoch: 2 720 loss: 4.7549718380
PLACES TRAINING Epoch: 2 730 loss: 4.7531651497
PLACES TRAINING Epoch: 2 740 loss: 4.7569677830
PLACES TRAINING Epoch: 2 750 loss: 4.7556926250
PLACES TRAINING Epoch: 2 760 loss: 4.7526400566
PLACES TRAINING Epoch: 2 770 loss: 4.7543015480
PLACES TRAINING Epoch: 2 780 loss: 4.7567759514
PLACES TRAINING Epoch: 2 790 loss: 4.7522702217
PLACES TRAINING Epoch: 2 800 loss: 4.7533467293
PLACES TRAINING Epoch: 2 810 loss: 4.7525015831
PLACES TRAINING Epoch: 2 820 loss: 4.7504560947
PLACES TRAINING Epoch: 2 830 loss: 4.7546021938
PLACES TRAINING Epoch: 2 840 loss: 4.7558770657
PLACES TRAINING Epoch: 2 850 loss: 4.7530806065
PLACES TRAINING Epoch: 2 860 loss: 4.7503634453
PLACES TRAINING Epoch: 2 870 loss: 4.7507242680
PLACES TRAINING Epoch: 2 880 loss: 4.7535276890
PLACES TRAINING Epoch: 2 890 loss: 4.7516391754
PLACES TRAINING Epoch: 2 900 loss: 4.7559876919
PLACES TRAINING Epoch: 2 910 loss: 4.7506335258
PLACES TRAINING Epoch: 2 920 loss: 4.7503094673