Filling queue with 20000 CIFAR images before starting to train. This will take a few minutes.
2017-10-21 12:30:41.051537: step 0, loss = 4.68 (277.8 examples/sec; 1.843 sec/batch)
2017-10-21 12:30:42.736577: step 10, loss = 4.61 (17847.1 examples/sec; 0.029 sec/batch)
2017-10-21 12:30:43.386587: step 20, loss = 4.48 (11746.0 examples/sec; 0.044 sec/batch)
2017-10-21 12:30:44.277388: step 30, loss = 4.33 (11253.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:30:45.165070: step 40, loss = 4.34 (11166.4 examples/sec; 0.046 sec/batch)
2017-10-21 12:30:46.068972: step 50, loss = 4.33 (11763.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:30:46.953018: step 60, loss = 4.15 (11901.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:30:47.854777: step 70, loss = 4.26 (11150.1 examples/sec; 0.046 sec/batch)
2017-10-21 12:30:48.740744: step 80, loss = 4.17 (11258.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:30:49.625224: step 90, loss = 4.13 (11899.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:30:50.510402: step 100, loss = 4.05 (12672.4 examples/sec; 0.040 sec/batch)
2017-10-21 12:30:51.856481: step 110, loss = 3.95 (16174.5 examples/sec; 0.032 sec/batch)
2017-10-21 12:30:52.460184: step 120, loss = 3.99 (15366.8 examples/sec; 0.033 sec/batch)
2017-10-21 12:30:53.354550: step 130, loss = 3.96 (10599.3 examples/sec; 0.048 sec/batch)
2017-10-21 12:30:54.220917: step 140, loss = 3.96 (11590.1 examples/sec; 0.044 sec/batch)
2017-10-21 12:30:55.099786: step 150, loss = 4.11 (11369.3 examples/sec; 0.045 sec/batch)
2017-10-21 12:30:55.973442: step 160, loss = 3.85 (11524.4 examples/sec; 0.044 sec/batch)
2017-10-21 12:30:56.829284: step 170, loss = 3.81 (12463.8 examples/sec; 0.041 sec/batch)
2017-10-21 12:30:57.690481: step 180, loss = 3.78 (12182.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:30:58.571223: step 190, loss = 3.76 (11939.8 examples/sec; 0.043 sec/batch)
2017-10-21 12:30:59.436424: step 200, loss = 3.68 (11950.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:31:00.775563: step 210, loss = 4.02 (18140.8 examples/sec; 0.028 sec/batch)
2017-10-21 12:31:01.381243: step 220, loss = 3.71 (17243.5 examples/sec; 0.030 sec/batch)
2017-10-21 12:31:02.299562: step 230, loss = 3.78 (10510.6 examples/sec; 0.049 sec/batch)
2017-10-21 12:31:03.165493: step 240, loss = 3.71 (11497.1 examples/sec; 0.045 sec/batch)
2017-10-21 12:31:04.031698: step 250, loss = 3.59 (11979.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:31:04.916319: step 260, loss = 3.77 (11444.7 examples/sec; 0.045 sec/batch)
2017-10-21 12:31:05.780762: step 270, loss = 3.75 (11861.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:31:06.675217: step 280, loss = 3.53 (10389.0 examples/sec; 0.049 sec/batch)
2017-10-21 12:31:07.532295: step 290, loss = 3.60 (13552.7 examples/sec; 0.038 sec/batch)
2017-10-21 12:31:08.384027: step 300, loss = 3.60 (12736.2 examples/sec; 0.040 sec/batch)
2017-10-21 12:31:09.722282: step 310, loss = 3.38 (17633.8 examples/sec; 0.029 sec/batch)
2017-10-21 12:31:10.323860: step 320, loss = 3.44 (14633.0 examples/sec; 0.035 sec/batch)
2017-10-21 12:31:11.183405: step 330, loss = 3.47 (11933.4 examples/sec; 0.043 sec/batch)
2017-10-21 12:31:12.045533: step 340, loss = 3.36 (12386.5 examples/sec; 0.041 sec/batch)
2017-10-21 12:31:12.941498: step 350, loss = 3.32 (11907.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:31:13.813814: step 360, loss = 3.55 (11216.5 examples/sec; 0.046 sec/batch)
2017-10-21 12:31:14.690096: step 370, loss = 3.42 (12935.4 examples/sec; 0.040 sec/batch)
2017-10-21 12:31:15.580775: step 380, loss = 3.33 (10578.9 examples/sec; 0.048 sec/batch)
2017-10-21 12:31:16.443715: step 390, loss = 3.37 (11890.9 examples/sec; 0.043 sec/batch)
2017-10-21 12:31:17.320530: step 400, loss = 3.31 (11147.6 examples/sec; 0.046 sec/batch)
2017-10-21 12:31:18.624618: step 410, loss = 3.20 (18593.7 examples/sec; 0.028 sec/batch)
2017-10-21 12:31:19.281180: step 420, loss = 3.24 (16016.3 examples/sec; 0.032 sec/batch)
2017-10-21 12:31:20.179468: step 430, loss = 3.23 (11643.9 examples/sec; 0.044 sec/batch)
2017-10-21 12:31:21.028476: step 440, loss = 3.48 (11502.8 examples/sec; 0.045 sec/batch)
2017-10-21 12:31:21.903108: step 450, loss = 3.24 (11511.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:31:22.784281: step 460, loss = 3.20 (10399.4 examples/sec; 0.049 sec/batch)
2017-10-21 12:31:23.659360: step 470, loss = 2.99 (12683.7 examples/sec; 0.040 sec/batch)
2017-10-21 12:31:24.537137: step 480, loss = 3.19 (12140.9 examples/sec; 0.042 sec/batch)
2017-10-21 12:31:25.429671: step 490, loss = 3.04 (12071.9 examples/sec; 0.042 sec/batch)
2017-10-21 12:31:26.291773: step 500, loss = 3.15 (12110.5 examples/sec; 0.042 sec/batch)
2017-10-21 12:31:27.586351: step 510, loss = 3.08 (18321.1 examples/sec; 0.028 sec/batch)
2017-10-21 12:31:28.193427: step 520, loss = 3.17 (13535.9 examples/sec; 0.038 sec/batch)
2017-10-21 12:31:29.069402: step 530, loss = 2.99 (11974.7 examples/sec; 0.043 sec/batch)
2017-10-21 12:31:29.949672: step 540, loss = 2.94 (11349.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:31:30.829899: step 550, loss = 2.93 (11295.9 examples/sec; 0.045 sec/batch)
2017-10-21 12:31:31.748512: step 560, loss = 3.04 (11457.6 examples/sec; 0.045 sec/batch)
2017-10-21 12:31:32.629241: step 570, loss = 3.11 (11472.4 examples/sec; 0.045 sec/batch)
2017-10-21 12:31:33.500251: step 580, loss = 3.08 (12165.8 examples/sec; 0.042 sec/batch)
2017-10-21 12:31:34.372527: step 590, loss = 3.02 (11116.4 examples/sec; 0.046 sec/batch)
2017-10-21 12:31:35.248458: step 600, loss = 2.81 (11778.5 examples/sec; 0.043 sec/batch)
2017-10-21 12:31:36.544254: step 610, loss = 2.96 (17384.9 examples/sec; 0.029 sec/batch)
2017-10-21 12:31:37.179706: step 620, loss = 2.86 (10917.4 examples/sec; 0.047 sec/batch)
2017-10-21 12:31:38.066097: step 630, loss = 2.79 (11940.5 examples/sec; 0.043 sec/batch)
2017-10-21 12:31:38.937556: step 640, loss = 2.79 (12178.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:31:39.803705: step 650, loss = 2.86 (12457.7 examples/sec; 0.041 sec/batch)
2017-10-21 12:31:40.698726: step 660, loss = 2.78 (10751.3 examples/sec; 0.048 sec/batch)
2017-10-21 12:31:41.564566: step 670, loss = 2.70 (11087.9 examples/sec; 0.046 sec/batch)
2017-10-21 12:31:42.473679: step 680, loss = 2.73 (11225.9 examples/sec; 0.046 sec/batch)
2017-10-21 12:31:43.336792: step 690, loss = 2.81 (12237.0 examples/sec; 0.042 sec/batch)
2017-10-21 12:31:44.211129: step 700, loss = 2.78 (11407.3 examples/sec; 0.045 sec/batch)
2017-10-21 12:31:45.487144: step 710, loss = 2.60 (18198.3 examples/sec; 0.028 sec/batch)
2017-10-21 12:31:46.101378: step 720, loss = 2.70 (16278.1 examples/sec; 0.031 sec/batch)
2017-10-21 12:31:46.977315: step 730, loss = 2.76 (11954.7 examples/sec; 0.043 sec/batch)
2017-10-21 12:31:47.852571: step 740, loss = 2.68 (11975.6 examples/sec; 0.043 sec/batch)
2017-10-21 12:31:48.725336: step 750, loss = 2.75 (11363.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:31:49.590120: step 760, loss = 2.62 (12551.2 examples/sec; 0.041 sec/batch)
2017-10-21 12:31:50.467413: step 770, loss = 2.49 (11989.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:31:51.342112: step 780, loss = 2.84 (12308.0 examples/sec; 0.042 sec/batch)
2017-10-21 12:31:52.239727: step 790, loss = 2.60 (11351.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:31:53.145806: step 800, loss = 2.60 (12410.5 examples/sec; 0.041 sec/batch)
2017-10-21 12:31:54.443406: step 810, loss = 2.51 (19105.3 examples/sec; 0.027 sec/batch)
2017-10-21 12:31:55.057838: step 820, loss = 2.56 (12130.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:31:55.916397: step 830, loss = 2.51 (12374.2 examples/sec; 0.041 sec/batch)
2017-10-21 12:31:56.787940: step 840, loss = 2.48 (11382.3 examples/sec; 0.045 sec/batch)
2017-10-21 12:31:57.657210: step 850, loss = 2.60 (11285.4 examples/sec; 0.045 sec/batch)
2017-10-21 12:31:58.555715: step 860, loss = 2.43 (10989.3 examples/sec; 0.047 sec/batch)
2017-10-21 12:31:59.432113: step 870, loss = 2.56 (12120.1 examples/sec; 0.042 sec/batch)
2017-10-21 12:32:00.309274: step 880, loss = 2.37 (11290.8 examples/sec; 0.045 sec/batch)
2017-10-21 12:32:01.203291: step 890, loss = 2.40 (11902.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:32:02.070369: step 900, loss = 2.45 (11692.5 examples/sec; 0.044 sec/batch)
2017-10-21 12:32:03.368961: step 910, loss = 2.42 (17195.1 examples/sec; 0.030 sec/batch)
2017-10-21 12:32:03.978102: step 920, loss = 2.40 (18536.8 examples/sec; 0.028 sec/batch)
2017-10-21 12:32:04.850829: step 930, loss = 2.35 (11122.7 examples/sec; 0.046 sec/batch)
2017-10-21 12:32:05.728512: step 940, loss = 2.34 (12271.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:32:06.609704: step 950, loss = 2.41 (12974.5 examples/sec; 0.039 sec/batch)
2017-10-21 12:32:07.493094: step 960, loss = 2.56 (11194.6 examples/sec; 0.046 sec/batch)
2017-10-21 12:32:08.368741: step 970, loss = 2.32 (12206.5 examples/sec; 0.042 sec/batch)
2017-10-21 12:32:09.225001: step 980, loss = 2.35 (11714.5 examples/sec; 0.044 sec/batch)
2017-10-21 12:32:10.093379: step 990, loss = 2.35 (13891.7 examples/sec; 0.037 sec/batch)
2017-10-21 12:32:10.964604: step 1000, loss = 2.44 (11136.4 examples/sec; 0.046 sec/batch)
2017-10-21 12:32:12.354888: step 1010, loss = 2.38 (17813.6 examples/sec; 0.029 sec/batch)
2017-10-21 12:32:12.953107: step 1020, loss = 2.27 (18055.3 examples/sec; 0.028 sec/batch)
2017-10-21 12:32:13.812694: step 1030, loss = 2.36 (12088.6 examples/sec; 0.042 sec/batch)
2017-10-21 12:32:14.690819: step 1040, loss = 2.40 (11652.5 examples/sec; 0.044 sec/batch)
2017-10-21 12:32:15.578637: step 1050, loss = 2.25 (10602.3 examples/sec; 0.048 sec/batch)
2017-10-21 12:32:16.447674: step 1060, loss = 2.23 (11491.2 examples/sec; 0.045 sec/batch)
2017-10-21 12:32:17.334336: step 1070, loss = 2.21 (12663.5 examples/sec; 0.040 sec/batch)
2017-10-21 12:32:18.222834: step 1080, loss = 2.39 (11675.1 examples/sec; 0.044 sec/batch)
2017-10-21 12:32:19.111262: step 1090, loss = 2.30 (11775.6 examples/sec; 0.043 sec/batch)
2017-10-21 12:32:19.990807: step 1100, loss = 2.30 (11792.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:32:21.278193: step 1110, loss = 2.23 (17549.3 examples/sec; 0.029 sec/batch)
2017-10-21 12:32:21.882903: step 1120, loss = 2.26 (14079.4 examples/sec; 0.036 sec/batch)
2017-10-21 12:32:22.767826: step 1130, loss = 2.30 (12114.6 examples/sec; 0.042 sec/batch)
2017-10-21 12:32:23.637183: step 1140, loss = 2.10 (11542.9 examples/sec; 0.044 sec/batch)
2017-10-21 12:32:24.515316: step 1150, loss = 2.11 (11318.7 examples/sec; 0.045 sec/batch)
2017-10-21 12:32:25.398423: step 1160, loss = 2.10 (11478.3 examples/sec; 0.045 sec/batch)
2017-10-21 12:32:26.286040: step 1170, loss = 2.11 (11385.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:32:27.177585: step 1180, loss = 2.28 (11527.5 examples/sec; 0.044 sec/batch)
2017-10-21 12:32:28.050917: step 1190, loss = 2.05 (11983.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:32:28.935740: step 1200, loss = 2.06 (11311.9 examples/sec; 0.045 sec/batch)
2017-10-21 12:32:30.192844: step 1210, loss = 2.10 (18194.8 examples/sec; 0.028 sec/batch)
2017-10-21 12:32:30.804495: step 1220, loss = 2.11 (11860.4 examples/sec; 0.043 sec/batch)
2017-10-21 12:32:31.681886: step 1230, loss = 1.98 (11936.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:32:32.537324: step 1240, loss = 2.03 (12131.5 examples/sec; 0.042 sec/batch)
2017-10-21 12:32:33.420860: step 1250, loss = 2.14 (11111.3 examples/sec; 0.046 sec/batch)
2017-10-21 12:32:34.306038: step 1260, loss = 2.00 (10503.6 examples/sec; 0.049 sec/batch)
2017-10-21 12:32:35.170587: step 1270, loss = 1.93 (11923.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:32:36.060723: step 1280, loss = 2.11 (11176.3 examples/sec; 0.046 sec/batch)
2017-10-21 12:32:36.947428: step 1290, loss = 1.99 (10970.8 examples/sec; 0.047 sec/batch)
2017-10-21 12:32:37.811720: step 1300, loss = 2.04 (12017.9 examples/sec; 0.043 sec/batch)
2017-10-21 12:32:39.075196: step 1310, loss = 1.82 (18196.8 examples/sec; 0.028 sec/batch)
2017-10-21 12:32:39.677776: step 1320, loss = 1.93 (17862.0 examples/sec; 0.029 sec/batch)
2017-10-21 12:32:40.539517: step 1330, loss = 1.97 (11212.6 examples/sec; 0.046 sec/batch)
2017-10-21 12:32:41.414027: step 1340, loss = 1.94 (12246.4 examples/sec; 0.042 sec/batch)
2017-10-21 12:32:42.282865: step 1350, loss = 1.96 (11755.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:32:43.155724: step 1360, loss = 1.99 (12086.3 examples/sec; 0.042 sec/batch)
2017-10-21 12:32:44.035763: step 1370, loss = 1.92 (10957.4 examples/sec; 0.047 sec/batch)
2017-10-21 12:32:44.894919: step 1380, loss = 1.91 (12346.7 examples/sec; 0.041 sec/batch)
2017-10-21 12:32:45.764153: step 1390, loss = 1.88 (12402.8 examples/sec; 0.041 sec/batch)
2017-10-21 12:32:46.644129: step 1400, loss = 1.91 (11362.7 examples/sec; 0.045 sec/batch)
2017-10-21 12:32:47.911023: step 1410, loss = 1.91 (18478.7 examples/sec; 0.028 sec/batch)
2017-10-21 12:32:48.519430: step 1420, loss = 1.93 (12057.3 examples/sec; 0.042 sec/batch)
2017-10-21 12:32:49.382650: step 1430, loss = 1.95 (11447.8 examples/sec; 0.045 sec/batch)
2017-10-21 12:32:50.258771: step 1440, loss = 1.82 (11923.4 examples/sec; 0.043 sec/batch)
2017-10-21 12:32:51.132126: step 1450, loss = 1.89 (12627.1 examples/sec; 0.041 sec/batch)
2017-10-21 12:32:52.013478: step 1460, loss = 1.96 (12202.2 examples/sec; 0.042 sec/batch)
2017-10-21 12:32:52.896116: step 1470, loss = 1.88 (10600.4 examples/sec; 0.048 sec/batch)
2017-10-21 12:32:53.769528: step 1480, loss = 1.82 (13223.9 examples/sec; 0.039 sec/batch)
2017-10-21 12:32:54.644306: step 1490, loss = 1.82 (12023.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:32:55.526417: step 1500, loss = 1.76 (12274.3 examples/sec; 0.042 sec/batch)
2017-10-21 12:32:56.784389: step 1510, loss = 1.78 (17902.9 examples/sec; 0.029 sec/batch)
2017-10-21 12:32:57.419729: step 1520, loss = 1.67 (11109.4 examples/sec; 0.046 sec/batch)
2017-10-21 12:32:58.282811: step 1530, loss = 1.91 (11456.3 examples/sec; 0.045 sec/batch)
2017-10-21 12:32:59.161573: step 1540, loss = 1.82 (12093.3 examples/sec; 0.042 sec/batch)
2017-10-21 12:33:00.039645: step 1550, loss = 1.75 (11814.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:33:00.922327: step 1560, loss = 1.98 (11578.1 examples/sec; 0.044 sec/batch)
2017-10-21 12:33:01.788345: step 1570, loss = 1.79 (12587.5 examples/sec; 0.041 sec/batch)
2017-10-21 12:33:02.678426: step 1580, loss = 1.78 (11917.8 examples/sec; 0.043 sec/batch)
2017-10-21 12:33:03.545326: step 1590, loss = 1.79 (11206.1 examples/sec; 0.046 sec/batch)
2017-10-21 12:33:04.437968: step 1600, loss = 1.79 (10988.1 examples/sec; 0.047 sec/batch)
2017-10-21 12:33:05.699598: step 1610, loss = 1.79 (18367.5 examples/sec; 0.028 sec/batch)
2017-10-21 12:33:06.309600: step 1620, loss = 1.70 (15790.4 examples/sec; 0.032 sec/batch)
2017-10-21 12:33:07.203870: step 1630, loss = 1.75 (11824.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:33:08.076292: step 1640, loss = 1.68 (13165.2 examples/sec; 0.039 sec/batch)
2017-10-21 12:33:08.959397: step 1650, loss = 1.66 (11731.5 examples/sec; 0.044 sec/batch)
2017-10-21 12:33:09.832182: step 1660, loss = 1.81 (11548.3 examples/sec; 0.044 sec/batch)
2017-10-21 12:33:10.701956: step 1670, loss = 1.71 (11768.3 examples/sec; 0.044 sec/batch)
2017-10-21 12:33:11.568780: step 1680, loss = 1.73 (12038.7 examples/sec; 0.043 sec/batch)
2017-10-21 12:33:12.445785: step 1690, loss = 1.71 (11112.0 examples/sec; 0.046 sec/batch)
2017-10-21 12:33:13.341331: step 1700, loss = 1.68 (11228.4 examples/sec; 0.046 sec/batch)
2017-10-21 12:33:14.612109: step 1710, loss = 1.72 (17478.7 examples/sec; 0.029 sec/batch)
2017-10-21 12:33:15.213502: step 1720, loss = 1.64 (11532.1 examples/sec; 0.044 sec/batch)
2017-10-21 12:33:16.091027: step 1730, loss = 1.90 (11787.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:33:16.970220: step 1740, loss = 1.67 (11000.4 examples/sec; 0.047 sec/batch)
2017-10-21 12:33:17.855496: step 1750, loss = 1.66 (10758.8 examples/sec; 0.048 sec/batch)
2017-10-21 12:33:18.728295: step 1760, loss = 1.62 (10712.9 examples/sec; 0.048 sec/batch)
2017-10-21 12:33:19.602718: step 1770, loss = 1.54 (11217.3 examples/sec; 0.046 sec/batch)
2017-10-21 12:33:20.470026: step 1780, loss = 1.70 (11489.3 examples/sec; 0.045 sec/batch)
2017-10-21 12:33:21.351199: step 1790, loss = 1.67 (10704.0 examples/sec; 0.048 sec/batch)
2017-10-21 12:33:22.244150: step 1800, loss = 1.60 (11450.2 examples/sec; 0.045 sec/batch)
2017-10-21 12:33:23.534287: step 1810, loss = 1.62 (15482.4 examples/sec; 0.033 sec/batch)
2017-10-21 12:33:24.134232: step 1820, loss = 1.59 (17368.8 examples/sec; 0.029 sec/batch)
2017-10-21 12:33:24.978433: step 1830, loss = 1.62 (11768.3 examples/sec; 0.044 sec/batch)
2017-10-21 12:33:25.855490: step 1840, loss = 1.74 (11866.4 examples/sec; 0.043 sec/batch)
2017-10-21 12:33:26.729386: step 1850, loss = 1.61 (11751.7 examples/sec; 0.044 sec/batch)
2017-10-21 12:33:27.598111: step 1860, loss = 1.64 (12464.7 examples/sec; 0.041 sec/batch)
2017-10-21 12:33:28.484705: step 1870, loss = 1.54 (11696.2 examples/sec; 0.044 sec/batch)
2017-10-21 12:33:29.361337: step 1880, loss = 1.53 (11873.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:33:30.246112: step 1890, loss = 1.57 (12521.9 examples/sec; 0.041 sec/batch)
2017-10-21 12:33:31.126056: step 1900, loss = 1.49 (12847.6 examples/sec; 0.040 sec/batch)
2017-10-21 12:33:32.413215: step 1910, loss = 1.54 (17706.7 examples/sec; 0.029 sec/batch)
2017-10-21 12:33:33.013459: step 1920, loss = 1.55 (17420.1 examples/sec; 0.029 sec/batch)
2017-10-21 12:33:33.898329: step 1930, loss = 1.49 (11273.2 examples/sec; 0.045 sec/batch)
2017-10-21 12:33:34.770716: step 1940, loss = 1.59 (11324.7 examples/sec; 0.045 sec/batch)
2017-10-21 12:33:35.647257: step 1950, loss = 1.52 (11306.4 examples/sec; 0.045 sec/batch)
2017-10-21 12:33:36.519814: step 1960, loss = 1.52 (12075.0 examples/sec; 0.042 sec/batch)
2017-10-21 12:33:37.408679: step 1970, loss = 1.53 (11171.3 examples/sec; 0.046 sec/batch)
2017-10-21 12:33:38.289415: step 1980, loss = 1.47 (11983.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:33:39.169667: step 1990, loss = 1.47 (11835.5 examples/sec; 0.043 sec/batch)
2017-10-21 12:33:40.059845: step 2000, loss = 1.53 (11598.1 examples/sec; 0.044 sec/batch)
2017-10-21 12:33:41.463960: step 2010, loss = 1.47 (16352.0 examples/sec; 0.031 sec/batch)
2017-10-21 12:33:42.060697: step 2020, loss = 1.54 (15763.4 examples/sec; 0.032 sec/batch)
2017-10-21 12:33:42.945429: step 2030, loss = 1.51 (11739.8 examples/sec; 0.044 sec/batch)
2017-10-21 12:33:43.806788: step 2040, loss = 1.60 (12391.7 examples/sec; 0.041 sec/batch)
2017-10-21 12:33:44.679777: step 2050, loss = 1.46 (13431.6 examples/sec; 0.038 sec/batch)
2017-10-21 12:33:45.580168: step 2060, loss = 1.40 (12548.5 examples/sec; 0.041 sec/batch)
2017-10-21 12:33:46.463922: step 2070, loss = 1.39 (12326.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:33:47.358742: step 2080, loss = 1.47 (12563.6 examples/sec; 0.041 sec/batch)
2017-10-21 12:33:48.225236: step 2090, loss = 1.47 (12062.8 examples/sec; 0.042 sec/batch)
2017-10-21 12:33:49.112915: step 2100, loss = 1.43 (10757.3 examples/sec; 0.048 sec/batch)
2017-10-21 12:33:50.385281: step 2110, loss = 1.71 (16453.5 examples/sec; 0.031 sec/batch)
2017-10-21 12:33:50.968643: step 2120, loss = 1.45 (17525.5 examples/sec; 0.029 sec/batch)
2017-10-21 12:33:51.856641: step 2130, loss = 1.36 (10713.0 examples/sec; 0.048 sec/batch)
2017-10-21 12:33:52.713852: step 2140, loss = 1.52 (12235.0 examples/sec; 0.042 sec/batch)
2017-10-21 12:33:53.595359: step 2150, loss = 1.45 (10773.2 examples/sec; 0.048 sec/batch)
2017-10-21 12:33:54.472832: step 2160, loss = 1.43 (11704.9 examples/sec; 0.044 sec/batch)
2017-10-21 12:33:55.368674: step 2170, loss = 1.38 (12188.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:33:56.256693: step 2180, loss = 1.39 (11962.8 examples/sec; 0.043 sec/batch)
2017-10-21 12:33:57.120013: step 2190, loss = 1.33 (12273.4 examples/sec; 0.042 sec/batch)
2017-10-21 12:33:58.017461: step 2200, loss = 1.42 (10886.3 examples/sec; 0.047 sec/batch)
2017-10-21 12:33:59.292503: step 2210, loss = 1.55 (18188.5 examples/sec; 0.028 sec/batch)
2017-10-21 12:33:59.908023: step 2220, loss = 1.44 (12530.1 examples/sec; 0.041 sec/batch)
2017-10-21 12:34:00.815624: step 2230, loss = 1.45 (12246.9 examples/sec; 0.042 sec/batch)
2017-10-21 12:34:01.688796: step 2240, loss = 1.42 (11776.4 examples/sec; 0.043 sec/batch)
2017-10-21 12:34:02.565809: step 2250, loss = 1.42 (11687.1 examples/sec; 0.044 sec/batch)
2017-10-21 12:34:03.428766: step 2260, loss = 1.35 (11834.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:34:04.322782: step 2270, loss = 1.31 (11408.7 examples/sec; 0.045 sec/batch)
2017-10-21 12:34:05.196315: step 2280, loss = 1.39 (11277.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:34:06.064627: step 2290, loss = 1.34 (12166.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:34:06.933026: step 2300, loss = 1.37 (12072.5 examples/sec; 0.042 sec/batch)
2017-10-21 12:34:08.212677: step 2310, loss = 1.36 (17753.2 examples/sec; 0.029 sec/batch)
2017-10-21 12:34:08.826251: step 2320, loss = 1.32 (15377.5 examples/sec; 0.033 sec/batch)
2017-10-21 12:34:09.693530: step 2330, loss = 1.29 (11597.3 examples/sec; 0.044 sec/batch)
2017-10-21 12:34:10.570414: step 2340, loss = 1.27 (11475.7 examples/sec; 0.045 sec/batch)
2017-10-21 12:34:11.456657: step 2350, loss = 1.22 (11321.5 examples/sec; 0.045 sec/batch)
2017-10-21 12:34:12.316967: step 2360, loss = 1.28 (11871.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:34:13.195515: step 2370, loss = 1.37 (12273.9 examples/sec; 0.042 sec/batch)
2017-10-21 12:34:14.068540: step 2380, loss = 1.33 (11825.5 examples/sec; 0.043 sec/batch)
2017-10-21 12:34:14.965204: step 2390, loss = 1.38 (11266.6 examples/sec; 0.045 sec/batch)
2017-10-21 12:34:15.850618: step 2400, loss = 1.38 (11710.3 examples/sec; 0.044 sec/batch)
2017-10-21 12:34:17.110193: step 2410, loss = 1.28 (17708.6 examples/sec; 0.029 sec/batch)
2017-10-21 12:34:17.706954: step 2420, loss = 1.32 (17039.7 examples/sec; 0.030 sec/batch)
2017-10-21 12:34:18.596551: step 2430, loss = 1.32 (11278.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:34:19.459722: step 2440, loss = 1.33 (11965.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:34:20.335879: step 2450, loss = 1.39 (11421.4 examples/sec; 0.045 sec/batch)
2017-10-21 12:34:21.220798: step 2460, loss = 1.23 (12178.6 examples/sec; 0.042 sec/batch)
2017-10-21 12:34:22.095581: step 2470, loss = 1.22 (12068.1 examples/sec; 0.042 sec/batch)
2017-10-21 12:34:22.964431: step 2480, loss = 1.34 (12503.5 examples/sec; 0.041 sec/batch)
2017-10-21 12:34:23.831808: step 2490, loss = 1.25 (11365.6 examples/sec; 0.045 sec/batch)
2017-10-21 12:34:24.732641: step 2500, loss = 1.45 (10007.1 examples/sec; 0.051 sec/batch)
2017-10-21 12:34:26.045491: step 2510, loss = 1.38 (16599.5 examples/sec; 0.031 sec/batch)
2017-10-21 12:34:26.654304: step 2520, loss = 1.24 (13246.1 examples/sec; 0.039 sec/batch)
2017-10-21 12:34:27.536309: step 2530, loss = 1.39 (11943.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:34:28.415532: step 2540, loss = 1.19 (11421.7 examples/sec; 0.045 sec/batch)
2017-10-21 12:34:29.290224: step 2550, loss = 1.48 (13327.3 examples/sec; 0.038 sec/batch)
2017-10-21 12:34:30.177378: step 2560, loss = 1.27 (11515.8 examples/sec; 0.044 sec/batch)
2017-10-21 12:34:31.044073: step 2570, loss = 1.23 (11593.4 examples/sec; 0.044 sec/batch)
2017-10-21 12:34:31.933295: step 2580, loss = 1.12 (11889.7 examples/sec; 0.043 sec/batch)
2017-10-21 12:34:32.826345: step 2590, loss = 1.17 (11952.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:34:33.713255: step 2600, loss = 1.25 (11274.8 examples/sec; 0.045 sec/batch)
2017-10-21 12:34:35.014781: step 2610, loss = 1.16 (18079.1 examples/sec; 0.028 sec/batch)
2017-10-21 12:34:35.618517: step 2620, loss = 1.25 (17438.2 examples/sec; 0.029 sec/batch)
2017-10-21 12:34:36.506117: step 2630, loss = 1.36 (11843.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:34:37.380469: step 2640, loss = 1.34 (11651.0 examples/sec; 0.044 sec/batch)
2017-10-21 12:34:38.255027: step 2650, loss = 1.21 (12228.5 examples/sec; 0.042 sec/batch)
2017-10-21 12:34:39.122586: step 2660, loss = 1.15 (12223.0 examples/sec; 0.042 sec/batch)
2017-10-21 12:34:39.984864: step 2670, loss = 1.37 (11348.5 examples/sec; 0.045 sec/batch)
2017-10-21 12:34:40.862897: step 2680, loss = 1.26 (11207.3 examples/sec; 0.046 sec/batch)
2017-10-21 12:34:41.723826: step 2690, loss = 1.31 (12094.5 examples/sec; 0.042 sec/batch)
2017-10-21 12:34:42.602464: step 2700, loss = 1.14 (12458.8 examples/sec; 0.041 sec/batch)
2017-10-21 12:34:43.887553: step 2710, loss = 1.25 (16598.3 examples/sec; 0.031 sec/batch)
2017-10-21 12:34:44.489775: step 2720, loss = 1.43 (17910.8 examples/sec; 0.029 sec/batch)
2017-10-21 12:34:45.358912: step 2730, loss = 1.25 (11708.3 examples/sec; 0.044 sec/batch)
2017-10-21 12:34:46.240853: step 2740, loss = 1.15 (11795.7 examples/sec; 0.043 sec/batch)
2017-10-21 12:34:47.117491: step 2750, loss = 1.28 (12833.0 examples/sec; 0.040 sec/batch)
2017-10-21 12:34:47.974749: step 2760, loss = 1.36 (11841.4 examples/sec; 0.043 sec/batch)
2017-10-21 12:34:48.833453: step 2770, loss = 1.17 (12674.0 examples/sec; 0.040 sec/batch)
2017-10-21 12:34:49.713830: step 2780, loss = 1.21 (11455.3 examples/sec; 0.045 sec/batch)
2017-10-21 12:34:50.578681: step 2790, loss = 1.38 (12477.6 examples/sec; 0.041 sec/batch)
2017-10-21 12:34:51.453735: step 2800, loss = 1.13 (11584.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:34:52.722754: step 2810, loss = 1.11 (18512.5 examples/sec; 0.028 sec/batch)
2017-10-21 12:34:53.342020: step 2820, loss = 1.28 (12530.5 examples/sec; 0.041 sec/batch)
2017-10-21 12:34:54.234089: step 2830, loss = 1.19 (12576.0 examples/sec; 0.041 sec/batch)
2017-10-21 12:34:55.114740: step 2840, loss = 1.22 (12179.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:34:55.995795: step 2850, loss = 1.09 (10986.8 examples/sec; 0.047 sec/batch)
2017-10-21 12:34:56.869909: step 2860, loss = 1.17 (11615.8 examples/sec; 0.044 sec/batch)
2017-10-21 12:34:57.742599: step 2870, loss = 1.13 (12062.0 examples/sec; 0.042 sec/batch)
2017-10-21 12:34:58.628643: step 2880, loss = 1.24 (12237.2 examples/sec; 0.042 sec/batch)
2017-10-21 12:34:59.508937: step 2890, loss = 1.12 (10878.5 examples/sec; 0.047 sec/batch)
2017-10-21 12:35:00.371963: step 2900, loss = 1.20 (11186.5 examples/sec; 0.046 sec/batch)
2017-10-21 12:35:01.731101: step 2910, loss = 1.17 (17106.4 examples/sec; 0.030 sec/batch)
2017-10-21 12:35:02.332962: step 2920, loss = 1.18 (17641.3 examples/sec; 0.029 sec/batch)
2017-10-21 12:35:03.204968: step 2930, loss = 1.17 (10497.5 examples/sec; 0.049 sec/batch)
2017-10-21 12:35:04.089249: step 2940, loss = 1.00 (11407.4 examples/sec; 0.045 sec/batch)
2017-10-21 12:35:04.976165: step 2950, loss = 1.22 (10855.0 examples/sec; 0.047 sec/batch)
2017-10-21 12:35:05.877815: step 2960, loss = 1.13 (10445.3 examples/sec; 0.049 sec/batch)
2017-10-21 12:35:06.751388: step 2970, loss = 1.08 (11582.9 examples/sec; 0.044 sec/batch)
2017-10-21 12:35:07.638236: step 2980, loss = 1.13 (12099.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:35:08.522233: step 2990, loss = 1.14 (11280.5 examples/sec; 0.045 sec/batch)
2017-10-21 12:35:09.422521: step 3000, loss = 1.12 (11088.1 examples/sec; 0.046 sec/batch)
2017-10-21 12:35:10.786600: step 3010, loss = 1.10 (17285.0 examples/sec; 0.030 sec/batch)
2017-10-21 12:35:11.406617: step 3020, loss = 1.11 (15964.0 examples/sec; 0.032 sec/batch)
2017-10-21 12:35:12.278148: step 3030, loss = 1.11 (11021.9 examples/sec; 0.046 sec/batch)
2017-10-21 12:35:13.149828: step 3040, loss = 1.02 (12932.4 examples/sec; 0.040 sec/batch)
2017-10-21 12:35:14.012552: step 3050, loss = 1.17 (12056.9 examples/sec; 0.042 sec/batch)
2017-10-21 12:35:14.889065: step 3060, loss = 1.11 (12548.9 examples/sec; 0.041 sec/batch)
2017-10-21 12:35:15.765318: step 3070, loss = 1.26 (11200.7 examples/sec; 0.046 sec/batch)
2017-10-21 12:35:16.644714: step 3080, loss = 1.02 (10972.6 examples/sec; 0.047 sec/batch)
2017-10-21 12:35:17.510391: step 3090, loss = 1.15 (12252.6 examples/sec; 0.042 sec/batch)
2017-10-21 12:35:18.383354: step 3100, loss = 1.03 (12257.3 examples/sec; 0.042 sec/batch)
2017-10-21 12:35:19.673965: step 3110, loss = 1.09 (17883.6 examples/sec; 0.029 sec/batch)
2017-10-21 12:35:20.266992: step 3120, loss = 1.05 (17916.1 examples/sec; 0.029 sec/batch)
2017-10-21 12:35:21.125521: step 3130, loss = 1.01 (11786.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:35:22.017780: step 3140, loss = 1.13 (12442.3 examples/sec; 0.041 sec/batch)
2017-10-21 12:35:22.899684: step 3150, loss = 1.12 (11079.7 examples/sec; 0.046 sec/batch)
2017-10-21 12:35:23.770139: step 3160, loss = 1.08 (12240.4 examples/sec; 0.042 sec/batch)
2017-10-21 12:35:24.643591: step 3170, loss = 1.14 (11721.9 examples/sec; 0.044 sec/batch)
2017-10-21 12:35:25.530501: step 3180, loss = 1.05 (11576.7 examples/sec; 0.044 sec/batch)
2017-10-21 12:35:26.407032: step 3190, loss = 1.01 (11327.2 examples/sec; 0.045 sec/batch)
2017-10-21 12:35:27.284473: step 3200, loss = 1.20 (11507.9 examples/sec; 0.044 sec/batch)
2017-10-21 12:35:28.558429: step 3210, loss = 1.02 (18077.0 examples/sec; 0.028 sec/batch)
2017-10-21 12:35:29.173487: step 3220, loss = 1.02 (17348.9 examples/sec; 0.030 sec/batch)
2017-10-21 12:35:30.034879: step 3230, loss = 0.97 (11170.1 examples/sec; 0.046 sec/batch)
2017-10-21 12:35:30.897999: step 3240, loss = 1.12 (11709.4 examples/sec; 0.044 sec/batch)
2017-10-21 12:35:31.777706: step 3250, loss = 1.11 (11875.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:35:32.656641: step 3260, loss = 1.17 (11209.2 examples/sec; 0.046 sec/batch)
2017-10-21 12:35:33.534422: step 3270, loss = 1.07 (11425.1 examples/sec; 0.045 sec/batch)
2017-10-21 12:35:34.439441: step 3280, loss = 1.09 (10932.4 examples/sec; 0.047 sec/batch)
2017-10-21 12:35:35.356210: step 3290, loss = 1.23 (11142.5 examples/sec; 0.046 sec/batch)
2017-10-21 12:35:36.262055: step 3300, loss = 1.10 (11494.7 examples/sec; 0.045 sec/batch)
2017-10-21 12:35:37.545106: step 3310, loss = 1.04 (17900.9 examples/sec; 0.029 sec/batch)
2017-10-21 12:35:38.141077: step 3320, loss = 1.05 (15360.4 examples/sec; 0.033 sec/batch)
2017-10-21 12:35:39.019359: step 3330, loss = 1.03 (11268.9 examples/sec; 0.045 sec/batch)
2017-10-21 12:35:39.950093: step 3340, loss = 1.05 (13222.5 examples/sec; 0.039 sec/batch)
2017-10-21 12:35:40.872287: step 3350, loss = 0.92 (11021.5 examples/sec; 0.046 sec/batch)
2017-10-21 12:35:41.775426: step 3360, loss = 0.98 (10401.5 examples/sec; 0.049 sec/batch)
2017-10-21 12:35:42.701302: step 3370, loss = 1.06 (11036.7 examples/sec; 0.046 sec/batch)
2017-10-21 12:35:43.578799: step 3380, loss = 1.09 (12024.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:35:44.448376: step 3390, loss = 1.03 (11381.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:35:45.376922: step 3400, loss = 1.01 (11506.7 examples/sec; 0.044 sec/batch)
2017-10-21 12:35:46.711357: step 3410, loss = 1.03 (18373.8 examples/sec; 0.028 sec/batch)
2017-10-21 12:35:47.324924: step 3420, loss = 1.12 (13074.9 examples/sec; 0.039 sec/batch)
2017-10-21 12:35:48.246286: step 3430, loss = 1.04 (10792.0 examples/sec; 0.047 sec/batch)
2017-10-21 12:35:49.139291: step 3440, loss = 1.07 (12430.3 examples/sec; 0.041 sec/batch)
2017-10-21 12:35:50.087779: step 3450, loss = 1.01 (10003.4 examples/sec; 0.051 sec/batch)
2017-10-21 12:35:50.995231: step 3460, loss = 1.00 (10901.7 examples/sec; 0.047 sec/batch)
2017-10-21 12:35:51.879584: step 3470, loss = 1.04 (12231.9 examples/sec; 0.042 sec/batch)
2017-10-21 12:35:52.771971: step 3480, loss = 1.00 (11818.9 examples/sec; 0.043 sec/batch)
2017-10-21 12:35:53.691146: step 3490, loss = 0.96 (10419.6 examples/sec; 0.049 sec/batch)
2017-10-21 12:35:54.639208: step 3500, loss = 1.05 (10517.8 examples/sec; 0.049 sec/batch)
2017-10-21 12:35:55.917838: step 3510, loss = 1.07 (18445.4 examples/sec; 0.028 sec/batch)
2017-10-21 12:35:56.509930: step 3520, loss = 1.09 (14993.4 examples/sec; 0.034 sec/batch)
2017-10-21 12:35:57.403307: step 3530, loss = 0.98 (11666.7 examples/sec; 0.044 sec/batch)
2017-10-21 12:35:58.332314: step 3540, loss = 0.94 (10182.5 examples/sec; 0.050 sec/batch)
2017-10-21 12:35:59.283184: step 3550, loss = 1.09 (11477.8 examples/sec; 0.045 sec/batch)
2017-10-21 12:36:00.161091: step 3560, loss = 0.89 (11477.1 examples/sec; 0.045 sec/batch)
2017-10-21 12:36:01.139262: step 3570, loss = 1.01 (11266.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:36:02.043391: step 3580, loss = 1.04 (10884.2 examples/sec; 0.047 sec/batch)
2017-10-21 12:36:02.915350: step 3590, loss = 0.94 (11617.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:36:03.811404: step 3600, loss = 0.97 (10288.6 examples/sec; 0.050 sec/batch)
2017-10-21 12:36:05.102736: step 3610, loss = 1.07 (15297.7 examples/sec; 0.033 sec/batch)
2017-10-21 12:36:05.699331: step 3620, loss = 0.99 (15251.6 examples/sec; 0.034 sec/batch)
2017-10-21 12:36:06.617578: step 3630, loss = 1.05 (10538.0 examples/sec; 0.049 sec/batch)
2017-10-21 12:36:07.500309: step 3640, loss = 0.92 (12023.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:36:08.413381: step 3650, loss = 0.98 (12053.5 examples/sec; 0.042 sec/batch)
2017-10-21 12:36:09.300408: step 3660, loss = 1.01 (11129.4 examples/sec; 0.046 sec/batch)
2017-10-21 12:36:10.169313: step 3670, loss = 0.95 (11841.4 examples/sec; 0.043 sec/batch)
2017-10-21 12:36:11.066603: step 3680, loss = 1.11 (10271.2 examples/sec; 0.050 sec/batch)
2017-10-21 12:36:11.942003: step 3690, loss = 0.96 (11350.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:36:12.820045: step 3700, loss = 0.95 (11812.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:36:14.088220: step 3710, loss = 1.10 (18265.9 examples/sec; 0.028 sec/batch)
2017-10-21 12:36:14.688580: step 3720, loss = 0.95 (13088.0 examples/sec; 0.039 sec/batch)
2017-10-21 12:36:15.546446: step 3730, loss = 0.99 (11652.2 examples/sec; 0.044 sec/batch)
2017-10-21 12:36:16.426988: step 3740, loss = 0.99 (11165.5 examples/sec; 0.046 sec/batch)
2017-10-21 12:36:17.294618: step 3750, loss = 0.98 (11219.1 examples/sec; 0.046 sec/batch)
2017-10-21 12:36:18.171771: step 3760, loss = 0.96 (12282.2 examples/sec; 0.042 sec/batch)
2017-10-21 12:36:19.040297: step 3770, loss = 1.06 (11933.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:36:19.921797: step 3780, loss = 0.92 (11316.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:36:20.808021: step 3790, loss = 0.95 (11437.1 examples/sec; 0.045 sec/batch)
2017-10-21 12:36:21.665516: step 3800, loss = 1.11 (11778.7 examples/sec; 0.043 sec/batch)
2017-10-21 12:36:22.938897: step 3810, loss = 1.02 (17624.6 examples/sec; 0.029 sec/batch)
2017-10-21 12:36:23.571837: step 3820, loss = 0.95 (10531.5 examples/sec; 0.049 sec/batch)
2017-10-21 12:36:24.467169: step 3830, loss = 1.09 (10891.1 examples/sec; 0.047 sec/batch)
2017-10-21 12:36:25.346204: step 3840, loss = 0.87 (11005.1 examples/sec; 0.047 sec/batch)
2017-10-21 12:36:26.220849: step 3850, loss = 1.03 (11674.8 examples/sec; 0.044 sec/batch)
2017-10-21 12:36:27.097869: step 3860, loss = 0.97 (11421.4 examples/sec; 0.045 sec/batch)
2017-10-21 12:36:27.963450: step 3870, loss = 0.92 (12952.1 examples/sec; 0.040 sec/batch)
2017-10-21 12:36:28.828661: step 3880, loss = 0.92 (12359.3 examples/sec; 0.041 sec/batch)
2017-10-21 12:36:29.706627: step 3890, loss = 1.04 (11708.4 examples/sec; 0.044 sec/batch)
2017-10-21 12:36:30.594541: step 3900, loss = 0.91 (11360.8 examples/sec; 0.045 sec/batch)
2017-10-21 12:36:31.877976: step 3910, loss = 1.01 (16360.0 examples/sec; 0.031 sec/batch)
2017-10-21 12:36:32.464881: step 3920, loss = 0.96 (15732.3 examples/sec; 0.033 sec/batch)
2017-10-21 12:36:33.333005: step 3930, loss = 0.94 (11805.7 examples/sec; 0.043 sec/batch)
2017-10-21 12:36:34.227516: step 3940, loss = 0.91 (11690.8 examples/sec; 0.044 sec/batch)
2017-10-21 12:36:35.102593: step 3950, loss = 0.94 (11903.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:36:35.989164: step 3960, loss = 0.95 (11646.8 examples/sec; 0.044 sec/batch)
2017-10-21 12:36:36.883593: step 3970, loss = 0.95 (11517.8 examples/sec; 0.044 sec/batch)
2017-10-21 12:36:37.778012: step 3980, loss = 0.88 (11390.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:36:38.660972: step 3990, loss = 1.08 (12299.9 examples/sec; 0.042 sec/batch)
2017-10-21 12:36:39.532102: step 4000, loss = 0.93 (11790.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:36:40.916690: step 4010, loss = 0.94 (18860.2 examples/sec; 0.027 sec/batch)
2017-10-21 12:36:41.530124: step 4020, loss = 0.98 (17497.5 examples/sec; 0.029 sec/batch)
2017-10-21 12:36:42.404788: step 4030, loss = 1.00 (12194.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:36:43.304736: step 4040, loss = 0.97 (11251.5 examples/sec; 0.046 sec/batch)
2017-10-21 12:36:44.201152: step 4050, loss = 0.87 (11551.4 examples/sec; 0.044 sec/batch)
2017-10-21 12:36:45.078386: step 4060, loss = 0.94 (10690.3 examples/sec; 0.048 sec/batch)
2017-10-21 12:36:45.955759: step 4070, loss = 1.08 (11345.2 examples/sec; 0.045 sec/batch)
2017-10-21 12:36:46.845103: step 4080, loss = 0.94 (11730.2 examples/sec; 0.044 sec/batch)
2017-10-21 12:36:47.728308: step 4090, loss = 0.98 (12459.0 examples/sec; 0.041 sec/batch)
2017-10-21 12:36:48.608608: step 4100, loss = 0.90 (11170.6 examples/sec; 0.046 sec/batch)
2017-10-21 12:36:49.895971: step 4110, loss = 0.91 (17562.6 examples/sec; 0.029 sec/batch)
2017-10-21 12:36:50.503595: step 4120, loss = 0.94 (14195.9 examples/sec; 0.036 sec/batch)
2017-10-21 12:36:51.380044: step 4130, loss = 0.99 (12603.3 examples/sec; 0.041 sec/batch)
2017-10-21 12:36:52.268803: step 4140, loss = 0.94 (10847.5 examples/sec; 0.047 sec/batch)
2017-10-21 12:36:53.162757: step 4150, loss = 0.97 (12077.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:36:54.056820: step 4160, loss = 0.98 (13428.8 examples/sec; 0.038 sec/batch)
2017-10-21 12:36:54.938969: step 4170, loss = 1.00 (12067.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:36:55.807378: step 4180, loss = 0.94 (11623.3 examples/sec; 0.044 sec/batch)
2017-10-21 12:36:56.705351: step 4190, loss = 0.87 (11047.7 examples/sec; 0.046 sec/batch)
2017-10-21 12:36:57.583414: step 4200, loss = 0.90 (12623.0 examples/sec; 0.041 sec/batch)
2017-10-21 12:36:58.865181: step 4210, loss = 0.87 (18148.4 examples/sec; 0.028 sec/batch)
2017-10-21 12:36:59.461789: step 4220, loss = 0.92 (16552.5 examples/sec; 0.031 sec/batch)
2017-10-21 12:37:00.349960: step 4230, loss = 0.88 (11464.6 examples/sec; 0.045 sec/batch)
2017-10-21 12:37:01.269322: step 4240, loss = 0.85 (10669.0 examples/sec; 0.048 sec/batch)
2017-10-21 12:37:02.151286: step 4250, loss = 0.88 (11629.8 examples/sec; 0.044 sec/batch)
2017-10-21 12:37:03.027533: step 4260, loss = 0.97 (12482.4 examples/sec; 0.041 sec/batch)
2017-10-21 12:37:03.907963: step 4270, loss = 0.90 (11553.7 examples/sec; 0.044 sec/batch)
2017-10-21 12:37:04.777372: step 4280, loss = 0.88 (11683.2 examples/sec; 0.044 sec/batch)
2017-10-21 12:37:05.674638: step 4290, loss = 0.86 (10276.2 examples/sec; 0.050 sec/batch)
2017-10-21 12:37:06.557233: step 4300, loss = 0.95 (12143.5 examples/sec; 0.042 sec/batch)
2017-10-21 12:37:07.829643: step 4310, loss = 0.97 (18155.8 examples/sec; 0.028 sec/batch)
2017-10-21 12:37:08.455241: step 4320, loss = 0.91 (18055.0 examples/sec; 0.028 sec/batch)
2017-10-21 12:37:09.335396: step 4330, loss = 1.13 (12335.8 examples/sec; 0.042 sec/batch)
2017-10-21 12:37:10.206820: step 4340, loss = 0.91 (12610.0 examples/sec; 0.041 sec/batch)
2017-10-21 12:37:11.092212: step 4350, loss = 0.96 (12025.9 examples/sec; 0.043 sec/batch)
2017-10-21 12:37:11.975164: step 4360, loss = 0.98 (11600.2 examples/sec; 0.044 sec/batch)
2017-10-21 12:37:12.852154: step 4370, loss = 0.94 (11618.0 examples/sec; 0.044 sec/batch)
2017-10-21 12:37:13.727804: step 4380, loss = 0.89 (11649.7 examples/sec; 0.044 sec/batch)
2017-10-21 12:37:14.614374: step 4390, loss = 0.93 (11913.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:37:15.482193: step 4400, loss = 0.83 (13182.1 examples/sec; 0.039 sec/batch)
2017-10-21 12:37:16.764195: step 4410, loss = 0.87 (18120.2 examples/sec; 0.028 sec/batch)
2017-10-21 12:37:17.367528: step 4420, loss = 0.81 (13103.4 examples/sec; 0.039 sec/batch)
2017-10-21 12:37:18.253642: step 4430, loss = 0.78 (11457.1 examples/sec; 0.045 sec/batch)
2017-10-21 12:37:19.154860: step 4440, loss = 0.95 (9883.4 examples/sec; 0.052 sec/batch)
2017-10-21 12:37:20.023535: step 4450, loss = 0.82 (11373.6 examples/sec; 0.045 sec/batch)
2017-10-21 12:37:20.927301: step 4460, loss = 0.93 (11693.5 examples/sec; 0.044 sec/batch)
2017-10-21 12:37:21.804154: step 4470, loss = 0.93 (10788.9 examples/sec; 0.047 sec/batch)
2017-10-21 12:37:22.691103: step 4480, loss = 0.87 (11712.0 examples/sec; 0.044 sec/batch)
2017-10-21 12:37:23.562887: step 4490, loss = 0.83 (12172.9 examples/sec; 0.042 sec/batch)
2017-10-21 12:37:24.447522: step 4500, loss = 0.87 (11598.4 examples/sec; 0.044 sec/batch)
2017-10-21 12:37:25.723170: step 4510, loss = 0.86 (17670.8 examples/sec; 0.029 sec/batch)
2017-10-21 12:37:26.329305: step 4520, loss = 0.95 (17439.6 examples/sec; 0.029 sec/batch)
2017-10-21 12:37:27.217124: step 4530, loss = 0.88 (10962.6 examples/sec; 0.047 sec/batch)
2017-10-21 12:37:28.095597: step 4540, loss = 1.02 (11589.7 examples/sec; 0.044 sec/batch)
2017-10-21 12:37:29.005500: step 4550, loss = 0.82 (10627.6 examples/sec; 0.048 sec/batch)
2017-10-21 12:37:29.871741: step 4560, loss = 0.82 (13543.3 examples/sec; 0.038 sec/batch)
2017-10-21 12:37:30.748170: step 4570, loss = 0.86 (12026.4 examples/sec; 0.043 sec/batch)
2017-10-21 12:37:31.637754: step 4580, loss = 0.76 (11331.7 examples/sec; 0.045 sec/batch)
2017-10-21 12:37:32.525126: step 4590, loss = 1.02 (12159.4 examples/sec; 0.042 sec/batch)
2017-10-21 12:37:33.413979: step 4600, loss = 0.89 (10266.2 examples/sec; 0.050 sec/batch)
2017-10-21 12:37:34.671424: step 4610, loss = 0.94 (17147.6 examples/sec; 0.030 sec/batch)
2017-10-21 12:37:35.277289: step 4620, loss = 0.87 (12669.1 examples/sec; 0.040 sec/batch)
2017-10-21 12:37:36.161875: step 4630, loss = 0.88 (11274.9 examples/sec; 0.045 sec/batch)
2017-10-21 12:37:37.021390: step 4640, loss = 0.93 (11153.5 examples/sec; 0.046 sec/batch)
2017-10-21 12:37:37.901223: step 4650, loss = 0.92 (11327.9 examples/sec; 0.045 sec/batch)
2017-10-21 12:37:38.788003: step 4660, loss = 0.80 (11628.3 examples/sec; 0.044 sec/batch)
2017-10-21 12:37:39.668937: step 4670, loss = 0.83 (11638.7 examples/sec; 0.044 sec/batch)
2017-10-21 12:37:40.545210: step 4680, loss = 0.79 (12118.9 examples/sec; 0.042 sec/batch)
2017-10-21 12:37:41.453611: step 4690, loss = 0.87 (11062.3 examples/sec; 0.046 sec/batch)
2017-10-21 12:37:42.340349: step 4700, loss = 0.79 (12059.8 examples/sec; 0.042 sec/batch)
2017-10-21 12:37:43.608644: step 4710, loss = 0.81 (17597.5 examples/sec; 0.029 sec/batch)
2017-10-21 12:37:44.218510: step 4720, loss = 0.89 (17500.4 examples/sec; 0.029 sec/batch)
2017-10-21 12:37:45.094533: step 4730, loss = 0.87 (11546.4 examples/sec; 0.044 sec/batch)
2017-10-21 12:37:45.979240: step 4740, loss = 0.93 (11644.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:37:46.870113: step 4750, loss = 0.89 (10677.2 examples/sec; 0.048 sec/batch)
2017-10-21 12:37:47.755738: step 4760, loss = 1.04 (12002.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:37:48.622448: step 4770, loss = 0.84 (11609.2 examples/sec; 0.044 sec/batch)
2017-10-21 12:37:49.511385: step 4780, loss = 0.79 (11057.5 examples/sec; 0.046 sec/batch)
2017-10-21 12:37:50.401712: step 4790, loss = 0.81 (11403.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:37:51.286104: step 4800, loss = 0.85 (12193.4 examples/sec; 0.042 sec/batch)
2017-10-21 12:37:52.556709: step 4810, loss = 0.80 (18923.5 examples/sec; 0.027 sec/batch)
2017-10-21 12:37:53.150987: step 4820, loss = 0.85 (16457.5 examples/sec; 0.031 sec/batch)
2017-10-21 12:37:54.039345: step 4830, loss = 0.76 (10979.3 examples/sec; 0.047 sec/batch)
2017-10-21 12:37:54.944147: step 4840, loss = 0.97 (10524.8 examples/sec; 0.049 sec/batch)
2017-10-21 12:37:55.811850: step 4850, loss = 0.86 (14340.4 examples/sec; 0.036 sec/batch)
2017-10-21 12:37:56.732446: step 4860, loss = 0.98 (10863.5 examples/sec; 0.047 sec/batch)
2017-10-21 12:37:57.607919: step 4870, loss = 0.80 (11304.8 examples/sec; 0.045 sec/batch)
2017-10-21 12:37:58.484965: step 4880, loss = 0.92 (11107.9 examples/sec; 0.046 sec/batch)
2017-10-21 12:37:59.358418: step 4890, loss = 0.96 (12801.8 examples/sec; 0.040 sec/batch)
2017-10-21 12:38:00.240227: step 4900, loss = 0.85 (11957.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:38:01.528894: step 4910, loss = 0.79 (17809.5 examples/sec; 0.029 sec/batch)
2017-10-21 12:38:02.154410: step 4920, loss = 0.92 (17939.3 examples/sec; 0.029 sec/batch)
2017-10-21 12:38:03.036938: step 4930, loss = 1.00 (10858.1 examples/sec; 0.047 sec/batch)
2017-10-21 12:38:03.906134: step 4940, loss = 0.80 (12049.1 examples/sec; 0.042 sec/batch)
2017-10-21 12:38:04.779518: step 4950, loss = 0.77 (11660.3 examples/sec; 0.044 sec/batch)
2017-10-21 12:38:05.672505: step 4960, loss = 0.85 (11114.2 examples/sec; 0.046 sec/batch)
2017-10-21 12:38:06.582678: step 4970, loss = 0.92 (11532.5 examples/sec; 0.044 sec/batch)
2017-10-21 12:38:07.447817: step 4980, loss = 0.90 (12895.8 examples/sec; 0.040 sec/batch)
2017-10-21 12:38:08.352250: step 4990, loss = 0.80 (12721.4 examples/sec; 0.040 sec/batch)
2017-10-21 12:38:09.217270: step 5000, loss = 0.90 (11566.0 examples/sec; 0.044 sec/batch)
2017-10-21 12:38:10.600078: step 5010, loss = 0.74 (18956.6 examples/sec; 0.027 sec/batch)
2017-10-21 12:38:11.204561: step 5020, loss = 0.83 (14653.1 examples/sec; 0.035 sec/batch)
2017-10-21 12:38:12.075403: step 5030, loss = 0.86 (11687.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:38:12.948637: step 5040, loss = 0.86 (12382.9 examples/sec; 0.041 sec/batch)
2017-10-21 12:38:13.815093: step 5050, loss = 0.88 (11800.4 examples/sec; 0.043 sec/batch)
2017-10-21 12:38:14.712504: step 5060, loss = 0.90 (11632.2 examples/sec; 0.044 sec/batch)
2017-10-21 12:38:15.600452: step 5070, loss = 0.79 (11642.4 examples/sec; 0.044 sec/batch)
2017-10-21 12:38:16.477170: step 5080, loss = 0.84 (10916.8 examples/sec; 0.047 sec/batch)
2017-10-21 12:38:17.347746: step 5090, loss = 0.91 (11703.2 examples/sec; 0.044 sec/batch)
2017-10-21 12:38:18.212607: step 5100, loss = 0.81 (11826.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:38:19.503128: step 5110, loss = 0.78 (16250.1 examples/sec; 0.032 sec/batch)
2017-10-21 12:38:20.104966: step 5120, loss = 0.81 (17502.5 examples/sec; 0.029 sec/batch)
2017-10-21 12:38:20.990071: step 5130, loss = 0.80 (12617.3 examples/sec; 0.041 sec/batch)
2017-10-21 12:38:21.860282: step 5140, loss = 0.92 (11566.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:38:22.735724: step 5150, loss = 0.83 (12354.7 examples/sec; 0.041 sec/batch)
2017-10-21 12:38:23.621751: step 5160, loss = 0.90 (11331.8 examples/sec; 0.045 sec/batch)
2017-10-21 12:38:24.487693: step 5170, loss = 0.75 (12107.1 examples/sec; 0.042 sec/batch)
2017-10-21 12:38:25.367089: step 5180, loss = 0.86 (11862.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:38:26.240068: step 5190, loss = 0.82 (11401.6 examples/sec; 0.045 sec/batch)
2017-10-21 12:38:27.140859: step 5200, loss = 0.85 (11492.3 examples/sec; 0.045 sec/batch)
2017-10-21 12:38:28.428048: step 5210, loss = 0.85 (17552.5 examples/sec; 0.029 sec/batch)
2017-10-21 12:38:29.025738: step 5220, loss = 0.74 (17024.2 examples/sec; 0.030 sec/batch)
2017-10-21 12:38:29.910034: step 5230, loss = 0.78 (11315.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:38:30.778852: step 5240, loss = 0.96 (11554.1 examples/sec; 0.044 sec/batch)
2017-10-21 12:38:31.648337: step 5250, loss = 0.80 (11893.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:38:32.516868: step 5260, loss = 0.86 (10778.4 examples/sec; 0.048 sec/batch)
2017-10-21 12:38:33.387450: step 5270, loss = 0.90 (11346.7 examples/sec; 0.045 sec/batch)
2017-10-21 12:38:34.269817: step 5280, loss = 0.93 (11666.2 examples/sec; 0.044 sec/batch)
2017-10-21 12:38:35.138105: step 5290, loss = 0.78 (12039.7 examples/sec; 0.043 sec/batch)
2017-10-21 12:38:36.013427: step 5300, loss = 0.77 (11818.5 examples/sec; 0.043 sec/batch)
2017-10-21 12:38:37.269548: step 5310, loss = 0.80 (18013.6 examples/sec; 0.028 sec/batch)
2017-10-21 12:38:37.883436: step 5320, loss = 0.85 (16977.7 examples/sec; 0.030 sec/batch)
2017-10-21 12:38:38.745239: step 5330, loss = 0.75 (11377.3 examples/sec; 0.045 sec/batch)
2017-10-21 12:38:39.626812: step 5340, loss = 0.79 (11542.8 examples/sec; 0.044 sec/batch)
2017-10-21 12:38:40.518346: step 5350, loss = 0.82 (11950.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:38:41.390444: step 5360, loss = 0.79 (11377.3 examples/sec; 0.045 sec/batch)
2017-10-21 12:38:42.262004: step 5370, loss = 0.80 (11802.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:38:43.149659: step 5380, loss = 0.82 (10558.5 examples/sec; 0.048 sec/batch)
2017-10-21 12:38:44.040893: step 5390, loss = 0.87 (11855.8 examples/sec; 0.043 sec/batch)
2017-10-21 12:38:44.918157: step 5400, loss = 0.79 (12025.9 examples/sec; 0.043 sec/batch)
2017-10-21 12:38:46.194311: step 5410, loss = 0.81 (17240.1 examples/sec; 0.030 sec/batch)
2017-10-21 12:38:46.794249: step 5420, loss = 0.74 (17760.7 examples/sec; 0.029 sec/batch)
2017-10-21 12:38:47.674765: step 5430, loss = 0.83 (11181.7 examples/sec; 0.046 sec/batch)
2017-10-21 12:38:48.549183: step 5440, loss = 0.82 (12029.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:38:49.417508: step 5450, loss = 0.79 (10992.6 examples/sec; 0.047 sec/batch)
2017-10-21 12:38:50.288596: step 5460, loss = 0.90 (11777.6 examples/sec; 0.043 sec/batch)
2017-10-21 12:38:51.172011: step 5470, loss = 0.81 (10443.5 examples/sec; 0.049 sec/batch)
2017-10-21 12:38:52.065801: step 5480, loss = 0.76 (12087.2 examples/sec; 0.042 sec/batch)
2017-10-21 12:38:52.928331: step 5490, loss = 0.86 (12565.4 examples/sec; 0.041 sec/batch)
2017-10-21 12:38:53.819134: step 5500, loss = 0.76 (11329.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:38:55.114817: step 5510, loss = 0.81 (16953.8 examples/sec; 0.030 sec/batch)
2017-10-21 12:38:55.715897: step 5520, loss = 0.79 (16966.1 examples/sec; 0.030 sec/batch)
2017-10-21 12:38:56.576897: step 5530, loss = 0.80 (11499.1 examples/sec; 0.045 sec/batch)
2017-10-21 12:38:57.456879: step 5540, loss = 0.73 (11764.0 examples/sec; 0.044 sec/batch)
2017-10-21 12:38:58.326520: step 5550, loss = 0.85 (11849.5 examples/sec; 0.043 sec/batch)
2017-10-21 12:38:59.209284: step 5560, loss = 0.95 (11758.0 examples/sec; 0.044 sec/batch)
2017-10-21 12:39:00.082173: step 5570, loss = 0.88 (11658.1 examples/sec; 0.044 sec/batch)
2017-10-21 12:39:01.001954: step 5580, loss = 0.74 (12018.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:39:01.893979: step 5590, loss = 0.95 (11170.9 examples/sec; 0.046 sec/batch)
2017-10-21 12:39:02.781100: step 5600, loss = 0.85 (11838.6 examples/sec; 0.043 sec/batch)
2017-10-21 12:39:04.095806: step 5610, loss = 0.77 (17078.6 examples/sec; 0.030 sec/batch)
2017-10-21 12:39:04.742635: step 5620, loss = 0.81 (10734.5 examples/sec; 0.048 sec/batch)
2017-10-21 12:39:05.653436: step 5630, loss = 0.84 (12078.6 examples/sec; 0.042 sec/batch)
2017-10-21 12:39:06.597640: step 5640, loss = 0.75 (11519.4 examples/sec; 0.044 sec/batch)
2017-10-21 12:39:07.497663: step 5650, loss = 0.88 (11375.8 examples/sec; 0.045 sec/batch)
2017-10-21 12:39:08.384051: step 5660, loss = 0.85 (11422.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:39:09.266804: step 5670, loss = 0.92 (11740.8 examples/sec; 0.044 sec/batch)
2017-10-21 12:39:10.146097: step 5680, loss = 0.89 (11978.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:39:11.024620: step 5690, loss = 0.78 (12199.9 examples/sec; 0.042 sec/batch)
2017-10-21 12:39:11.903092: step 5700, loss = 0.77 (11857.8 examples/sec; 0.043 sec/batch)
2017-10-21 12:39:13.190842: step 5710, loss = 0.82 (17783.2 examples/sec; 0.029 sec/batch)
2017-10-21 12:39:13.786097: step 5720, loss = 0.84 (16605.5 examples/sec; 0.031 sec/batch)
2017-10-21 12:39:14.651778: step 5730, loss = 0.77 (12424.6 examples/sec; 0.041 sec/batch)
2017-10-21 12:39:15.535715: step 5740, loss = 0.88 (11909.9 examples/sec; 0.043 sec/batch)
2017-10-21 12:39:16.409147: step 5750, loss = 0.76 (11263.8 examples/sec; 0.045 sec/batch)
2017-10-21 12:39:17.293590: step 5760, loss = 0.89 (13032.4 examples/sec; 0.039 sec/batch)
2017-10-21 12:39:18.173938: step 5770, loss = 0.76 (12047.9 examples/sec; 0.042 sec/batch)
2017-10-21 12:39:19.060621: step 5780, loss = 0.85 (11647.8 examples/sec; 0.044 sec/batch)
2017-10-21 12:39:19.928796: step 5790, loss = 0.84 (11892.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:39:20.818417: step 5800, loss = 0.76 (11790.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:39:22.072275: step 5810, loss = 0.77 (17812.2 examples/sec; 0.029 sec/batch)
2017-10-21 12:39:22.708865: step 5820, loss = 0.74 (11113.4 examples/sec; 0.046 sec/batch)
2017-10-21 12:39:23.583680: step 5830, loss = 0.75 (12324.1 examples/sec; 0.042 sec/batch)
2017-10-21 12:39:24.479355: step 5840, loss = 0.86 (11629.7 examples/sec; 0.044 sec/batch)
2017-10-21 12:39:25.359551: step 5850, loss = 0.73 (11930.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:39:26.268733: step 5860, loss = 0.78 (11448.8 examples/sec; 0.045 sec/batch)
2017-10-21 12:39:27.144921: step 5870, loss = 0.80 (11910.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:39:28.024290: step 5880, loss = 0.81 (11355.1 examples/sec; 0.045 sec/batch)
2017-10-21 12:39:28.903681: step 5890, loss = 0.80 (11251.2 examples/sec; 0.046 sec/batch)
2017-10-21 12:39:29.769983: step 5900, loss = 0.73 (12310.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:39:31.057510: step 5910, loss = 0.75 (16894.4 examples/sec; 0.030 sec/batch)
2017-10-21 12:39:31.648653: step 5920, loss = 0.81 (17372.8 examples/sec; 0.029 sec/batch)
2017-10-21 12:39:32.529219: step 5930, loss = 0.70 (11521.1 examples/sec; 0.044 sec/batch)
2017-10-21 12:39:33.395905: step 5940, loss = 0.82 (12856.6 examples/sec; 0.040 sec/batch)
2017-10-21 12:39:34.290905: step 5950, loss = 0.75 (11150.4 examples/sec; 0.046 sec/batch)
2017-10-21 12:39:35.179512: step 5960, loss = 0.73 (11629.0 examples/sec; 0.044 sec/batch)
2017-10-21 12:39:36.050069: step 5970, loss = 0.85 (11314.7 examples/sec; 0.045 sec/batch)
2017-10-21 12:39:36.923186: step 5980, loss = 0.74 (12010.6 examples/sec; 0.043 sec/batch)
2017-10-21 12:39:37.801273: step 5990, loss = 0.72 (11341.7 examples/sec; 0.045 sec/batch)
2017-10-21 12:39:38.683810: step 6000, loss = 0.97 (11408.9 examples/sec; 0.045 sec/batch)
2017-10-21 12:39:40.112255: step 6010, loss = 0.72 (18095.2 examples/sec; 0.028 sec/batch)
2017-10-21 12:39:40.732861: step 6020, loss = 0.72 (13521.9 examples/sec; 0.038 sec/batch)
2017-10-21 12:39:41.617004: step 6030, loss = 0.79 (11168.8 examples/sec; 0.046 sec/batch)
2017-10-21 12:39:42.491095: step 6040, loss = 0.76 (11415.6 examples/sec; 0.045 sec/batch)
2017-10-21 12:39:43.371363: step 6050, loss = 0.81 (11369.3 examples/sec; 0.045 sec/batch)
2017-10-21 12:39:44.252480: step 6060, loss = 0.90 (10996.2 examples/sec; 0.047 sec/batch)
2017-10-21 12:39:45.118621: step 6070, loss = 0.75 (12239.1 examples/sec; 0.042 sec/batch)
2017-10-21 12:39:45.977235: step 6080, loss = 0.84 (12043.7 examples/sec; 0.043 sec/batch)
2017-10-21 12:39:46.857804: step 6090, loss = 0.77 (11255.5 examples/sec; 0.045 sec/batch)
2017-10-21 12:39:47.732857: step 6100, loss = 0.84 (11440.9 examples/sec; 0.045 sec/batch)
2017-10-21 12:39:49.001041: step 6110, loss = 0.72 (17773.1 examples/sec; 0.029 sec/batch)
2017-10-21 12:39:49.600679: step 6120, loss = 0.76 (15364.8 examples/sec; 0.033 sec/batch)
2017-10-21 12:39:50.470658: step 6130, loss = 0.73 (11606.2 examples/sec; 0.044 sec/batch)
2017-10-21 12:39:51.362167: step 6140, loss = 0.84 (11320.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:39:52.237964: step 6150, loss = 0.74 (11517.9 examples/sec; 0.044 sec/batch)
2017-10-21 12:39:53.106023: step 6160, loss = 0.76 (12123.4 examples/sec; 0.042 sec/batch)
2017-10-21 12:39:53.968868: step 6170, loss = 0.83 (12929.8 examples/sec; 0.040 sec/batch)
2017-10-21 12:39:54.846420: step 6180, loss = 0.81 (11461.9 examples/sec; 0.045 sec/batch)
2017-10-21 12:39:55.721943: step 6190, loss = 0.74 (11641.4 examples/sec; 0.044 sec/batch)
2017-10-21 12:39:56.604646: step 6200, loss = 0.80 (12473.1 examples/sec; 0.041 sec/batch)
2017-10-21 12:39:57.883830: step 6210, loss = 0.64 (17240.2 examples/sec; 0.030 sec/batch)
2017-10-21 12:39:58.485403: step 6220, loss = 0.83 (17021.3 examples/sec; 0.030 sec/batch)
2017-10-21 12:39:59.349581: step 6230, loss = 0.80 (11914.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:40:00.223347: step 6240, loss = 0.74 (12700.3 examples/sec; 0.040 sec/batch)
2017-10-21 12:40:01.141136: step 6250, loss = 0.74 (11879.5 examples/sec; 0.043 sec/batch)
2017-10-21 12:40:02.023885: step 6260, loss = 0.76 (11183.4 examples/sec; 0.046 sec/batch)
2017-10-21 12:40:02.896176: step 6270, loss = 0.78 (12338.1 examples/sec; 0.041 sec/batch)
2017-10-21 12:40:03.766635: step 6280, loss = 0.77 (11996.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:40:04.655674: step 6290, loss = 0.77 (11183.7 examples/sec; 0.046 sec/batch)
2017-10-21 12:40:05.535437: step 6300, loss = 0.76 (11501.2 examples/sec; 0.045 sec/batch)
2017-10-21 12:40:06.813894: step 6310, loss = 0.78 (18029.6 examples/sec; 0.028 sec/batch)
2017-10-21 12:40:07.410388: step 6320, loss = 0.83 (16942.9 examples/sec; 0.030 sec/batch)
2017-10-21 12:40:08.271271: step 6330, loss = 0.76 (11722.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:40:09.134250: step 6340, loss = 0.81 (12108.5 examples/sec; 0.042 sec/batch)
2017-10-21 12:40:10.019697: step 6350, loss = 0.78 (12366.0 examples/sec; 0.041 sec/batch)
2017-10-21 12:40:10.896542: step 6360, loss = 0.80 (11569.7 examples/sec; 0.044 sec/batch)
2017-10-21 12:40:11.757699: step 6370, loss = 0.79 (11938.6 examples/sec; 0.043 sec/batch)
2017-10-21 12:40:12.629901: step 6380, loss = 0.69 (12966.6 examples/sec; 0.039 sec/batch)
2017-10-21 12:40:13.512415: step 6390, loss = 0.73 (11272.8 examples/sec; 0.045 sec/batch)
2017-10-21 12:40:14.381987: step 6400, loss = 0.72 (11554.9 examples/sec; 0.044 sec/batch)
2017-10-21 12:40:15.656046: step 6410, loss = 0.73 (16640.9 examples/sec; 0.031 sec/batch)
2017-10-21 12:40:16.248335: step 6420, loss = 0.80 (17750.0 examples/sec; 0.029 sec/batch)
2017-10-21 12:40:17.138291: step 6430, loss = 0.87 (11337.5 examples/sec; 0.045 sec/batch)
2017-10-21 12:40:18.035542: step 6440, loss = 0.80 (11962.6 examples/sec; 0.043 sec/batch)
2017-10-21 12:40:18.904577: step 6450, loss = 0.68 (11481.6 examples/sec; 0.045 sec/batch)
2017-10-21 12:40:19.793886: step 6460, loss = 0.86 (12062.3 examples/sec; 0.042 sec/batch)
2017-10-21 12:40:20.669297: step 6470, loss = 0.85 (11362.3 examples/sec; 0.045 sec/batch)
2017-10-21 12:40:21.549218: step 6480, loss = 0.76 (10694.8 examples/sec; 0.048 sec/batch)
2017-10-21 12:40:22.418858: step 6490, loss = 0.76 (13013.7 examples/sec; 0.039 sec/batch)
2017-10-21 12:40:23.306217: step 6500, loss = 0.79 (11480.2 examples/sec; 0.045 sec/batch)
2017-10-21 12:40:24.605370: step 6510, loss = 0.78 (17723.6 examples/sec; 0.029 sec/batch)
2017-10-21 12:40:25.210446: step 6520, loss = 0.80 (17019.9 examples/sec; 0.030 sec/batch)
2017-10-21 12:40:26.073627: step 6530, loss = 0.74 (11965.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:40:26.955098: step 6540, loss = 0.81 (11096.1 examples/sec; 0.046 sec/batch)
2017-10-21 12:40:27.830245: step 6550, loss = 0.69 (10833.3 examples/sec; 0.047 sec/batch)
2017-10-21 12:40:28.713689: step 6560, loss = 0.82 (12165.9 examples/sec; 0.042 sec/batch)
2017-10-21 12:40:29.595201: step 6570, loss = 0.75 (11929.7 examples/sec; 0.043 sec/batch)
2017-10-21 12:40:30.471789: step 6580, loss = 0.83 (11657.9 examples/sec; 0.044 sec/batch)
2017-10-21 12:40:31.344296: step 6590, loss = 0.82 (12614.2 examples/sec; 0.041 sec/batch)
2017-10-21 12:40:32.249921: step 6600, loss = 0.77 (10734.0 examples/sec; 0.048 sec/batch)
2017-10-21 12:40:33.521787: step 6610, loss = 0.74 (16666.4 examples/sec; 0.031 sec/batch)
2017-10-21 12:40:34.117209: step 6620, loss = 0.68 (17970.0 examples/sec; 0.028 sec/batch)
2017-10-21 12:40:34.979211: step 6630, loss = 0.69 (11554.1 examples/sec; 0.044 sec/batch)
2017-10-21 12:40:35.854278: step 6640, loss = 0.76 (10457.0 examples/sec; 0.049 sec/batch)
2017-10-21 12:40:36.729680: step 6650, loss = 0.83 (11009.1 examples/sec; 0.047 sec/batch)
2017-10-21 12:40:37.599825: step 6660, loss = 0.76 (11944.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:40:38.464260: step 6670, loss = 0.82 (11979.7 examples/sec; 0.043 sec/batch)
2017-10-21 12:40:39.352403: step 6680, loss = 0.77 (10580.3 examples/sec; 0.048 sec/batch)
2017-10-21 12:40:40.227217: step 6690, loss = 0.89 (11479.1 examples/sec; 0.045 sec/batch)
2017-10-21 12:40:41.109459: step 6700, loss = 0.72 (12363.0 examples/sec; 0.041 sec/batch)
2017-10-21 12:40:42.476502: step 6710, loss = 0.72 (14776.2 examples/sec; 0.035 sec/batch)
2017-10-21 12:40:43.092183: step 6720, loss = 0.67 (17536.2 examples/sec; 0.029 sec/batch)
2017-10-21 12:40:43.899054: step 6730, loss = 0.78 (11000.6 examples/sec; 0.047 sec/batch)
2017-10-21 12:40:44.784015: step 6740, loss = 0.82 (11749.3 examples/sec; 0.044 sec/batch)
2017-10-21 12:40:45.667790: step 6750, loss = 0.75 (12020.9 examples/sec; 0.043 sec/batch)
2017-10-21 12:40:46.541704: step 6760, loss = 0.85 (12335.3 examples/sec; 0.042 sec/batch)
2017-10-21 12:40:47.408187: step 6770, loss = 0.69 (11523.8 examples/sec; 0.044 sec/batch)
2017-10-21 12:40:48.285713: step 6780, loss = 0.73 (11600.0 examples/sec; 0.044 sec/batch)
2017-10-21 12:40:49.149009: step 6790, loss = 0.85 (11533.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:40:50.035898: step 6800, loss = 0.75 (11697.2 examples/sec; 0.044 sec/batch)
2017-10-21 12:40:51.312592: step 6810, loss = 0.73 (17628.8 examples/sec; 0.029 sec/batch)
2017-10-21 12:40:51.915214: step 6820, loss = 0.75 (18189.4 examples/sec; 0.028 sec/batch)
2017-10-21 12:40:52.794897: step 6830, loss = 0.71 (11806.4 examples/sec; 0.043 sec/batch)
2017-10-21 12:40:53.664323: step 6840, loss = 0.77 (11971.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:40:54.541299: step 6850, loss = 0.76 (11373.5 examples/sec; 0.045 sec/batch)
2017-10-21 12:40:55.417336: step 6860, loss = 0.71 (12196.3 examples/sec; 0.042 sec/batch)
2017-10-21 12:40:56.308893: step 6870, loss = 0.75 (10519.7 examples/sec; 0.049 sec/batch)
2017-10-21 12:40:57.197023: step 6880, loss = 0.78 (12189.8 examples/sec; 0.042 sec/batch)
2017-10-21 12:40:58.084808: step 6890, loss = 0.68 (11483.5 examples/sec; 0.045 sec/batch)
2017-10-21 12:40:58.957571: step 6900, loss = 0.79 (11472.1 examples/sec; 0.045 sec/batch)
2017-10-21 12:41:00.233493: step 6910, loss = 0.87 (17634.1 examples/sec; 0.029 sec/batch)
2017-10-21 12:41:00.856788: step 6920, loss = 0.79 (13765.0 examples/sec; 0.037 sec/batch)
2017-10-21 12:41:01.737118: step 6930, loss = 0.73 (10334.1 examples/sec; 0.050 sec/batch)
2017-10-21 12:41:02.623158: step 6940, loss = 0.74 (11543.7 examples/sec; 0.044 sec/batch)
2017-10-21 12:41:03.520981: step 6950, loss = 0.71 (11558.9 examples/sec; 0.044 sec/batch)
2017-10-21 12:41:04.394702: step 6960, loss = 0.71 (12302.5 examples/sec; 0.042 sec/batch)
2017-10-21 12:41:05.268834: step 6970, loss = 0.86 (11567.8 examples/sec; 0.044 sec/batch)
2017-10-21 12:41:06.140416: step 6980, loss = 0.71 (11759.9 examples/sec; 0.044 sec/batch)
2017-10-21 12:41:07.015811: step 6990, loss = 0.78 (12586.7 examples/sec; 0.041 sec/batch)
2017-10-21 12:41:07.889859: step 7000, loss = 0.67 (11120.5 examples/sec; 0.046 sec/batch)
2017-10-21 12:41:09.269628: step 7010, loss = 0.91 (17779.3 examples/sec; 0.029 sec/batch)
2017-10-21 12:41:09.856991: step 7020, loss = 0.61 (17248.5 examples/sec; 0.030 sec/batch)
2017-10-21 12:41:10.708020: step 7030, loss = 0.72 (11840.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:41:11.586212: step 7040, loss = 0.78 (11638.8 examples/sec; 0.044 sec/batch)
2017-10-21 12:41:12.470230: step 7050, loss = 0.70 (11377.2 examples/sec; 0.045 sec/batch)
2017-10-21 12:41:13.354320: step 7060, loss = 0.73 (12499.4 examples/sec; 0.041 sec/batch)
2017-10-21 12:41:14.228739: step 7070, loss = 0.76 (11838.4 examples/sec; 0.043 sec/batch)
2017-10-21 12:41:15.104416: step 7080, loss = 0.84 (11850.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:41:15.968749: step 7090, loss = 0.74 (11222.1 examples/sec; 0.046 sec/batch)
2017-10-21 12:41:16.852924: step 7100, loss = 0.64 (11718.9 examples/sec; 0.044 sec/batch)
2017-10-21 12:41:18.144439: step 7110, loss = 0.72 (17115.0 examples/sec; 0.030 sec/batch)
2017-10-21 12:41:18.756856: step 7120, loss = 0.73 (16355.7 examples/sec; 0.031 sec/batch)
2017-10-21 12:41:19.583491: step 7130, loss = 0.72 (11552.5 examples/sec; 0.044 sec/batch)
2017-10-21 12:41:20.472216: step 7140, loss = 0.68 (11347.2 examples/sec; 0.045 sec/batch)
2017-10-21 12:41:21.339532: step 7150, loss = 0.60 (11904.7 examples/sec; 0.043 sec/batch)
2017-10-21 12:41:22.243157: step 7160, loss = 0.67 (11834.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:41:23.132160: step 7170, loss = 0.74 (11852.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:41:24.018719: step 7180, loss = 0.75 (11056.6 examples/sec; 0.046 sec/batch)
2017-10-21 12:41:24.887525: step 7190, loss = 0.74 (12119.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:41:25.766211: step 7200, loss = 0.68 (10880.0 examples/sec; 0.047 sec/batch)
2017-10-21 12:41:27.054319: step 7210, loss = 0.68 (18211.3 examples/sec; 0.028 sec/batch)
2017-10-21 12:41:27.654327: step 7220, loss = 0.71 (17349.1 examples/sec; 0.030 sec/batch)
2017-10-21 12:41:28.529416: step 7230, loss = 0.83 (11048.6 examples/sec; 0.046 sec/batch)
2017-10-21 12:41:29.400057: step 7240, loss = 0.81 (11699.7 examples/sec; 0.044 sec/batch)
2017-10-21 12:41:30.268241: step 7250, loss = 0.66 (11556.3 examples/sec; 0.044 sec/batch)
2017-10-21 12:41:31.146127: step 7260, loss = 0.81 (12481.7 examples/sec; 0.041 sec/batch)
2017-10-21 12:41:32.025369: step 7270, loss = 0.75 (12458.5 examples/sec; 0.041 sec/batch)
2017-10-21 12:41:32.893709: step 7280, loss = 0.81 (11267.3 examples/sec; 0.045 sec/batch)
2017-10-21 12:41:33.750114: step 7290, loss = 0.66 (11479.1 examples/sec; 0.045 sec/batch)
2017-10-21 12:41:34.638323: step 7300, loss = 0.71 (11237.1 examples/sec; 0.046 sec/batch)
2017-10-21 12:41:35.923139: step 7310, loss = 0.72 (16985.2 examples/sec; 0.030 sec/batch)
2017-10-21 12:41:36.530792: step 7320, loss = 0.70 (15797.9 examples/sec; 0.032 sec/batch)
2017-10-21 12:41:37.397076: step 7330, loss = 0.68 (11022.0 examples/sec; 0.046 sec/batch)
2017-10-21 12:41:38.284814: step 7340, loss = 0.77 (11376.6 examples/sec; 0.045 sec/batch)
2017-10-21 12:41:39.146511: step 7350, loss = 0.71 (11818.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:41:40.011170: step 7360, loss = 0.83 (12444.1 examples/sec; 0.041 sec/batch)
2017-10-21 12:41:40.899676: step 7370, loss = 0.65 (10402.5 examples/sec; 0.049 sec/batch)
2017-10-21 12:41:41.774055: step 7380, loss = 0.75 (11545.1 examples/sec; 0.044 sec/batch)
2017-10-21 12:41:42.655039: step 7390, loss = 0.72 (11331.4 examples/sec; 0.045 sec/batch)
2017-10-21 12:41:43.527041: step 7400, loss = 0.64 (11274.5 examples/sec; 0.045 sec/batch)
2017-10-21 12:41:44.867167: step 7410, loss = 0.72 (16689.9 examples/sec; 0.031 sec/batch)
2017-10-21 12:41:45.459806: step 7420, loss = 0.69 (16258.8 examples/sec; 0.031 sec/batch)
2017-10-21 12:41:46.316486: step 7430, loss = 0.80 (11999.7 examples/sec; 0.043 sec/batch)
2017-10-21 12:41:47.198390: step 7440, loss = 0.69 (12035.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:41:48.069207: step 7450, loss = 0.82 (11992.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:41:48.952161: step 7460, loss = 0.77 (10885.8 examples/sec; 0.047 sec/batch)
2017-10-21 12:41:49.844292: step 7470, loss = 0.80 (12122.3 examples/sec; 0.042 sec/batch)
2017-10-21 12:41:50.719345: step 7480, loss = 0.74 (11608.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:41:51.590926: step 7490, loss = 0.76 (12766.5 examples/sec; 0.040 sec/batch)
2017-10-21 12:41:52.470046: step 7500, loss = 0.71 (11510.1 examples/sec; 0.044 sec/batch)
2017-10-21 12:41:53.736996: step 7510, loss = 0.73 (18137.5 examples/sec; 0.028 sec/batch)
2017-10-21 12:41:54.352634: step 7520, loss = 0.68 (14596.7 examples/sec; 0.035 sec/batch)
2017-10-21 12:41:55.238686: step 7530, loss = 0.70 (11042.2 examples/sec; 0.046 sec/batch)
2017-10-21 12:41:56.106659: step 7540, loss = 0.72 (11738.9 examples/sec; 0.044 sec/batch)
2017-10-21 12:41:56.970804: step 7550, loss = 0.70 (11302.8 examples/sec; 0.045 sec/batch)
2017-10-21 12:41:57.835583: step 7560, loss = 0.76 (12907.6 examples/sec; 0.040 sec/batch)
2017-10-21 12:41:58.700842: step 7570, loss = 0.68 (11903.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:41:59.580402: step 7580, loss = 0.69 (10788.9 examples/sec; 0.047 sec/batch)
2017-10-21 12:42:00.465444: step 7590, loss = 0.68 (10719.0 examples/sec; 0.048 sec/batch)
2017-10-21 12:42:01.339826: step 7600, loss = 0.79 (10963.6 examples/sec; 0.047 sec/batch)
2017-10-21 12:42:02.613671: step 7610, loss = 0.69 (17358.7 examples/sec; 0.029 sec/batch)
2017-10-21 12:42:03.224970: step 7620, loss = 0.74 (15898.5 examples/sec; 0.032 sec/batch)
2017-10-21 12:42:04.083755: step 7630, loss = 0.75 (11409.5 examples/sec; 0.045 sec/batch)
2017-10-21 12:42:04.949622: step 7640, loss = 0.81 (11596.8 examples/sec; 0.044 sec/batch)
2017-10-21 12:42:05.808659: step 7650, loss = 0.67 (11082.6 examples/sec; 0.046 sec/batch)
2017-10-21 12:42:06.685251: step 7660, loss = 0.68 (12193.1 examples/sec; 0.042 sec/batch)
2017-10-21 12:42:07.550638: step 7670, loss = 0.71 (11152.3 examples/sec; 0.046 sec/batch)
2017-10-21 12:42:08.427944: step 7680, loss = 0.76 (11586.9 examples/sec; 0.044 sec/batch)
2017-10-21 12:42:09.298164: step 7690, loss = 0.64 (11878.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:42:10.172398: step 7700, loss = 0.70 (12529.7 examples/sec; 0.041 sec/batch)
2017-10-21 12:42:11.454393: step 7710, loss = 0.83 (17808.0 examples/sec; 0.029 sec/batch)
2017-10-21 12:42:12.072589: step 7720, loss = 0.72 (19143.7 examples/sec; 0.027 sec/batch)
2017-10-21 12:42:12.939304: step 7730, loss = 0.69 (11934.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:42:13.811466: step 7740, loss = 0.66 (12248.2 examples/sec; 0.042 sec/batch)
2017-10-21 12:42:14.701055: step 7750, loss = 0.71 (13373.1 examples/sec; 0.038 sec/batch)
2017-10-21 12:42:15.572595: step 7760, loss = 0.75 (11302.3 examples/sec; 0.045 sec/batch)
2017-10-21 12:42:16.444253: step 7770, loss = 0.67 (11372.4 examples/sec; 0.045 sec/batch)
2017-10-21 12:42:17.327892: step 7780, loss = 0.63 (12778.6 examples/sec; 0.040 sec/batch)
2017-10-21 12:42:18.201964: step 7790, loss = 0.73 (11562.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:42:19.085767: step 7800, loss = 0.69 (11736.1 examples/sec; 0.044 sec/batch)
2017-10-21 12:42:20.355260: step 7810, loss = 0.80 (17383.3 examples/sec; 0.029 sec/batch)
2017-10-21 12:42:20.959087: step 7820, loss = 0.68 (16108.3 examples/sec; 0.032 sec/batch)
2017-10-21 12:42:21.847291: step 7830, loss = 0.74 (11726.0 examples/sec; 0.044 sec/batch)
2017-10-21 12:42:22.722761: step 7840, loss = 0.69 (11738.3 examples/sec; 0.044 sec/batch)
2017-10-21 12:42:23.587240: step 7850, loss = 0.78 (11718.2 examples/sec; 0.044 sec/batch)
2017-10-21 12:42:24.469225: step 7860, loss = 0.72 (12449.0 examples/sec; 0.041 sec/batch)
2017-10-21 12:42:25.339413: step 7870, loss = 0.73 (12867.1 examples/sec; 0.040 sec/batch)
2017-10-21 12:42:26.194735: step 7880, loss = 0.73 (12200.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:42:27.076677: step 7890, loss = 0.71 (11277.1 examples/sec; 0.045 sec/batch)
2017-10-21 12:42:27.965047: step 7900, loss = 0.73 (12152.6 examples/sec; 0.042 sec/batch)
2017-10-21 12:42:29.243647: step 7910, loss = 0.70 (17939.6 examples/sec; 0.029 sec/batch)
2017-10-21 12:42:29.855417: step 7920, loss = 0.68 (16274.7 examples/sec; 0.031 sec/batch)
2017-10-21 12:42:30.710603: step 7930, loss = 0.74 (11969.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:42:31.592171: step 7940, loss = 0.79 (11815.9 examples/sec; 0.043 sec/batch)
2017-10-21 12:42:32.471610: step 7950, loss = 0.67 (12382.1 examples/sec; 0.041 sec/batch)
2017-10-21 12:42:33.348621: step 7960, loss = 0.59 (12108.4 examples/sec; 0.042 sec/batch)
2017-10-21 12:42:34.229733: step 7970, loss = 0.80 (11233.3 examples/sec; 0.046 sec/batch)
2017-10-21 12:42:35.100640: step 7980, loss = 0.71 (11801.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:42:35.978820: step 7990, loss = 0.69 (11617.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:42:36.850476: step 8000, loss = 0.79 (11468.2 examples/sec; 0.045 sec/batch)
2017-10-21 12:42:38.315307: step 8010, loss = 0.65 (18333.6 examples/sec; 0.028 sec/batch)
2017-10-21 12:42:38.916012: step 8020, loss = 0.72 (17618.4 examples/sec; 0.029 sec/batch)
2017-10-21 12:42:39.777883: step 8030, loss = 0.74 (11033.7 examples/sec; 0.046 sec/batch)
2017-10-21 12:42:40.686150: step 8040, loss = 0.75 (11596.3 examples/sec; 0.044 sec/batch)
2017-10-21 12:42:41.568909: step 8050, loss = 0.71 (11646.7 examples/sec; 0.044 sec/batch)
2017-10-21 12:42:42.447512: step 8060, loss = 0.76 (11723.9 examples/sec; 0.044 sec/batch)
2017-10-21 12:42:43.326154: step 8070, loss = 0.81 (11553.1 examples/sec; 0.044 sec/batch)
2017-10-21 12:42:44.200881: step 8080, loss = 0.71 (11559.8 examples/sec; 0.044 sec/batch)
2017-10-21 12:42:45.093339: step 8090, loss = 0.63 (11294.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:42:45.956923: step 8100, loss = 0.63 (12043.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:42:47.232614: step 8110, loss = 0.78 (17014.0 examples/sec; 0.030 sec/batch)
2017-10-21 12:42:47.839123: step 8120, loss = 0.68 (14429.3 examples/sec; 0.035 sec/batch)
2017-10-21 12:42:48.713754: step 8130, loss = 0.80 (11473.3 examples/sec; 0.045 sec/batch)
2017-10-21 12:42:49.583477: step 8140, loss = 0.63 (11971.8 examples/sec; 0.043 sec/batch)
2017-10-21 12:42:50.474617: step 8150, loss = 0.70 (12510.8 examples/sec; 0.041 sec/batch)
2017-10-21 12:42:51.350792: step 8160, loss = 0.68 (12421.7 examples/sec; 0.041 sec/batch)
2017-10-21 12:42:52.224585: step 8170, loss = 0.82 (12121.3 examples/sec; 0.042 sec/batch)
2017-10-21 12:42:53.086467: step 8180, loss = 0.77 (12353.6 examples/sec; 0.041 sec/batch)
2017-10-21 12:42:53.969907: step 8190, loss = 0.64 (12080.6 examples/sec; 0.042 sec/batch)
2017-10-21 12:42:54.842656: step 8200, loss = 0.76 (11603.2 examples/sec; 0.044 sec/batch)
2017-10-21 12:42:56.128028: step 8210, loss = 0.62 (17651.9 examples/sec; 0.029 sec/batch)
2017-10-21 12:42:56.723603: step 8220, loss = 0.70 (17229.8 examples/sec; 0.030 sec/batch)
2017-10-21 12:42:57.588868: step 8230, loss = 0.65 (11839.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:42:58.466794: step 8240, loss = 0.71 (11081.7 examples/sec; 0.046 sec/batch)
2017-10-21 12:42:59.337659: step 8250, loss = 0.70 (11465.5 examples/sec; 0.045 sec/batch)
2017-10-21 12:43:00.210833: step 8260, loss = 0.74 (12455.9 examples/sec; 0.041 sec/batch)
2017-10-21 12:43:01.117504: step 8270, loss = 0.73 (11405.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:43:01.997010: step 8280, loss = 0.71 (10673.2 examples/sec; 0.048 sec/batch)
2017-10-21 12:43:02.863013: step 8290, loss = 0.87 (12176.6 examples/sec; 0.042 sec/batch)
2017-10-21 12:43:03.737876: step 8300, loss = 0.63 (11855.1 examples/sec; 0.043 sec/batch)
2017-10-21 12:43:05.015006: step 8310, loss = 0.64 (18123.6 examples/sec; 0.028 sec/batch)
2017-10-21 12:43:05.605639: step 8320, loss = 0.75 (17571.4 examples/sec; 0.029 sec/batch)
2017-10-21 12:43:06.466726: step 8330, loss = 0.72 (11634.5 examples/sec; 0.044 sec/batch)
2017-10-21 12:43:07.341222: step 8340, loss = 0.62 (11388.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:43:08.212553: step 8350, loss = 0.68 (12414.7 examples/sec; 0.041 sec/batch)
2017-10-21 12:43:09.090417: step 8360, loss = 0.71 (10956.5 examples/sec; 0.047 sec/batch)
2017-10-21 12:43:09.969436: step 8370, loss = 0.74 (11228.5 examples/sec; 0.046 sec/batch)
2017-10-21 12:43:10.855001: step 8380, loss = 0.77 (11534.0 examples/sec; 0.044 sec/batch)
2017-10-21 12:43:11.733657: step 8390, loss = 0.76 (11828.6 examples/sec; 0.043 sec/batch)
2017-10-21 12:43:12.605217: step 8400, loss = 0.69 (11811.8 examples/sec; 0.043 sec/batch)
2017-10-21 12:43:13.883156: step 8410, loss = 0.65 (17857.7 examples/sec; 0.029 sec/batch)
2017-10-21 12:43:14.482799: step 8420, loss = 0.69 (15061.5 examples/sec; 0.034 sec/batch)
2017-10-21 12:43:15.353779: step 8430, loss = 0.88 (12767.5 examples/sec; 0.040 sec/batch)
2017-10-21 12:43:16.234454: step 8440, loss = 0.68 (11816.5 examples/sec; 0.043 sec/batch)
2017-10-21 12:43:17.108589: step 8450, loss = 0.75 (12652.4 examples/sec; 0.040 sec/batch)
2017-10-21 12:43:17.981279: step 8460, loss = 0.69 (12012.6 examples/sec; 0.043 sec/batch)
2017-10-21 12:43:18.856837: step 8470, loss = 0.60 (11978.5 examples/sec; 0.043 sec/batch)
2017-10-21 12:43:19.748379: step 8480, loss = 0.74 (12179.8 examples/sec; 0.042 sec/batch)
2017-10-21 12:43:20.602513: step 8490, loss = 0.78 (13101.3 examples/sec; 0.039 sec/batch)
2017-10-21 12:43:21.504009: step 8500, loss = 0.81 (10765.2 examples/sec; 0.048 sec/batch)
2017-10-21 12:43:22.797922: step 8510, loss = 0.66 (18161.6 examples/sec; 0.028 sec/batch)
2017-10-21 12:43:23.399077: step 8520, loss = 0.73 (15695.4 examples/sec; 0.033 sec/batch)
2017-10-21 12:43:24.255853: step 8530, loss = 0.84 (12196.4 examples/sec; 0.042 sec/batch)
2017-10-21 12:43:25.139022: step 8540, loss = 0.66 (11557.4 examples/sec; 0.044 sec/batch)
2017-10-21 12:43:26.011212: step 8550, loss = 0.64 (11447.9 examples/sec; 0.045 sec/batch)
2017-10-21 12:43:26.900828: step 8560, loss = 0.70 (11625.0 examples/sec; 0.044 sec/batch)
2017-10-21 12:43:27.766728: step 8570, loss = 0.76 (11589.0 examples/sec; 0.044 sec/batch)
2017-10-21 12:43:28.624096: step 8580, loss = 0.73 (11903.8 examples/sec; 0.043 sec/batch)
2017-10-21 12:43:29.504350: step 8590, loss = 0.78 (12494.0 examples/sec; 0.041 sec/batch)
2017-10-21 12:43:30.377281: step 8600, loss = 0.70 (12496.2 examples/sec; 0.041 sec/batch)
2017-10-21 12:43:31.678299: step 8610, loss = 0.83 (17423.3 examples/sec; 0.029 sec/batch)
2017-10-21 12:43:32.272625: step 8620, loss = 0.68 (17701.1 examples/sec; 0.029 sec/batch)
2017-10-21 12:43:33.147212: step 8630, loss = 0.75 (11020.2 examples/sec; 0.046 sec/batch)
2017-10-21 12:43:34.018413: step 8640, loss = 0.79 (10877.5 examples/sec; 0.047 sec/batch)
2017-10-21 12:43:34.902964: step 8650, loss = 0.75 (11657.5 examples/sec; 0.044 sec/batch)
2017-10-21 12:43:35.770240: step 8660, loss = 0.89 (11589.2 examples/sec; 0.044 sec/batch)
2017-10-21 12:43:36.650750: step 8670, loss = 0.66 (12446.1 examples/sec; 0.041 sec/batch)
2017-10-21 12:43:37.527853: step 8680, loss = 0.64 (12441.1 examples/sec; 0.041 sec/batch)
2017-10-21 12:43:38.407606: step 8690, loss = 0.68 (10939.7 examples/sec; 0.047 sec/batch)
2017-10-21 12:43:39.277756: step 8700, loss = 0.72 (12323.9 examples/sec; 0.042 sec/batch)
2017-10-21 12:43:40.549837: step 8710, loss = 0.64 (19141.7 examples/sec; 0.027 sec/batch)
2017-10-21 12:43:41.159931: step 8720, loss = 0.69 (16448.8 examples/sec; 0.031 sec/batch)
2017-10-21 12:43:42.024268: step 8730, loss = 0.66 (12373.7 examples/sec; 0.041 sec/batch)
2017-10-21 12:43:42.911087: step 8740, loss = 0.74 (12742.8 examples/sec; 0.040 sec/batch)
2017-10-21 12:43:43.793509: step 8750, loss = 0.70 (10994.2 examples/sec; 0.047 sec/batch)
2017-10-21 12:43:44.653065: step 8760, loss = 0.72 (11783.7 examples/sec; 0.043 sec/batch)
2017-10-21 12:43:45.522270: step 8770, loss = 0.74 (11708.7 examples/sec; 0.044 sec/batch)
2017-10-21 12:43:46.397766: step 8780, loss = 0.64 (11276.2 examples/sec; 0.045 sec/batch)
2017-10-21 12:43:47.277254: step 8790, loss = 0.64 (13683.7 examples/sec; 0.037 sec/batch)
2017-10-21 12:43:48.160389: step 8800, loss = 0.71 (12173.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:43:49.440341: step 8810, loss = 0.71 (16711.5 examples/sec; 0.031 sec/batch)
2017-10-21 12:43:50.042024: step 8820, loss = 0.64 (16277.4 examples/sec; 0.031 sec/batch)
2017-10-21 12:43:50.915955: step 8830, loss = 0.67 (11270.6 examples/sec; 0.045 sec/batch)
2017-10-21 12:43:51.805631: step 8840, loss = 0.76 (12212.7 examples/sec; 0.042 sec/batch)
2017-10-21 12:43:52.676538: step 8850, loss = 0.70 (12172.2 examples/sec; 0.042 sec/batch)
2017-10-21 12:43:53.562616: step 8860, loss = 0.64 (10638.2 examples/sec; 0.048 sec/batch)
2017-10-21 12:43:54.448460: step 8870, loss = 0.67 (11341.2 examples/sec; 0.045 sec/batch)
2017-10-21 12:43:55.301687: step 8880, loss = 0.64 (11685.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:43:56.187932: step 8890, loss = 0.71 (11338.0 examples/sec; 0.045 sec/batch)
2017-10-21 12:43:57.056646: step 8900, loss = 0.69 (11625.5 examples/sec; 0.044 sec/batch)
2017-10-21 12:43:58.319984: step 8910, loss = 0.86 (17909.1 examples/sec; 0.029 sec/batch)
2017-10-21 12:43:58.922077: step 8920, loss = 0.64 (13321.9 examples/sec; 0.038 sec/batch)
2017-10-21 12:43:59.805027: step 8930, loss = 0.67 (11567.4 examples/sec; 0.044 sec/batch)
2017-10-21 12:44:00.717953: step 8940, loss = 0.61 (12662.3 examples/sec; 0.040 sec/batch)
2017-10-21 12:44:01.595711: step 8950, loss = 0.76 (12121.0 examples/sec; 0.042 sec/batch)
2017-10-21 12:44:02.489603: step 8960, loss = 0.80 (12254.2 examples/sec; 0.042 sec/batch)
2017-10-21 12:44:03.376400: step 8970, loss = 0.63 (11210.1 examples/sec; 0.046 sec/batch)
2017-10-21 12:44:04.256292: step 8980, loss = 0.68 (11185.9 examples/sec; 0.046 sec/batch)
2017-10-21 12:44:05.119581: step 8990, loss = 0.88 (11729.7 examples/sec; 0.044 sec/batch)
2017-10-21 12:44:06.006945: step 9000, loss = 0.77 (10545.6 examples/sec; 0.049 sec/batch)
2017-10-21 12:44:07.420659: step 9010, loss = 0.68 (18624.2 examples/sec; 0.027 sec/batch)
2017-10-21 12:44:08.027871: step 9020, loss = 0.65 (13462.1 examples/sec; 0.038 sec/batch)
2017-10-21 12:44:08.904373: step 9030, loss = 0.73 (12196.6 examples/sec; 0.042 sec/batch)
2017-10-21 12:44:09.796592: step 9040, loss = 0.67 (10964.5 examples/sec; 0.047 sec/batch)
2017-10-21 12:44:10.669885: step 9050, loss = 0.62 (12294.6 examples/sec; 0.042 sec/batch)
2017-10-21 12:44:11.564144: step 9060, loss = 0.65 (11977.4 examples/sec; 0.043 sec/batch)
2017-10-21 12:44:12.441061: step 9070, loss = 0.64 (12202.4 examples/sec; 0.042 sec/batch)
2017-10-21 12:44:13.327136: step 9080, loss = 0.67 (11662.1 examples/sec; 0.044 sec/batch)
2017-10-21 12:44:14.199932: step 9090, loss = 0.75 (11669.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:44:15.069120: step 9100, loss = 0.64 (12029.7 examples/sec; 0.043 sec/batch)
2017-10-21 12:44:16.340266: step 9110, loss = 0.67 (18949.3 examples/sec; 0.027 sec/batch)
2017-10-21 12:44:16.962321: step 9120, loss = 0.73 (16719.4 examples/sec; 0.031 sec/batch)
2017-10-21 12:44:17.839175: step 9130, loss = 0.67 (11158.5 examples/sec; 0.046 sec/batch)
2017-10-21 12:44:18.722838: step 9140, loss = 0.63 (11904.0 examples/sec; 0.043 sec/batch)
2017-10-21 12:44:19.604574: step 9150, loss = 0.71 (11958.5 examples/sec; 0.043 sec/batch)
2017-10-21 12:44:20.490279: step 9160, loss = 0.64 (10888.5 examples/sec; 0.047 sec/batch)
2017-10-21 12:44:21.378081: step 9170, loss = 0.74 (12149.6 examples/sec; 0.042 sec/batch)
2017-10-21 12:44:22.246160: step 9180, loss = 0.56 (11231.1 examples/sec; 0.046 sec/batch)
2017-10-21 12:44:23.111046: step 9190, loss = 0.63 (11957.9 examples/sec; 0.043 sec/batch)
2017-10-21 12:44:24.004923: step 9200, loss = 0.76 (10122.7 examples/sec; 0.051 sec/batch)
2017-10-21 12:44:25.271614: step 9210, loss = 0.68 (16278.4 examples/sec; 0.031 sec/batch)
2017-10-21 12:44:25.869005: step 9220, loss = 0.69 (18173.4 examples/sec; 0.028 sec/batch)
2017-10-21 12:44:26.753820: step 9230, loss = 0.64 (12161.2 examples/sec; 0.042 sec/batch)
2017-10-21 12:44:27.667324: step 9240, loss = 0.83 (10225.1 examples/sec; 0.050 sec/batch)
2017-10-21 12:44:28.623522: step 9250, loss = 0.65 (10956.9 examples/sec; 0.047 sec/batch)
2017-10-21 12:44:29.642439: step 9260, loss = 0.69 (10384.7 examples/sec; 0.049 sec/batch)
2017-10-21 12:44:30.638247: step 9270, loss = 0.71 (9847.8 examples/sec; 0.052 sec/batch)
2017-10-21 12:44:31.635201: step 9280, loss = 0.64 (10703.4 examples/sec; 0.048 sec/batch)
2017-10-21 12:44:32.627205: step 9290, loss = 0.69 (10195.5 examples/sec; 0.050 sec/batch)
2017-10-21 12:44:33.610968: step 9300, loss = 0.67 (9535.9 examples/sec; 0.054 sec/batch)
2017-10-21 12:44:34.880394: step 9310, loss = 0.60 (17155.7 examples/sec; 0.030 sec/batch)
2017-10-21 12:44:35.573366: step 9320, loss = 0.71 (12111.0 examples/sec; 0.042 sec/batch)
2017-10-21 12:44:36.460548: step 9330, loss = 0.61 (12676.7 examples/sec; 0.040 sec/batch)
2017-10-21 12:44:37.341135: step 9340, loss = 0.76 (11121.7 examples/sec; 0.046 sec/batch)
2017-10-21 12:44:38.221530: step 9350, loss = 0.73 (12881.6 examples/sec; 0.040 sec/batch)
2017-10-21 12:44:39.121454: step 9360, loss = 0.78 (12242.0 examples/sec; 0.042 sec/batch)
2017-10-21 12:44:40.004480: step 9370, loss = 0.71 (11828.5 examples/sec; 0.043 sec/batch)
2017-10-21 12:44:40.866319: step 9380, loss = 0.75 (11392.8 examples/sec; 0.045 sec/batch)
2017-10-21 12:44:41.762528: step 9390, loss = 0.62 (12213.0 examples/sec; 0.042 sec/batch)
2017-10-21 12:44:42.660796: step 9400, loss = 0.67 (10897.0 examples/sec; 0.047 sec/batch)
2017-10-21 12:44:43.945519: step 9410, loss = 0.61 (17371.6 examples/sec; 0.029 sec/batch)
2017-10-21 12:44:44.554554: step 9420, loss = 0.68 (15999.7 examples/sec; 0.032 sec/batch)
2017-10-21 12:44:45.443186: step 9430, loss = 0.68 (11859.4 examples/sec; 0.043 sec/batch)
2017-10-21 12:44:46.314230: step 9440, loss = 0.69 (12554.4 examples/sec; 0.041 sec/batch)
2017-10-21 12:44:47.193783: step 9450, loss = 0.68 (11748.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:44:48.084179: step 9460, loss = 0.66 (11814.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:44:48.968260: step 9470, loss = 0.71 (11279.5 examples/sec; 0.045 sec/batch)
2017-10-21 12:44:49.849057: step 9480, loss = 0.82 (12464.0 examples/sec; 0.041 sec/batch)
2017-10-21 12:44:50.727697: step 9490, loss = 0.62 (11847.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:44:51.616236: step 9500, loss = 0.68 (12166.8 examples/sec; 0.042 sec/batch)
2017-10-21 12:44:52.909427: step 9510, loss = 0.68 (17415.2 examples/sec; 0.029 sec/batch)
2017-10-21 12:44:53.519675: step 9520, loss = 0.66 (17391.6 examples/sec; 0.029 sec/batch)
2017-10-21 12:44:54.397623: step 9530, loss = 0.69 (11345.9 examples/sec; 0.045 sec/batch)
2017-10-21 12:44:55.272407: step 9540, loss = 0.70 (11732.4 examples/sec; 0.044 sec/batch)
2017-10-21 12:44:56.143175: step 9550, loss = 0.70 (11058.8 examples/sec; 0.046 sec/batch)
2017-10-21 12:44:57.031065: step 9560, loss = 0.68 (11518.2 examples/sec; 0.044 sec/batch)
2017-10-21 12:44:57.919944: step 9570, loss = 0.65 (12213.9 examples/sec; 0.042 sec/batch)
2017-10-21 12:44:58.803749: step 9580, loss = 0.64 (11303.7 examples/sec; 0.045 sec/batch)
2017-10-21 12:44:59.680515: step 9590, loss = 0.80 (12589.5 examples/sec; 0.041 sec/batch)
2017-10-21 12:45:00.602348: step 9600, loss = 0.63 (10539.4 examples/sec; 0.049 sec/batch)
2017-10-21 12:45:01.870054: step 9610, loss = 0.70 (17564.7 examples/sec; 0.029 sec/batch)
2017-10-21 12:45:02.491226: step 9620, loss = 0.76 (13401.5 examples/sec; 0.038 sec/batch)
2017-10-21 12:45:03.372190: step 9630, loss = 0.70 (11148.5 examples/sec; 0.046 sec/batch)
2017-10-21 12:45:04.237739: step 9640, loss = 0.67 (11464.8 examples/sec; 0.045 sec/batch)
2017-10-21 12:45:05.123733: step 9650, loss = 0.67 (11824.7 examples/sec; 0.043 sec/batch)
2017-10-21 12:45:06.018505: step 9660, loss = 0.61 (12109.2 examples/sec; 0.042 sec/batch)
2017-10-21 12:45:06.891952: step 9670, loss = 0.65 (12128.9 examples/sec; 0.042 sec/batch)
2017-10-21 12:45:07.764433: step 9680, loss = 0.64 (12079.0 examples/sec; 0.042 sec/batch)
2017-10-21 12:45:08.626462: step 9690, loss = 0.66 (11611.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:45:09.494394: step 9700, loss = 0.75 (12879.7 examples/sec; 0.040 sec/batch)
2017-10-21 12:45:10.769939: step 9710, loss = 0.68 (17441.7 examples/sec; 0.029 sec/batch)
2017-10-21 12:45:11.367310: step 9720, loss = 0.74 (17350.0 examples/sec; 0.030 sec/batch)
2017-10-21 12:45:12.253270: step 9730, loss = 0.71 (10215.0 examples/sec; 0.050 sec/batch)
2017-10-21 12:45:13.138044: step 9740, loss = 0.77 (11743.1 examples/sec; 0.044 sec/batch)
2017-10-21 12:45:14.026193: step 9750, loss = 0.63 (11816.8 examples/sec; 0.043 sec/batch)
2017-10-21 12:45:14.914625: step 9760, loss = 0.61 (12152.6 examples/sec; 0.042 sec/batch)
2017-10-21 12:45:15.796983: step 9770, loss = 0.73 (10888.5 examples/sec; 0.047 sec/batch)
2017-10-21 12:45:16.684787: step 9780, loss = 0.72 (11786.2 examples/sec; 0.043 sec/batch)
2017-10-21 12:45:17.573292: step 9790, loss = 0.66 (11574.2 examples/sec; 0.044 sec/batch)
2017-10-21 12:45:18.467020: step 9800, loss = 0.77 (12324.5 examples/sec; 0.042 sec/batch)
2017-10-21 12:45:19.727433: step 9810, loss = 0.61 (18098.6 examples/sec; 0.028 sec/batch)
2017-10-21 12:45:20.343462: step 9820, loss = 0.62 (12813.7 examples/sec; 0.040 sec/batch)
2017-10-21 12:45:21.223228: step 9830, loss = 0.62 (11649.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:45:22.144504: step 9840, loss = 0.73 (10979.3 examples/sec; 0.047 sec/batch)
2017-10-21 12:45:23.046621: step 9850, loss = 0.66 (11173.5 examples/sec; 0.046 sec/batch)
2017-10-21 12:45:23.928478: step 9860, loss = 0.60 (11909.3 examples/sec; 0.043 sec/batch)
2017-10-21 12:45:24.805849: step 9870, loss = 0.64 (12159.1 examples/sec; 0.042 sec/batch)
2017-10-21 12:45:25.661789: step 9880, loss = 0.74 (11760.5 examples/sec; 0.044 sec/batch)
2017-10-21 12:45:26.527829: step 9890, loss = 0.66 (12378.7 examples/sec; 0.041 sec/batch)
2017-10-21 12:45:27.406908: step 9900, loss = 0.71 (11656.6 examples/sec; 0.044 sec/batch)
2017-10-21 12:45:28.711197: step 9910, loss = 0.61 (17626.8 examples/sec; 0.029 sec/batch)
2017-10-21 12:45:29.315606: step 9920, loss = 0.72 (16662.7 examples/sec; 0.031 sec/batch)
2017-10-21 12:45:30.191506: step 9930, loss = 0.66 (11154.3 examples/sec; 0.046 sec/batch)
2017-10-21 12:45:31.052132: step 9940, loss = 0.70 (11445.6 examples/sec; 0.045 sec/batch)
2017-10-21 12:45:31.933559: step 9950, loss = 0.77 (11944.8 examples/sec; 0.043 sec/batch)
2017-10-21 12:45:32.814685: step 9960, loss = 0.70 (11700.4 examples/sec; 0.044 sec/batch)
2017-10-21 12:45:33.707847: step 9970, loss = 0.54 (10704.0 examples/sec; 0.048 sec/batch)
2017-10-21 12:45:34.563367: step 9980, loss = 0.69 (11764.7 examples/sec; 0.044 sec/batch)
2017-10-21 12:45:35.442545: step 9990, loss = 0.67 (11469.2 examples/sec; 0.045 sec/batch)