-
Notifications
You must be signed in to change notification settings - Fork 8
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
1 parent
afec7b6
commit ead3785
Showing
31 changed files
with
234,417 additions
and
0 deletions.
There are no files selected for viewing
77 changes: 77 additions & 0 deletions
77
examples/long_text/0.5.yelp.250.512.dynamiclr.par2.gan.new.out
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,77 @@ | ||
train_file:/home/syp/zwr/txtgen/examples/long_text/yelp_data/pos/pos.train.txt | ||
valid_file:/home/syp/zwr/txtgen/examples/long_text/yelp_data/pos/pos.valid.txt | ||
logdir:./log_dir/pos.bsize250.epoch150.seqlen16.dynamic_lr.fixed_mask.present0.5.partition2.hidden512.gan/ | ||
step:35601 source:(250, 17) g_loss:1.548106 d_loss:0.000000 ppl:4.689394 lr:0.000234 | ||
step:35801 source:(250, 17) g_loss:1.579333 d_loss:0.000000 ppl:4.838124 lr:0.000234 | ||
step:36001 source:(250, 17) g_loss:1.582944 d_loss:0.000000 ppl:4.855135 lr:0.000233 | ||
step:36201 source:(250, 17) g_loss:1.534200 d_loss:0.000000 ppl:4.624115 lr:0.000232 | ||
step:36401 source:(250, 17) g_loss:1.558126 d_loss:0.000000 ppl:4.735978 lr:0.000232 | ||
step:36601 source:(250, 17) g_loss:1.533349 d_loss:0.000000 ppl:4.620012 lr:0.000231 | ||
step:36801 source:(250, 17) g_loss:1.530993 d_loss:0.000000 ppl:4.609141 lr:0.000230 | ||
step:37001 source:(250, 17) g_loss:1.530852 d_loss:0.000000 ppl:4.608454 lr:0.000230 | ||
epoch:125 test_bleu:42.54753291606903 template_bleu:32.497718930244446 test_loss:4.474950790405273 test_ppl:100.27088165283203 | ||
epoch:125 train_g_bleu:71.63092494010925 template_bleu:32.185012102127075 train_g_loss:1.936802625656128 train_g_ppl:7.196612358093262 | ||
step:37201 source:(250, 17) g_loss:1.536745 d_loss:0.000000 ppl:4.635620 lr:0.000229 | ||
step:37401 source:(250, 17) g_loss:1.540416 d_loss:0.000000 ppl:4.652607 lr:0.000229 | ||
step:37601 source:(250, 17) g_loss:1.537606 d_loss:0.000000 ppl:4.639566 lr:0.000228 | ||
step:37801 source:(250, 17) g_loss:1.536425 d_loss:0.000000 ppl:4.634108 lr:0.000227 | ||
step:38001 source:(250, 17) g_loss:1.547389 d_loss:0.000000 ppl:4.685195 lr:0.000227 | ||
step:38201 source:(250, 17) g_loss:1.531610 d_loss:0.000000 ppl:4.612047 lr:0.000226 | ||
step:38401 source:(250, 17) g_loss:1.527591 d_loss:0.000000 ppl:4.593348 lr:0.000226 | ||
step:38601 source:(250, 17) g_loss:1.520976 d_loss:0.000000 ppl:4.563033 lr:0.000225 | ||
step:38801 source:(250, 17) g_loss:1.539520 d_loss:0.000000 ppl:4.648195 lr:0.000224 | ||
step:39001 source:(250, 17) g_loss:1.546572 d_loss:0.000000 ppl:4.681340 lr:0.000224 | ||
epoch:130 test_bleu:41.79803431034088 template_bleu:32.497718930244446 test_loss:4.5916924476623535 test_ppl:114.98224639892578 | ||
epoch:130 train_g_bleu:73.499596118927 template_bleu:32.185012102127075 train_g_loss:1.9240975379943848 train_g_ppl:7.118496894836426 | ||
step:39201 source:(250, 17) g_loss:1.560775 d_loss:0.000000 ppl:4.748385 lr:0.000223 | ||
step:39401 source:(250, 17) g_loss:1.524132 d_loss:0.000000 ppl:4.577285 lr:0.000223 | ||
step:39601 source:(250, 17) g_loss:1.525652 d_loss:0.000000 ppl:4.584320 lr:0.000222 | ||
step:39801 source:(250, 17) g_loss:1.519901 d_loss:0.000000 ppl:4.558082 lr:0.000222 | ||
step:40001 source:(250, 17) g_loss:1.539388 d_loss:0.000000 ppl:4.647891 lr:0.000221 | ||
step:40201 source:(250, 17) g_loss:1.527815 d_loss:0.000000 ppl:4.594221 lr:0.000220 | ||
step:40401 source:(250, 17) g_loss:1.521839 d_loss:0.000000 ppl:4.566904 lr:0.000220 | ||
step:40601 source:(250, 17) g_loss:1.510919 d_loss:0.000000 ppl:4.517366 lr:0.000219 | ||
step:40801 source:(250, 17) g_loss:1.502446 d_loss:0.000000 ppl:4.479192 lr:0.000219 | ||
step:41001 source:(250, 17) g_loss:1.509020 d_loss:0.000000 ppl:4.508721 lr:0.000218 | ||
step:41201 source:(250, 17) g_loss:1.511285 d_loss:0.000000 ppl:4.518903 lr:0.000218 | ||
epoch:135 test_bleu:42.75813102722168 template_bleu:32.497718930244446 test_loss:4.483145713806152 test_ppl:102.34559631347656 | ||
epoch:135 train_g_bleu:77.19091773033142 template_bleu:32.185012102127075 train_g_loss:1.8359583616256714 train_g_ppl:6.5244832038879395 | ||
step:41401 source:(250, 17) g_loss:1.523722 d_loss:0.000000 ppl:4.575447 lr:0.000217 | ||
step:41601 source:(250, 17) g_loss:1.498592 d_loss:0.000000 ppl:4.461993 lr:0.000217 | ||
step:41801 source:(250, 17) g_loss:1.499550 d_loss:0.000000 ppl:4.466269 lr:0.000216 | ||
step:42001 source:(250, 17) g_loss:1.504244 d_loss:0.000000 ppl:4.487231 lr:0.000216 | ||
step:42201 source:(250, 17) g_loss:1.511527 d_loss:0.000000 ppl:4.519944 lr:0.000215 | ||
step:42401 source:(250, 17) g_loss:1.512354 d_loss:0.000000 ppl:4.523767 lr:0.000215 | ||
step:42601 source:(250, 17) g_loss:1.514217 d_loss:0.000000 ppl:4.532143 lr:0.000214 | ||
step:42801 source:(250, 17) g_loss:1.508126 d_loss:0.000000 ppl:4.504862 lr:0.000214 | ||
step:43001 source:(250, 17) g_loss:1.525540 d_loss:0.000000 ppl:4.584385 lr:0.000213 | ||
step:43201 source:(250, 17) g_loss:1.506420 d_loss:0.000000 ppl:4.497983 lr:0.000213 | ||
epoch:140 test_bleu:42.565733194351196 template_bleu:32.497718930244446 test_loss:4.477306842803955 test_ppl:102.73019409179688 | ||
epoch:140 train_g_bleu:74.13867712020874 template_bleu:32.185012102127075 train_g_loss:1.9106131792068481 train_g_ppl:7.050699710845947 | ||
step:43401 source:(250, 17) g_loss:1.502327 d_loss:0.000000 ppl:4.479233 lr:0.000212 | ||
step:43601 source:(250, 17) g_loss:1.504553 d_loss:0.000000 ppl:4.489278 lr:0.000212 | ||
step:43801 source:(250, 17) g_loss:1.504979 d_loss:0.000000 ppl:4.491137 lr:0.000211 | ||
step:44001 source:(250, 17) g_loss:1.516961 d_loss:0.000000 ppl:4.545187 lr:0.000211 | ||
step:44201 source:(250, 17) g_loss:1.517699 d_loss:0.000000 ppl:4.548413 lr:0.000210 | ||
step:44401 source:(250, 17) g_loss:1.501388 d_loss:0.000000 ppl:4.475166 lr:0.000210 | ||
step:44601 source:(250, 17) g_loss:1.500668 d_loss:0.000000 ppl:4.471860 lr:0.000209 | ||
step:44801 source:(250, 17) g_loss:1.491412 d_loss:0.000000 ppl:4.430841 lr:0.000209 | ||
step:45001 source:(250, 17) g_loss:1.536006 d_loss:0.000000 ppl:4.632726 lr:0.000208 | ||
step:45201 source:(250, 17) g_loss:1.491506 d_loss:0.000000 ppl:4.430898 lr:0.000208 | ||
step:45401 source:(250, 17) g_loss:1.503309 d_loss:0.000000 ppl:4.483981 lr:0.000207 | ||
epoch:145 test_bleu:42.51361787319183 template_bleu:32.497718930244446 test_loss:4.481626033782959 test_ppl:102.0898208618164 | ||
epoch:145 train_g_bleu:76.32933855056763 template_bleu:32.185012102127075 train_g_loss:1.8196110725402832 train_g_ppl:6.396180629730225 | ||
step:45601 source:(250, 17) g_loss:1.485476 d_loss:0.000000 ppl:4.404444 lr:0.000207 | ||
step:45801 source:(250, 17) g_loss:1.504960 d_loss:0.000000 ppl:4.490861 lr:0.000207 | ||
step:46001 source:(250, 17) g_loss:1.502108 d_loss:0.000000 ppl:4.478140 lr:0.000206 | ||
step:46201 source:(250, 17) g_loss:1.491864 d_loss:0.000000 ppl:4.432919 lr:0.000206 | ||
step:46401 source:(250, 17) g_loss:1.538060 d_loss:0.000000 ppl:4.641941 lr:0.000205 | ||
step:46601 source:(250, 17) g_loss:1.498397 d_loss:0.000000 ppl:4.461542 lr:0.000205 | ||
step:46801 source:(250, 17) g_loss:1.515822 d_loss:0.000000 ppl:4.539786 lr:0.000204 | ||
step:47001 source:(250, 17) g_loss:1.501776 d_loss:0.000000 ppl:4.477035 lr:0.000204 | ||
epoch:149 test_bleu:42.534586787223816 template_bleu:32.497718930244446 test_loss:4.508484363555908 test_ppl:107.5583724975586 | ||
epoch:149 train_g_bleu:77.55188941955566 template_bleu:32.185012102127075 train_g_loss:1.8097453117370605 train_g_ppl:6.355318069458008 | ||
step:47201 source:(250, 17) g_loss:1.492661 d_loss:0.000000 ppl:4.436079 lr:0.000203 | ||
step:47401 source:(250, 17) g_loss:1.473005 d_loss:0.000000 ppl:4.349542 lr:0.000203 | ||
/home/syp/miniconda2/envs/newtf/lib/python3.6/importlib/_bootstrap.py:205: RuntimeWarning: compiletime version 3.5 of module 'tensorflow.python.framework.fast_tensor_util' does not match runtime version 3.6 | ||
return f(*args, **kwds) |
78 changes: 78 additions & 0 deletions
78
examples/long_text/0.6.yelp.250.512.dynamiclr.par2.gan.new.out
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,78 @@ | ||
train_file:/home/syp/zwr/txtgen/examples/long_text/yelp_data/pos/pos.train.txt | ||
valid_file:/home/syp/zwr/txtgen/examples/long_text/yelp_data/pos/pos.valid.txt | ||
logdir:./log_dir/pos.bsize250.epoch150.seqlen16.dynamic_lr.fixed_mask.present0.6.partition2.hidden512.gan/ | ||
step:33401 source:(250, 17) g_loss:1.559684 d_loss:0.000000 ppl:4.744217 lr:0.000242 | ||
step:33601 source:(250, 17) g_loss:1.542443 d_loss:0.000000 ppl:4.662971 lr:0.000241 | ||
step:33801 source:(250, 17) g_loss:1.574657 d_loss:0.000000 ppl:4.815094 lr:0.000240 | ||
step:34001 source:(250, 17) g_loss:1.564921 d_loss:0.000000 ppl:4.768654 lr:0.000240 | ||
step:34201 source:(250, 17) g_loss:1.602528 d_loss:0.000000 ppl:4.951150 lr:0.000239 | ||
step:34401 source:(250, 17) g_loss:1.561305 d_loss:0.000000 ppl:4.751001 lr:0.000238 | ||
step:34601 source:(250, 17) g_loss:1.551343 d_loss:0.000000 ppl:4.704056 lr:0.000238 | ||
step:34801 source:(250, 17) g_loss:1.553272 d_loss:0.000000 ppl:4.713269 lr:0.000237 | ||
step:35001 source:(250, 17) g_loss:1.520344 d_loss:0.000000 ppl:4.560683 lr:0.000236 | ||
epoch:125 test_bleu:55.420440435409546 template_bleu:38.278621435165405 test_loss:3.945925712585449 test_ppl:91.28235626220703 | ||
epoch:125 train_g_bleu:94.07399296760559 template_bleu:37.78722882270813 train_g_loss:1.5239386558532715 train_g_ppl:4.61153507232666 | ||
step:35201 source:(250, 17) g_loss:1.516096 d_loss:0.000000 ppl:4.541181 lr:0.000236 | ||
step:35401 source:(250, 17) g_loss:1.540540 d_loss:0.000000 ppl:4.653816 lr:0.000235 | ||
step:35601 source:(250, 17) g_loss:1.495307 d_loss:0.000000 ppl:4.447869 lr:0.000234 | ||
step:35801 source:(250, 17) g_loss:1.510601 d_loss:0.000000 ppl:4.516692 lr:0.000234 | ||
step:36001 source:(250, 17) g_loss:1.514494 d_loss:0.000000 ppl:4.534194 lr:0.000233 | ||
step:36201 source:(250, 17) g_loss:1.492944 d_loss:0.000000 ppl:4.437470 lr:0.000232 | ||
step:36401 source:(250, 17) g_loss:1.502883 d_loss:0.000000 ppl:4.481976 lr:0.000232 | ||
step:36601 source:(250, 17) g_loss:1.516687 d_loss:0.000000 ppl:4.544329 lr:0.000231 | ||
step:36801 source:(250, 17) g_loss:1.488550 d_loss:0.000000 ppl:4.418117 lr:0.000230 | ||
step:37001 source:(250, 17) g_loss:1.506315 d_loss:0.000000 ppl:4.497335 lr:0.000230 | ||
epoch:130 test_bleu:55.194246768951416 template_bleu:38.278621435165405 test_loss:3.90920352935791 test_ppl:74.25850677490234 | ||
epoch:130 train_g_bleu:95.7224190235138 template_bleu:37.78722882270813 train_g_loss:1.4806938171386719 train_g_ppl:4.4083943367004395 | ||
step:37201 source:(250, 17) g_loss:1.509537 d_loss:0.000000 ppl:4.511879 lr:0.000229 | ||
step:37401 source:(250, 17) g_loss:1.502420 d_loss:0.000000 ppl:4.479804 lr:0.000229 | ||
step:37601 source:(250, 17) g_loss:1.503928 d_loss:0.000000 ppl:4.486559 lr:0.000228 | ||
step:37801 source:(250, 17) g_loss:1.491947 d_loss:0.000000 ppl:4.433016 lr:0.000227 | ||
step:38001 source:(250, 17) g_loss:1.493441 d_loss:0.000000 ppl:4.439795 lr:0.000227 | ||
step:38201 source:(250, 17) g_loss:1.494921 d_loss:0.000000 ppl:4.446208 lr:0.000226 | ||
step:38401 source:(250, 17) g_loss:1.487812 d_loss:0.000000 ppl:4.414748 lr:0.000226 | ||
step:38601 source:(250, 17) g_loss:1.493444 d_loss:0.000000 ppl:4.439651 lr:0.000225 | ||
step:38801 source:(250, 17) g_loss:1.481276 d_loss:0.000000 ppl:4.385992 lr:0.000224 | ||
step:39001 source:(250, 17) g_loss:1.517557 d_loss:0.000000 ppl:4.548022 lr:0.000224 | ||
epoch:135 test_bleu:55.39852976799011 template_bleu:38.278621435165405 test_loss:3.942775011062622 test_ppl:84.02769470214844 | ||
epoch:135 train_g_bleu:93.65829825401306 template_bleu:37.78722882270813 train_g_loss:1.5224775075912476 train_g_ppl:4.621406555175781 | ||
step:39201 source:(250, 17) g_loss:1.490662 d_loss:0.000000 ppl:4.427538 lr:0.000223 | ||
step:39401 source:(250, 17) g_loss:1.493387 d_loss:0.000000 ppl:4.439529 lr:0.000223 | ||
step:39601 source:(250, 17) g_loss:1.475896 d_loss:0.000000 ppl:4.362800 lr:0.000222 | ||
step:39801 source:(250, 17) g_loss:1.492419 d_loss:0.000000 ppl:4.435294 lr:0.000222 | ||
step:40001 source:(250, 17) g_loss:1.500113 d_loss:0.000000 ppl:4.469399 lr:0.000221 | ||
step:40201 source:(250, 17) g_loss:1.490221 d_loss:0.000000 ppl:4.425404 lr:0.000220 | ||
step:40401 source:(250, 17) g_loss:1.489531 d_loss:0.000000 ppl:4.422766 lr:0.000220 | ||
step:40601 source:(250, 17) g_loss:1.466883 d_loss:0.000000 ppl:4.323574 lr:0.000219 | ||
step:40801 source:(250, 17) g_loss:1.477391 d_loss:0.000000 ppl:4.369344 lr:0.000219 | ||
step:41001 source:(250, 17) g_loss:1.471389 d_loss:0.000000 ppl:4.342805 lr:0.000218 | ||
step:41201 source:(250, 17) g_loss:1.488881 d_loss:0.000000 ppl:4.419534 lr:0.000218 | ||
epoch:140 test_bleu:55.49038648605347 template_bleu:38.278621435165405 test_loss:3.9374938011169434 test_ppl:72.12569427490234 | ||
epoch:140 train_g_bleu:94.95589137077332 template_bleu:37.78722882270813 train_g_loss:1.4875644445419312 train_g_ppl:4.452949047088623 | ||
step:41401 source:(250, 17) g_loss:1.470034 d_loss:0.000000 ppl:4.336848 lr:0.000217 | ||
step:41601 source:(250, 17) g_loss:1.469327 d_loss:0.000000 ppl:4.334231 lr:0.000217 | ||
step:41801 source:(250, 17) g_loss:1.481758 d_loss:0.000000 ppl:4.387790 lr:0.000216 | ||
step:42001 source:(250, 17) g_loss:1.471254 d_loss:0.000000 ppl:4.342154 lr:0.000216 | ||
step:42201 source:(250, 17) g_loss:1.466180 d_loss:0.000000 ppl:4.320355 lr:0.000215 | ||
step:42401 source:(250, 17) g_loss:1.471782 d_loss:0.000000 ppl:4.344584 lr:0.000215 | ||
step:42601 source:(250, 17) g_loss:1.466676 d_loss:0.000000 ppl:4.322161 lr:0.000214 | ||
step:42801 source:(250, 17) g_loss:1.473044 d_loss:0.000000 ppl:4.350119 lr:0.000214 | ||
step:43001 source:(250, 17) g_loss:1.463542 d_loss:0.000000 ppl:4.309168 lr:0.000213 | ||
step:43201 source:(250, 17) g_loss:1.465588 d_loss:0.000000 ppl:4.317548 lr:0.000213 | ||
epoch:145 test_bleu:54.751431941986084 template_bleu:38.278621435165405 test_loss:3.9903006553649902 test_ppl:74.64008331298828 | ||
epoch:145 train_g_bleu:97.59971499443054 template_bleu:37.78722882270813 train_g_loss:1.4349274635314941 train_g_ppl:4.207480430603027 | ||
step:43401 source:(250, 17) g_loss:1.467325 d_loss:0.000000 ppl:4.325232 lr:0.000212 | ||
step:43601 source:(250, 17) g_loss:1.478442 d_loss:0.000000 ppl:4.373446 lr:0.000212 | ||
step:43801 source:(250, 17) g_loss:1.474421 d_loss:0.000000 ppl:4.356277 lr:0.000211 | ||
step:44001 source:(250, 17) g_loss:1.467943 d_loss:0.000000 ppl:4.327763 lr:0.000211 | ||
step:44201 source:(250, 17) g_loss:1.467831 d_loss:0.000000 ppl:4.327116 lr:0.000210 | ||
step:44401 source:(250, 17) g_loss:1.460208 d_loss:0.000000 ppl:4.294605 lr:0.000210 | ||
step:44601 source:(250, 17) g_loss:1.475301 d_loss:0.000000 ppl:4.360077 lr:0.000209 | ||
step:44801 source:(250, 17) g_loss:1.466426 d_loss:0.000000 ppl:4.321571 lr:0.000209 | ||
step:45001 source:(250, 17) g_loss:1.491313 d_loss:0.000000 ppl:4.430133 lr:0.000208 | ||
epoch:149 test_bleu:55.32633662223816 template_bleu:38.278621435165405 test_loss:3.9137609004974365 test_ppl:70.1955337524414 | ||
epoch:149 train_g_bleu:95.74973583221436 template_bleu:37.78722882270813 train_g_loss:1.4678237438201904 train_g_ppl:4.358087062835693 | ||
step:45201 source:(250, 17) g_loss:1.474224 d_loss:0.000000 ppl:4.354743 lr:0.000208 | ||
step:45401 source:(250, 17) g_loss:1.462336 d_loss:0.000000 ppl:4.303623 lr:0.000207 | ||
/home/syp/miniconda2/envs/newtf/lib/python3.6/importlib/_bootstrap.py:205: RuntimeWarning: compiletime version 3.5 of module 'tensorflow.python.framework.fast_tensor_util' does not match runtime version 3.6 | ||
return f(*args, **kwds) |
Oops, something went wrong.