Skip to content

Commit

Permalink
remove more spurious perplexity references
Browse files Browse the repository at this point in the history
  • Loading branch information
soumith authored Dec 30, 2018
1 parent 97e3e13 commit e0929a4
Showing 1 changed file with 3 additions and 3 deletions.
6 changes: 3 additions & 3 deletions word_language_model/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,9 +5,9 @@ By default, the training script uses the Wikitext-2 dataset, provided.
The trained model can then be used by the generate script to generate new text.

```bash
python main.py --cuda --epochs 6 # Train a LSTM on Wikitext-2 with CUDA, reaching perplexity of 117.61
python main.py --cuda --epochs 6 --tied # Train a tied LSTM on Wikitext-2 with CUDA, reaching perplexity of 110.44
python main.py --cuda --tied # Train a tied LSTM on Wikitext-2 with CUDA for 40 epochs, reaching perplexity of 87.17
python main.py --cuda --epochs 6 # Train a LSTM on Wikitext-2 with CUDA
python main.py --cuda --epochs 6 --tied # Train a tied LSTM on Wikitext-2 with CUDA
python main.py --cuda --tied # Train a tied LSTM on Wikitext-2 with CUDA for 40 epochs
python generate.py # Generate samples from the trained LSTM model.
```

Expand Down

0 comments on commit e0929a4

Please sign in to comment.