Skip to content

deynabaevax/nlp-text-generation

Repository files navigation

Text Generation with Natural Language Processing

Introduction

A language model can predict the probability of the next word in the sequence, based on the words already observed in the sequence. Neural network models are a preferred method for developing statistical language models because they can use a distributed representation, where different words with similar meanings have similar representation and because they can use a large context of recently observed words when making predictions.

Goals

1. Prepare text for developing a word-based language model.
2. Design and fit a neural language model with a learned embedding and an LSTM hidden layer.
3. Use the learned language model to generate new text with similar statistical properties as the source text.

The Dataset

The Republic is the classical Greek philosopher Plato's most famous work. It is structured as a dialog on the topic of order and justice within a city state. I got the file from the Project Gutenberg's website. Link to the dataset.

Steps

1. Getting the Data
2. Data Preparation
3. Encode the Sequences
4. Sequence Inputs and Output
5. Train the Language Model
6. Use the Language Model

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published