Our paper is available here.
-
demo/ contains demo code, demonstrating how to run the full ParaGuide approach to transform texts from formal → informal.
-
training/ contains the logic for training a paraphrase-conditioned text diffusion model.
-
inference/ contains our code for generating inferences with ParaGuide (once you have preprocessed/paraphrased your data)
-
data/ contains the logic for generating reddit and enron (paraphrase, original text) data.
-
baselines/ contains our implementations of each baseline.
-
evaluations contains our automatic evaluation automatic and human eval data and code.
In your python environment (>=3.8), you can install dependencies via the requirements file:
pip install -r requirements.txt
Our models and data are available for download here.
We also provided corresponding scripts:
- Models: models/download.sh
- Data: data/enron/download_training_dataset.sh
We recommend first checking out demo/generate_examples.py, which demonstrates ParaGuide inference logic!