For RNN text generation, the main goal is to produce text that looks natural and meaningful. We often use perplexity to measure this. Perplexity tells us how well the model predicts the next word. A lower perplexity means the model is better at guessing the next word, so the generated text is more fluent.
Sometimes, we also check BLEU score if we have reference texts to compare. BLEU measures how similar the generated text is to real examples. But perplexity is the most common because it works even without exact references.