We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
When I want to compute the perplexity of GPT2, how many epoches is suitable for training the GPT2 model?
When the model is trained for 400 epoches, the perplexity is about 35. Will the perplexity go down after taking more epoches?
Activity