Ok since I last posted in this thread I have been playing with GPT-2 and testing it's capability.
In short, it's shit. I don't know what all the fuss is about.
I have better LSTM Text Generation models that I have already been working with that surpass the limited GPT-2 117M Model.
What a let down after all the hype.
