Open AI, a project founded with the support of Elon Musk, is able to generate news stories from a headline or first line of text.
In February, the firm released a limited version of its software for other developers to use, to explore its potential.
The firm, which Musk is no longer involved in, has since launched an updated version of the software with half of the power of the full AI.
Now, computer science master's students Aaron Gokaslan and Vanya Cohen from Brown University have shared code for what they say is the full version.
Built by Adam King (@AdamDanielKing) as an easier way to play with OpenAI's new machine learning model. In February, OpenAI unveiled a language model called GPT-2 that generates coherent paragraphs of text one word at a time.
For now OpenAI has decided only to release three smaller versions of it which aren't as coherent but still produce interesting results. This site runs the largest released model, 774M, which is half the size of the full model.