10 Facts About GPT-2

1.

GPT-2 was created as a "direct scale-up" of OpenAI's 2018 GPT model, with a ten-fold increase in both its parameter count and the size of its training dataset.

FactSnippet No. 683,341
2.

GPT-2 was to be followed by the 175-billion-parameter GPT-3, revealed to the public in 2020 .

FactSnippet No. 683,342
3.

GPT-2 was created as a direct scale-up of GPT, with both its parameter count and dataset size increased by a factor of 10.

FactSnippet No. 683,343
4.

Notably, GPT-2 was evaluated on its performance on tasks in a zero-shot setting.

FactSnippet No. 683,344
5.

GPT-2's flexibility was described as "impressive" by The Verge; specifically, its ability to translate text between languages, summarize long articles, and answer trivia questions were noted.

FactSnippet No. 683,345

Related searches

GPT-3 OpenAI RAM
6.

One commonly-cited justification was that, since generated text was usually completely novel, it could be used by spammers to evade automated filters; OpenAI demonstrated a version of GPT-2 fine-tuned to "generate infinite positive – or negative – reviews of products".

FactSnippet No. 683,346
7.

GPT-2 deployment is resource-intensive; the full version of the model is larger than five gigabytes, making it difficult to embed locally into applications, and consumes large amounts of RAM.

FactSnippet No. 683,347
8.

Possible applications of GPT-2 described by journalists included aiding humans in writing text like news articles.

FactSnippet No. 683,348
9.

In 2019, AI Dungeon was launched, which used GPT-2 to generate dynamic text adventures based on user input.

FactSnippet No. 683,349
10.

In February 2021, a crisis center for troubled teens announced that they would begin using a GPT-2-derived chatbot to help train counselors by allowing them to have conversations with simulated teens .

FactSnippet No. 683,350