(Q95726734)

English

GPT-3

2020 transformer-based language model

  • Generative Pre-trained Transformer 3
  • Generative Pretrained Transformer 3
  • GPT3

Statements

0 references
28 May 2020
0 references
0 references
0 references
GPT-3.5 (multiple languages)
0 references
0 references
0 references
125M
125,000,000 parameter
350M
350,000,000 parameter
760M
760,000,000 parameter
1.3B
1,300,000,000 parameter
2.7B
2,700,000,000 parameter
6.7B
6,700,000,000 parameter
13B
13,000,000,000 parameter
175B
175,000,000,000 parameter
1 reference
To study the dependence of ML performance on model size, we train 8 different sizes of model, ranging over three orders of magnitude from 125 million parameters to 175 billion parameters, with the last being the model we call GPT-3. (English)
22 July 2020
0 references
0 references
0 references
408,995
0 references

Identifiers

0 references
0 references
0 references
Generative Pretrained Transformers
17 July 2020
23 May 2023
0 references
 
edit
edit
    edit
      edit
        edit
          edit
            edit
              edit
                edit