- tags
- Transformers, GPT, NLP
- blog post
- AI21Labs blog
Architecture
This model is similar to GPT-3 with an improved tokenizer that increases the learning efficiency. It also has more parameters.
Parameter count
178B
This model is similar to GPT-3 with an improved tokenizer that increases the learning efficiency. It also has more parameters.
178B