Sparrow

tags
Transformers, GPT, Chinchilla
paper
(Glaese et al. 2022)
blog post
Deepmind announcement blog post

Architecture

Starts from the Chinchilla 70B model but adds RLHF (Reinforcement Learning with Human Feedback). It also adds inline evidence like GopherCite.

Parameter count

70B

Bibliography

  1. . . "Improving Alignment of Dialogue Agents via Targeted Human Judgements". arXiv. DOI.
Last changed | authored by

Comments


← Back to Notes