AlexaTM - 20 Billion Parameter Model With Impressive Performance

post by MrThink (ViktorThink) · 2022-09-09T21:46:30.151Z · LW · GW · 0 comments

Amazon trained a seq2seq model that outperforms GPT-3 on SuperGLUE, SQuADv2, and is not (that) far behind PaLM with 540 billion parameters.

Article: https://www.amazon.science/publications/alexatm-20b-few-shot-learning-using-a-large-scale-multilingual-seq2seq-model

Benchmarks:

0 comments

Comments sorted by top scores.