GPT-3 links

and some renewed interest in Transformers


HN comments where iI found the link

some intuition on relation between Graph and Transformer architecture

I should find a good intro to transformer, it seems that they scale the right way, even bigger models with ever better performances, like size drives performance.

UPDATE 24 may 2021: Are we in an AI overhang?


Leave a Reply

Fill in your details below or click an icon to log in: Logo

You are commenting using your account. Log Out /  Change )

Facebook photo

You are commenting using your Facebook account. Log Out /  Change )

Connecting to %s