How To Scale Transformers’ Memory up to 262K Tokens With a Minor Change?

Extending Transformers by memorizing up to 262K tokens

Reza Yazdanfar
Towards AI
Published in
6 min readMar 13, 2023

--

I am building nouswise🤗!! check it out and get on the waitlist😉 If you want early access hit me up…

--

--