missing @kbin.social
Posts 0
Comments 1
Retentive Network: A Successor to Transformer for Large Language Models 1 0
If the claims here are true.. wow research and development are moving very quickly
Reply
Retentive Network: A Successor to Transformer for Large Language Models 1 0
How does this compare with RWKV performance and architecture wise? Admittingly I am still reading the paper.
Reply
Next