Transformer-XL
Selective Attention
Large transformer networks work wonders with natural language, but they require enormous amounts of computation. New research slashes processor cycles without compromising performance.
2 Posts
Stay updated with weekly AI News and Insights delivered to your inbox