Deep Learning

1.Attention is All You Need

post-thumbnail

2.An Image Is Worth 16x16 Words : Transformers for Image Recognition at Scale

post-thumbnail

3.Swin Transformer: hierarchical Vision Transformer using Shifted Windows

post-thumbnail

4.FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness

post-thumbnail