반응형 Sliding Attention Window1 Taming Transformers for High Resolution Image Synthesis (VQGAN) https://arxiv.org/abs/2012.09841 Taming Transformers for High-Resolution Image Synthesis Designed to learn long-range interactions on sequential data, transformers continue to show state-of-the-art results on a wide variety of tasks. In contrast to CNNs, they contain no inductive bias that prioritizes local interactions. This makes them expres arxiv.org ## Background 본 논문에 들어가기 앞서 Background 지식을.. 2022. 7. 28. 이전 1 다음 반응형