DataLama / papers

๋…ผ๋ฌธ์„ ์ฝ์—ˆ์œผ๋ฉด, ์ •๋ฆฌ๋ฅผ ํ•˜์ž.
1 stars 0 forks source link

A Survey of Transformers (2021) #1

Open DataLama opened 2 years ago

DataLama commented 2 years ago

๐Ÿ“œ A Survey of Transformers

โšก ํ•œ์ค„์š”์•ฝ

2021๋…„ 6์›” ๊ธฐ์ค€์œผ๋กœ ์ •๋ฆฌํ•œ transformer ์•„ํ‚คํ…์ณ์— ๋Œ€ํ•œ ์„œ๋ฒ ์ด ๋…ผ๋ฌธ.

๐Ÿท๏ธ Abstract

Transformers have achieved great success in many artificial intelligence fields, such as natural language processing, computer vision, and audio processing. Therefore, it is natural to attract lots of interest from academic and industry researchers. Up to the present, a great variety of Transformer variants (a.k.a. X-formers) have been proposed, however, a systematic and comprehensive literature review on these Transformer variants is still missing. In this survey, we provide a comprehensive review of various X-formers. We first briefly introduce the vanilla Transformer and then propose a new taxonomy of X-formers. Next, we introduce the various X-formers from three perspectives: architectural modification, pre-training, and applications. Finally, we outline some potential directions for future research.

๐Ÿ—„๏ธ References

DataLama commented 2 years ago

1. Introduction

DataLama commented 2 years ago

2. Background

Vanila Transformer์— ๋Œ€ํ•œ ์„ค๋ช…. (All you need is attention 2017)

transformer architecture์— ๋Œ€ํ•œ ์„ค๋ช…

๋ชจ๋ธ ๋ณต์žก๋„

๋‹ค๋ฅธ ๋„คํŠธ์›Œํฌ์™€ ๋น„๊ต