Transformer Network 기반 모델 구조들
BERT 모든 사전 학습 transformer 기반 모델의 어머니 - BERT, Devlin, J.et al., "Bert: Pre-training of deep bidirectional transformers for language understanding.", 2018. General purpose - T5, Raffel, Colin, et al., "Exploring the limits of transfer learning with a unified text-to-text transformer.", 2019. - GPT, Radford, Alec, et al., "Improving language understanding by generative pre-training.", 2018. - GPT-2..