BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding - Crossminds
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
BERT is the first deeply bidirectional, unsupervised language representation model, pre-trained using only a plain text corpus. It has been widely used on various natural language processing tasks. This graph maps prior contextual representation models it builds on and notable pre-trained language models derived from BERT.
Other recommended papers
Yen-Chun Chen, Linjie Li, Licheng Yu, Ahmed El Kholy, Faisal Ahmed, Zhe Gan, Yu Cheng, Jingjing Liu
Weijie Su, Xizhou Zhu, Yue Cao, Bin Li, Lewei Lu, Furu Wei, Jifeng Dai
Liunian Harold Li, Mark Yatskar, Da Yin, Cho-Jui Hsieh, Kai-Wei Chang
Chen Sun, Fabien Baradel, Kevin Murphy, Cordelia Schmid
Chen Sun, Austin Myers, Carl Vondrick, Kevin Murphy, Cordelia Schmid
Andrew M. Dai, Quoc V. Le