⭐ 我的收藏
管理收藏的研究素材,快速查阅和复用
Attention Is All You Need
Vaswani, A. et al.NeurIPS2017
The dominant sequence transduction models are based on complex recurrent or convolutional neural networks...
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Devlin, J. et al.NAACL2019
We introduce a new language representation model called BERT, designed to pre-train deep bidirectional representations...
GPT-4 Technical Report
OpenAIarXiv2023
We report the development of GPT-4, a large-scale multimodal model that exhibits human-level performance on many benchmarks...
A Survey on Large Language Models: Applications and Challenges
Zhao, W.X. et al.ACM Computing Surveys2023
This paper presents a comprehensive survey on large language models, covering their architecture, training methods, and applications...