🤖 AI学习必读100篇重要论文

按时间顺序整理 | 包含论文名称、作者、年份和下载链接

🏛️ 基础理论时期 (1950-1989)
#年份论文名称作者
11950Computing Machinery and IntelligenceAlan Turing
21956A Proposal for the Dartmouth AI Summer Research ProjectJohn McCarthy, et al.
31966MIT Speech Recognition system - ELIZAJoseph Weizenbaum
41969Perceptrons: An Introduction to Computational GeometryMarvin Minsky, Seymour Papert
51974A Theory of the LearnableLeslie Valiant
61986Learning Representations by Back-Propagating ErrorsDavid Rumelhart, Geoffrey Hinton, Ronald Williams
71989Backpropagation Applied to Handwritten Zip Code RecognitionYann LeCun, et al.
🔬 统计学习与早期深度学习 (1990-2011)
#年份论文名称作者
81995Learning in the Presence of Concept Drift and Hidden ContextsLeslie Valiant
91997Long Short-Term Memory (LSTM)Sepp Hochreiter, Jürgen Schmidhuber
101998Gradient-Based Learning Applied to Document RecognitionYann LeCun, Léon Bottou, Yoshua Bengio
111998Boosting: A Weak Learning AlgorithmYoav Freund, Robert Schapire
122001Random ForestsLeo Breiman
132002An Improved Boosting AlgorithmAlexander Grove, Dale Schuurmans
142003A Tutorial on Support Vector MachinesNello Cristianini, John Shawe-Taylor
152006A Fast Learning Algorithm for Deep Belief NetsGeoffrey Hinton, Simon Osindero, Yee-Whye Teh
162008Sparse Feature Learning for Deep Belief NetworksMarc'Aurelio Ranzato, et al.
172010Variational Learning for DigitsA. Mnih, K. Kavukcuoglu
182011Neural Networks for NLPRichard Socher, et al.
🚀 深度学习爆发期 (2012-2016)
#年份论文名称作者
192012ImageNet Classification with Deep Convolutional Neural Networks (AlexNet)Alex Krizhevsky, Ilya Sutskever, Geoffrey Hinton
202013Efficient Estimation of Word Representations in Vector Space (Word2Vec)Tomas Mikolov, et al.
212013Playing Atari with Deep Reinforcement LearningVolodymyr Mnih, et al.
222014Generative Adversarial Networks (GAN)Ian Goodfellow, et al.
232014Sequence to Sequence Learning with Neural NetworksIlya Sutskever, Oriol Vinyals, Quoc V. Le
242014Neural Machine Translation by Jointly Learning to Align and TranslateDzmitry Bahdanau, Kyunghyun Cho, Yoshua Bengio
252014Dropout: A Simple Way to Prevent Neural Networks from OverfittingNitish Srivastava, et al.
262015Batch Normalization: Accelerating Deep Network TrainingSergey Ioffe, Christian Szegedy
272015Deep Residual Learning for Image Recognition (ResNet)Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun
282015Very Deep Convolutional Networks for Large-Scale Image Recognition (VGGNet)Karen Simonyan, Andrew Zisserman
292015Spatial Transformer NetworksMax Jaderberg, et al.
302016Mastering the Game of Go with Deep Neural Networks and Tree Search (AlphaGo)David Silver, Aja Huang, et al.
312016FastText: Bag of Tricks for Efficient Text ClassificationArmand Joulin, et al.
322016WaveNet: A Generative Model for Raw AudioAaron van den Oord, et al.
⚡ Transformer与预训练模型时代 (2017-2020)
#年份论文名称作者
332017Attention Is All You Need (Transformer)Ashish Vaswani, Noam Shazeer, et al.
342017Neural Machine Translation with Latent AlignmentDzmitry Bahdanau, et al.
352017Fast and Accurate Reading Comprehension by Neural NetworkRomain Gloannec, et al.
362018BERT: Pre-training of Deep Bidirectional Transformers for Language UnderstandingJacob Devlin, Ming-Wei Chang, et al.
372018Improving Language Understanding by Generative Pre-Training (GPT)Alec Radford, et al.
382018GLUE: A Multi-Task Benchmark and Evaluation for NLPAlex Wang, et al.
392019XLNet: Generalized Autoregressive Pretraining for Language UnderstandingZhilin Yang, et al.
402019Visualizing and Measuring the Geometry of BERTIan Tenney, et al.
412019RoBERTa: A Robustly Optimized BERT Pretraining ApproachYinhan Liu, et al.
422019ControlNet: Conditional Neural NetworksLvmin Zhang, et al.
432020Language Models are Few-Shot Learners (GPT-3)Tom Brown, et al.
442020Image GPTXi Chen, et al.
452020YOLOv4: Optimal Speed and Accuracy of Object DetectionAlexey Bochkovskiy, et al.
462020Bootstrap Your Own Latent (BYOL)Jean-Bastien Grill, et al.
472020SimCLRv2: Big Self-Supervised Models are Strong Semi-Supervised LearnersTing Chen, et al.
482020An Image is Worth 16x16 Words: Transformers for Image Recognition (ViT)Alexey Dosovitskiy, et al.
492020Exploring the Limits of Transfer Learning with T5Colin Raffel, et al.
502020Denoising Diffusion Probabilistic ModelsJonathan Ho, Ajay Jain, Pieter Abbeel
512020Score-Based Generative Modeling through Stochastic Differential EquationsYang Song, et al.
🤖 大模型与生成式AI时代 (2021-2022)
#年份论文名称作者
522021Highly Accurate Protein Structure Prediction with AlphaFold2John Jumper, Richard Evans, et al.
532021Learning Transferable Visual Models From Natural Language Supervision (CLIP)Alec Radford, et al.
542021Swin Transformer: Hierarchical Vision Transformer using Shifted WindowsZe Liu, et al.
552021Emerging Properties in Self-Supervised Vision Transformers (DINO)Mathilde Caron, et al.
562021DALL-E: Zero-Shot Text-to-Image GenerationAditya Ramesh, et al.
572021Chain-of-Thought Prompting Elicits Reasoning in Large Language ModelsJason Wei, et al.
582021Noised CLIP: CLIP with Positive-Negative Contrastive LearningDieuwke Hupkes, et al.
592021LLaMA: Open and Efficient Foundation Language ModelsHugo Touvron, et al.
602021OPT: Open Pre-trained Transformer Language ModelsSusan Zhang, et al.
612021Training Compute-Optimal Large Language Models (Chinchilla)Jordan Hoffmann, et al.
622022Self-consistency Improves Chain of Thought ReasoningXuezhi Wang, et al.
632022PaLM: Scaling Language Modeling with PathwaysAakanksha Chowdhery, et al.
642022Hierarchical Text-Conditional Image Generation with CLIP Latents (DALL-E 2)Aditya Ramesh, et al.
652022Make-An-AnimationYuan Amou, et al.
662022Emergent Abilities of Large Language ModelsJason Wei, et al.
672022A ConvNet for the 2020s (ConvNeXt)Zhuang Liu, et al.
682022DeiT III: Revenge of the ViTHugo Touvron, et al.
692022GiT: Graph Transformer for Graph ClassificationJiaxuan You, et al.
702022RLHF: Training Language Models to Follow Instructions with Human FeedbackLong Ouyang, et al.
712022Stable Diffusion: High-Resolution Image Synthesis with Latent Diffusion ModelsRobin Rombach, et al.
🌟 AGI与大模型突破期 (2023-2024)
#年份论文名称作者
722023GPT-4 Technical ReportOpenAI
732023LLaMA 2: Open Foundation and Fine-Tuned Chat ModelsHugo Touvron, et al.
742023LLaMA-Adapter V2: Large Language Models with Visual Instruction TuningPeng Gao, et al.
752023QLoRA: Efficient Finetuning of Quantized LLMsTim Dettmers, et al.
762023FlashAttention: Fast and Memory-Efficient Exact Attention with IO-AwarenessTri Dao, et al.
772023Toolformer: Language Models Can Teach Themselves to Use ToolsTimo Schick, et al.
782023Segment Anything Model (SAM)Alexander Kirillov, et al.
792023Llama 3 (Meta AI)Meta AI
802023Mistral 7BAlbert Jiang, et al.
812023Mixtral of ExpertsAlbert Jiang, et al.
822023MiniGPT-4: Enhancing Vision Language UnderstandingDeyao Zhu, et al.
832023LLaVA: Visual Instruction TuningHaotian Liu, et al.
842023Multimodal Large Language Models: A SurveyYongth P, et al.
852024DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language ModelDeepSeek Team
862024Qwen2 Technical ReportAlibaba
872024DINOv2: Self-Supervised Vision TransformersMaxime Oquab, et al.
882024Sora: Video Generation from TextOpenAI
892024Llama 3.1 405BMeta AI
902024The Llama 3 Herd of ModelsMeta AI
912024Gemini 1.5: Unlocking Multimodal Long-Context UnderstandingGoogle DeepMind
922024DeepSeek-R1: Incentivizing Reasoning Capability in LLMsDeepSeek Team
932024o1: OpenAI Reasoning ModelOpenAI
942024ReFT: Reasoning with Fine-Tuned Language ModelsStanford NLP
🔮 前沿研究方向 (2024-2025)
#年份论文名称作者
952024Agent Q:引导LLM进行自我改进Meta AI
962024Claude 3.5: Computer UseAnthropic
972024World Models: Generative AI for 3D ScenesKabra N, et al.
982025DeepSeek-R2DeepSeek Team
992025Gemini 2.0Google DeepMind
1002025o3-mini: OpenAI Reasoning ModelOpenAI
说明:
• 本列表按时间顺序排列,从1950年图灵测试到2025年最新研究
• 大部分论文可通过 arXiv 免费下载,部分链接可能需要科学上网
• 建议按时代分组学习:基础理论 → 深度学习 → Transformer → 大模型
• 更新日期:2025年3月