计算机视觉领域的经典论文
Survey
- A Comprehensive Survey on Source-free Domain Adaptation
- Domain Generalization in Computational Pathology: Survey and Guidelines
- A Survey on Generative Modeling with Limited Data, Few Shots, and Zero Shot
- Know Your Self-supervised Learning: A Survey on Image-based Generative and Discriminative Training
- On the Design Fundamentals of Diffusion Models: A Survey
Paper
图像分类
- AlexNet_ImageNet Classification with Deep Convolutional
- VGGNet_Very Deep Convolutional Networks for Large-Scale Image Recognition
- ResNet_Deep Residual Learning for Image Recognition
- DenseNet_Densely Connected Convolutional Networks
- ViT_An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale
目标检测
- R-CNN_Rich feature hierarchies for accurate object detection and semantic segmentation
- Fast R-CNN
- Mask R-CNN
- YOLO_You Only Look Once: Unified, Real-Time Object Detection
- DETR_End-to-End Object Detection with Transformers
语义分割
- BiSeNet_Bilateral Segmentation Network for Real-time Semantic Segmentation
- FCN_Fully Convolutional Networks for Semantic Segmentation
- OCRNet_Object-Contextual Representations for Semantic Segmentation
- U-Net_Convolutional Networks for Biomedical Image Segmentation
- Swin Transformer_Hierarchical Vision Transformer using Shifted Windows
生成模型
自然语言处理领域的经典论文
Survey
- Augmented Language Models: a Survey
- Model-tuning Via Prompts Makes NLP Models Adversarially Robust
- A Survey on In-context Learning
Paper
传统自然语言处理
- Word2Vec_Efficient Estimation of Word Representations in Vector Space
- CNN_Convolutional Neural Networks for Sentence Classification
- RNN_A Critical Review of Recurrent Neural Networks for Sequence Learning
- Seq2Seq_Sequence to Sequence Learning with Neural Networks
- Convolutional Sequence to Sequence Learning
- GloVe_Global Vectors for Word Representation
大模型(LLM)
- Transformer_Attention Is All You Need
- GPT_Improving Language Understanding by Generative Pre-Training
- GPT2_Language Models are Unsupervised Multitask Learners
- [GPT3_Language Models are Few-Shot Learners](2005.14165 (arxiv.org))
- [GPT3.5_Training language models to follow instructions with human feedback](2203.02155 (arxiv.org))
- [GPT-4 Technical Report](2303.08774 (arxiv.org))
- BERT_Pre-training of Deep Bidirectional Transformers for Language Understanding
大模型微调
- LORA: LOW-RANK ADAPTATION OF LARGE LANGUAGE MODEL
- [The Power of Scale for Parameter-Efficient Prompt Tuning]([2104.08691] The Power of Scale for Parameter-Efficient Prompt Tuning (arxiv.org))
- [Chain-of-Thought Prompting Elicits Reasoning in Large Language Models](2201.11903 (arxiv.org))
多模态(Multi-model Language Modal)
- PaLM-E: An Embodied Multimodal Language Model
- [Visual Instruction Tuning](2304.08485 (arxiv.org))
- [TALLRec: An Effective and Efficient Tuning Framework to Align Large Language Model with Recommendation](TALLRec: An Effective and Efficient Tuning Framework to Align Large Language Model with Recommendation (arxiv.org))
- LLaRA: Aligning Large Language Models with Sequential Recommenders
推荐系统领域的经典论文
Survey
- A Survey on User Behavior Modeling in Recommender Systems
- Disentangled Representation Learning
- A Cookbook of Self-Supervised Learning
- Self-Supervised Learning for Recommender Systems: A Survey
- Graph Neural Networks in Recommender Systems: A Survey
Paper
- Bayesian Personalized Ranking
- Neural Collaborative Filtering
- Neural Graph Collaborative Filtering
- LightGCN_Simplifying and Powering Graph Convolution Network for Recommendation
- Self-Attentive Sequential Recommendation
- Intent-aware Ranking Ensemble for Personalized Recommendation
- LightGT_A Light Graph Transformer for Multimedia Recommendation
- Graph Transformer for Recommendation
- Learning Disentangled Representations for Recommendation
- Deep Interest Network for Click-Through Rate Prediction
- Less is More: Reweighting Important Spectral Graph Features for Recommendation
- Search-based User Interest Modeling with Lifelong Sequential Behavior Data for Click-Through Rate Prediction
- Multi-behavior Self-supervised Learning for Recommendation
- Multi-Scenario Ranking with Adaptive Feature Learning
- Towards Multi-Interest Pre-training with Sparse Capsule Network