<aside>
<aside>
<aside>
🔥 Paper 관계형에 연결된 논문을 확인하고, ReadingLog 관계형에 각자 할당된 논문을 정리하면 됩니다. **** (다양한 리뷰자료들을 활용해 이해해도 되지만, 최종적으로는 논문을 꼭 확인해보세요.)
</aside>
[Week 0] Preliminaries - Word Embedding [Week 0] Preliminaries - Attention and Transformer [Week 1] Introduction - Overview and Survey Papers [Week 2] Generative/Predictive Self-Supervised Learning [Week 3] Contrastive (Self-)Supervised Learning [Week 4] Self-Distilled Self-Supervised Learning [Week 5] Word/Sentence/Document-level SSL for Text Encoders [Week 6] Node/Graph-level SSL for Graph Encoders [Week 7] Image-Text Alignment for Generation based on Single-Stream Fusion [Week 8] Image-Text Alignment for Retrieval/Generation based on Dual-Stream Fusion [Week 9] Large-scale Multimodal Alignment based on Multi-Stream Fusion [Week 10] Adapter-based Multimodal Models (toward MLLMs) [Week 11] Multimodal Large Language Models [Week 12] Multimodal Recommendation [Week 12+] Multimodal Recommendation based on Graph Learning
[Week 1] Introduction - Overview and Survey Papers [2023][MIR][-] VLP - A Survey on Vision-Language Pre-training (295) [2022][CVPR][METER] An Empirical Study of Training End-to-End Vision-and-Language Transformers (470) # dual-stream —————————— [2024][ACL-F][-] The Revolution of Multimodal Large Language Models - A Survey (113) [2024][TPAMI][-] A Survey on Multimodal Large Language Models (1984) [2024][ACL-F][-] MM-LLMs - Recent Advances in MultiModal Large Language Models (424) —————————— [2023][arXiv][-] A comprehensive survey on multimodal recommender systems - Taxonomy, evaluation, and future directions (109) [2023][TRS][-] A survey of graph neural networks for recommender systems - Challenges, methods, and directions (635) [2023][TIS][-] Contrastive Self-supervised Learning in Recommender Systems - A Survey (85) [2024][CSUR][-] Multimodal Recommender Systems - A Survey (123) [2024][TKDE][-] Self-Supervised Learning for Recommender Systems - A Survey (449)
[Week 2] Generative/Predictive Self-Supervised Learning [2018][OpenAI][GPT-1] Improving Language Understanding by Generative Pre-Training (15473) # decoder [2019][NAACL][BERT] Pre-training of Deep Bidirectional Transformers for Language Understanding (142678) # encoder, masked language modeling [2019][OpenAI][GPT-2] Language Models are Unsupervised Multitask Learners (18051) # decoder, autoregressive modeling [2020][JMLR][T5] Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer (26079) # encoder-decoder [2020][NIPS][GPT-3] Language Models are Few-Shot Learners (52889) # decoder [2023][OpenAI][GPT-4] Technical Report (5168) # decoder —————————— [2020][ICML][iGPT] Generative Pretraining from Pixels (2103) [2021][CVPR][MAE] Masked Autoencoders Are Scalable Vision Learners (11219) [2022][ICLR][BEiT] BERT Pre-Training of Image Transformers (3779) # Microsoft
[Week 3] Contrastive (Self-)Supervised Learning [2015][CVPR][FaceNet] A Unified Embedding for Face Recognition and Clustering (19437) # supervised, triplet loss [2015][ICLR][triplet network] Deep Metric Learning with Triplet Network (2888) # supervised, triplet loss [2016][CVPR][-] Deep Metric Learning via Lifted Structured Feature Embedding (2195) # hard negative, supervised [2020][NIPS][SupCon] Supervised Contrastive Learning (6708) # Google, hard negative, supervised —————————— ****[2018][arXiv][CPC] Representation Learning with Contrastive Predictive Coding (12471) # self-supervised [2018][CVPR][NPID] ****Unsupervised Feature Learning via Non-Parametric Instance Discrimination (4924) # self-supervised, instance discrimination [2020][ICML][SimCLR] A Simple Framework for Contrastive Learning of Visual Representations (26029) # Google, self-supervised [2020][CVPR][MoCo] Momentum Contrast for Unsupervised Visual Representation Learning (16840) # self-supervised, memory bank [2020][NIPS][SwAV] Unsupervised Learning of Visual Features by Contrasting Cluster Assignments (5141) # Facebook, clustering, self-supervised [2020][NIPS][SimCLRv2] Big Self-Supervised Models are Strong Semi-Supervised Learners (2916) [2020][arXiv][MoCo-2] Improved Baselines with Momentum Contrastive Learning (4272) [2021][ICCV][MoCo-3] An Empirical Study of Training Self-Supervised Vision Transformers (2430) # Transformer
[Week 4] Self-Distilled Self-Supervised Learning [2020][NIPS][BYOL] Bootstrap Your Own Latent - A New Approach to Self-Supervised Learning (8971) # DeepMind [2021][CVPR][SimSiam] Exploring Simple Siamese Representation Learning (5623) # Meta, BYOL 구조 단순화 [2021][ICCV][DINO] Emerging Properties in Self-Supervised Vision Transformers (8179) # Meta [2022][ICLR][iBOT] Image BERT Pre-Training with Online Tokenizer (1154) # Microsoft [2024][TMLR][DINOv2] Learning Robust Visual Features without Supervision (4465) # Meta [2021][ACL][BSL] Bootstrapped Unsupervised Sentence Representation Learning (43)
[Week 5] Word/Sentence/Document-level SSL for Text Encoders **[2019][EMNLP][SBERT] Sentence Embeddings using Siamese BERT-Networks (18399) # supervised [2018][ACL][USE] Universal Sentence Encoder (1788) # Google, supervised —————————— [2021][EMNLP-F][TSDAE] Using Transformer-based Sequential Denoising Auto-Encoder for Unsupervised Sentence Embedding Learning (274) # Generative/Predictive —————————— [2021][ACL][DeCLUTR] Deep Contrastive Learning for Unsupervised Textual Representations (603) # contrastive, unsupervised [2021][ACL][ConSERT] A Simple Contrastive Framework for Self-Supervised Sentence Representation Learning (679) # contrastive, unsupervised [2021][EMNLP][SimCSE] Simple Contrastive Learning of Sentence Embeddings (4193) # contrastive, (un)supervised [2022][EMNLP][PromptBERT] Improving BERT Sentence Embeddings with Prompts (224) # contrastive, supervised / weakly-supervised [2022][ACL-F][ST5] Scalable Sentence Encoders from Pre-trained Text-to-Text Models (660) # contrastive, Google, supervised [2022][TMLR][Contriever] Unsupervised Dense Information Retrieval with Contrastive Learning (1017) # ⭐, contrastive, Meta, unsupervised [2022][arXiv][E5] Text Embeddings by Weakly-Supervised Contrastive Pre-training (724) # ⭐, Microsoft, weakly-supervised [2023][ACL-F][ReContriever] Unsupervised Dense Retrieval with Relevance-Aware Contrastive Pre-Training (39) —————————— [2020][NIPS][MiniLM] Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers (1700) # 범용 언어모델, Microsoft, knowledge distillation [2024][ACL-F][BGE-M3] Multi-lingual, Multi-functionality, Multi-granularity Text Embeddings through Self-Knowledge Distillation (659) # ⭐, contrastive, knowledge distillation
[Week 6] Node/Graph-level SSL for Graph Encoders [2019][ICLR][DGI] Deep Graph Infomax (3530) # node-level, MIM, contrastive [2020][KDD][GCC] Graph Contrastive Coding for Graph Neural Network Pre-Training (1064) # node-level, graph-level, contrastive [2020][ICLR][InfoGraph] Unsupervised and Semi-supervised Graph-Level Representation Learning via Mutual Information Maximization (1334) # graph-level, MIM, contrastive [2020][KDD][GPT-GNN] Generative Pre-Training of Graph Neural Networks (661) # node-level, generative [2020][ICML][-] Contrastive Multi-View Representation Learning on Graphs (1870) # node-level, graph-level, contrastive [2022][KDD][GraphMAE] Self-Supervised Masked Graph Autoencoders (640) # node-level, graph-level, generative [2021][NIPS][InfoGCL] Information-Aware Graph Contrastive Learning (230) # node-level, graph-level, contrastive [2021][SIGIR][SGL] Self-supervised Graph Learning for Recommendation (1110) # contrastive, node-level