LLM / VLM paper list

2025. 2. 11. 17:49paper

LLM

ChatGPT Papers

  • GPT-1 : Improving Language Understanding by Generative Pre-Training. - Alec Radford, Karthik Narasimhan, Tim Salimans, Ilya Sutskever [pdf] 2018.6
  • GPT-2 : Language Models are Unsupervised Multitask Learners. - Alec Radford, Jeff Wu, Rewon Child, D. Luan, Dario Amodei, Ilya Sutskeve [pdf] 2019.2
  • GPT-3 : Language Models are Few-Shot Learners. - Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, Dario Amodei [pdf] 2020.5
  • GPT-4 : GPT-4 Technical Report. OpenAI [pdf] 2023.3
  • InstructGPT :Training language models to follow instructions with human feedback. - Long Ouyang, Jeff Wu, Xu Jiang, Diogo Almeida, Carroll L. Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, John Schulman, Jacob Hilton, Fraser Kelton, Luke Miller, Maddie Simens, Amanda Askell, Peter Welinder, Paul Christiano, Jan Leike, Ryan Lowe [pdf] 2022.3
  • RLHF - Augmenting Reinforcement Learning with Human Feedback.
  • W. Bradley Knox, Peter Stone [pdf] 2011.7

 

LLaMA papers

  • LLaMA: Open and Efficient Foundation Language Models. - Hugo Touvron, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Timothée Lacroix, Baptiste Rozière, Naman Goyal, Eric Hambro, Faisal Azhar, Aurelien Rodriguez, Armand Joulin, Edouard Grave, Guillaume Lample*. arxiv 2023. paper. code
  • Llama 2: Open Foundation and Fine-Tuned Chat Models.- Hugo Touvron, Louis Martin, Kevin Stone et al.* Meta AI 2023. paper. code
  • The Llama 3 Herd of Models. - Abhimanyu Dubey, Abhinav Jauhri, Abhinav Pandey et al.* arxiv 2024. paper. code

 

Deepseek papers

 

VLM

LLaVA paper

  • LLaVA - Visual Instruction Tuning (NeurIPS 2023, Oral) - Haotian Liu*, Chunyuan Li*, Qingyang Wu, Yong Jae Lee[Paper] [HF]
  • LLaVA-Next - Improved Baselines with Visual Instruction Tuning (CVPR 2024) - Haotian Liu, Chunyuan Li, Yuheng Li, Yong Jae Lee [Paper] [HF]
  • LLaVA - one vision .. 
  • LLaVA-NeXT-Interleave: Tackling Multi-image, Video, and 3D in Large Multimodal Models

 

LoRA