Finetuned Language Models are Zero-Shot LearnersInstruction 데이터셋을 통해 fine-tuning을 진행하고 이를 통해 zero-shot 성능을 높이는 방법Mismatch between LM objective and human preferences 문제가 있는데 RLHF 에서 개선 Introducing FLAN: More generalizable Language Models with Instruction Fine-Tuninghttps://ai.googleblog.com/2021/10/introducing-flan-more-generalizable.html?m=1Instruction Tuning이란?LLM에 사용되는 Instruction tuning에 대해 알아보자https://velog.io/@nellcome/Instruction-Tuning이란Finetuned Language Models Are Zero-Shot LearnersThis paper explores a simple method for improving the zero-shot learning abilities of language models. We show that instruction tuning -- finetuning language models on a collection of tasks...https://arxiv.org/abs/2109.01652The Flan Collection: Designing Data and Methods for Effective...We study the design decisions of publicly available instruction tuning methods, and break down the development of Flan 2022 (Chung et al., 2022). Through careful ablation studies on the Flan...https://arxiv.org/abs/2301.13688Scaling Instruction-Finetuned Language ModelsFinetuning language models on a collection of datasets phrased as instructions has been shown to improve model performance and generalization to unseen tasks. In this paper we explore instruction...https://arxiv.org/abs/2210.11416