Paper – LIMA

Large language models are trained in two stages: (1) unsupervised pretraining from raw text, to learn general-purpose representations, and (2) large-scale instruction tuning and reinforcement learning, to better align to end tasks and user preferences. The relative importance of these two stages is measured by training LIMA, a 65B parameter LLaMa language model fine-tuned with … Continue reading Paper – LIMA