Optimizing and Testing Instruction-Following: Analyzing the Impact of Fine-Grained Instruction Variants on instruction-tuned LLMs
arxiv(2024)
Abstract
The effective alignment of Large Language Models (LLMs) with precise
instructions is essential for their application in diverse real-world
scenarios. Current methods focus on enhancing the diversity and complexity of
training and evaluation samples, yet they fall short in accurately assessing
LLMs' ability to follow similar instruction variants. We introduce an effective
data augmentation technique that decomposes complex instructions into simpler
sub-components, modifies these, and reconstructs them into new variants,
thereby preserves the original instruction's context and complexity while
introducing variability, which is critical for training and evaluating LLMs'
instruction-following precision. We developed the DeMoRecon dataset using this
method to both fine-tune and evaluate LLMs. Our findings show that LLMs
fine-tuned with DeMoRecon will gain significant performance boost on both ours
and commonly used instructions-following benchmarks.
MoreTranslated text
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined