As instruction-tuned large language models (LLMs) evolve, aligning pretrained
foundation models presents increasing challenges. Existing alignment
strategies, which typically leverage diverse and high-quality data sources,
often overlook the intrinsic uncertainty of tasks, learning all