Large Language Models Can Self-Improve in Long-context Reasoning
https://arxiv.org/pdf/2411.08147Large language models (LLMs) have achieved substantial progress in processing long contexts but still struggle with long-context reasoning. Existing approaches typically involve fine-tuning LLMs with synthetic data, which depends on annotations from human experts or advanced models like GPT-4, thus restricting further advancements. To address this issue, we invest..