Evaluating Long-Context LLMs
This paper proposes a novel way to evaluate large language models (LLMs) that claim to handle long contexts effectively. The researchers introduce a benchmark known as N, enhancing the traditional Needle-in-a-Haystack (NIAH) tests by eliminating literal matches between the search context and the re