LAMBADA
A long-context language modeling benchmark where the final word of a passage must be predicted from broader discourse.
About LAMBADA
In LAMBADA, the correct final token is typically not predictable from the last sentence alone but becomes obvious given the full preceding context. This stresses long-range coherence, entity tracking, and narrative understanding that go beyond local n-gram statistics.
LAMBADA remains a concise, widely used signal of generative models’ ability to leverage extended context for accurate next-token prediction.