The field of natural language processing and multimodal models is moving towards improving temporal understanding and reasoning capabilities. Recent studies have highlighted the importance of evaluating and enhancing the temporal consistency of large language models, as well as their ability to interpret and reason about time in various contexts.
Noteworthy papers in this area include: Temporal Referential Consistency: Do LLMs Favor Sequences Over Absolute Time References?, which introduces a novel benchmark for evaluating temporal referential consistency in large language models. A Matter of Time: Revealing the Structure of Time in Vision-Language Models, which investigates the temporal awareness of vision-language models and proposes methods to derive an explicit timeline representation from the embedding space.