Advancements in Large Language Models and Ontology-Based Process Models

The field of artificial intelligence is witnessing significant developments in the areas of large language models (LLMs) and ontology-based process models. Researchers are focusing on enhancing the reasoning capabilities of LLMs, improving their ability to understand and represent real-world entities, and developing more robust evaluation frameworks to assess their performance. Notably, there is a growing interest in addressing the limitations of LLMs, such as their tendency to memorize rather than truly understand concepts, and their vulnerability to biases and perturbations. In the domain of ontology-based process models, researchers are working on developing more consistent and semantically coherent models, enabling the integration of standardized process semantics and formal mathematical constructs.

Some noteworthy papers in this area include: Measuring (a Sufficient) World Model in LLMs: A Variance Decomposition Framework, which proposes a formal framework for evaluating the robustness of LLMs' world models. ReasonGRM: Enhancing Generative Reward Models through Large Reasoning Models, which introduces a novel framework for improving the reasoning capabilities of generative reward models. The MedPerturb Dataset: What Non-Content Perturbations Reveal About Human and Clinical LLM Decision Making, which presents a dataset designed to evaluate the clinical robustness of LLMs under controlled perturbations.

Sources

Consistency Verification in Ontology-Based Process Models with Parameter Interdependencies

Locality in Many-Valued Structures

Can structural correspondences ground real world representational content in Large Language Models?

Measuring (a Sufficient) World Model in LLMs: A Variance Decomposition Framework

ReasonGRM: Enhancing Generative Reward Models through Large Reasoning Models

The Role of Model Confidence on Bias Effects in Measured Uncertainties

Dispositions and Roles of Generically Dependent Entities

A Note on Proper Relational Structures

The MedPerturb Dataset: What Non-Content Perturbations Reveal About Human and Clinical LLM Decision Making

Mirage of Mastery: Memorization Tricks LLMs into Artificially Inflated Self-Knowledge

HARPT: A Corpus for Analyzing Consumers' Trust and Privacy Concerns in Mobile Health Apps

Potemkin Understanding in Large Language Models

"What's Up, Doc?": Analyzing How Users Seek Health Information in Large-Scale Conversational AI Datasets

Built with on top of