Advancements in Clinical and Software Applications of Large Language Models

The field of large language models (LLMs) is rapidly advancing, with significant developments in clinical and software applications. Researchers are exploring the use of LLMs in clinical note generation, with studies investigating their reliability and consistency. The results show that LLMs can generate high-quality clinical notes, but their performance may vary depending on the specific model and task. In software engineering, LLMs are being used to automate requirements generation and release note generation, with promising results. However, more work is needed to refine these systems and ensure their accuracy and reliability. Noteworthy papers in this area include Assessing the Quality of AI-Generated Clinical Notes, which developed a framework to evaluate the quality of LLM-generated clinical notes, and ReqBrain, which introduced a task-specific instruction tuning method for LLMs in requirements generation. Another notable paper is SmartNote, which proposed a personalized release note generator using LLMs. Overall, the field is moving towards increased adoption of LLMs in clinical and software applications, with a focus on improving their performance, reliability, and usability.

Sources

Assessing the Quality of AI-Generated Clinical Notes: A Validated Evaluation of a Large Language Model Scribe

Are LLMs reliable? An exploration of the reliability of large language models in clinical note generation

ReqBrain: Task-Specific Instruction Tuning of LLMs for AI-Assisted Requirements Generation

SmartNote: An LLM-Powered, Personalised Release Note Generator That Just Works

LLM assisted web application functional requirements generation: A case study of four popular LLMs over a Mess Management System

OpenReview Should be Protected and Leveraged as a Community Asset for Research in the Era of Large Language Models

Assessing and Refining ChatGPT's Performance in Identifying Targeting and Inappropriate Language: A Comparative Study

Natural Language Processing in Support of Evidence-based Medicine: A Scoping Review

Can Large Language Models Match the Conclusions of Systematic Reviews?

Enhancing Study-Level Inference from Clinical Trial Papers via RL-based Numeric Reasoning

Evaluating the performance and fragility of large language models on the self-assessment for neurological surgeons

Built with on top of