Advancements in Large Language Models and Natural Language Processing

The field of natural language processing is witnessing significant developments, particularly in the area of large language models (LLMs). Researchers are exploring innovative ways to improve the performance and reliability of LLMs, including integrating traditional feature-based methods with modern PLM-based approaches. The use of LLMs in automated scoring, authorship attribution, and machine-generated text detection is becoming increasingly prevalent. Additionally, there is a growing interest in developing benchmarks and frameworks to evaluate the performance of LLMs, such as arXivBench and OpenTuringBench. Noteworthy papers in this area include the introduction of LazyReview, a dataset for uncovering lazy thinking in NLP peer reviews, and PlanGlow, a personalized study planning system that leverages LLMs to generate explainable and controllable study plans. These advancements have the potential to enhance various applications, from academic research to personalized learning, and highlight the importance of continued innovation in the field of natural language processing.

Sources

Integrated ensemble of BERT- and features-based models for authorship attribution in Japanese literary works

Lexical Bundle Frequency as a Construct-Relevant Candidate Feature in Automated Scoring of L2 Academic Writing

ArxivBench: Can LLMs Assist Researchers in Conducting Research?

LEMUR Neural Network Dataset: Towards Seamless AutoML

LazyReview A Dataset for Uncovering Lazy Thinking in NLP Peer Reviews

OpenTuringBench: An Open-Model-based Benchmark and Framework for Machine-Generated Text Detection and Attribution

You've Changed: Detecting Modification of Black-Box Large Language Models

PlanGlow: Personalized Study Planning with an Explainable and Controllable LLM-Driven System

Built with on top of