8000 GitHub - BBeeChu/RMTS
[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
Skip to content

BBeeChu/RMTS

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

RMTS 🤖

License: MIT code data data

📖 Overview

Image Existing automated essay scoring (AES) has solely relied on essay text without using explanatory rationales for the scores, thereby forgoing an opportunity to capture the specific aspects evaluated by rubric indicators in a fine-grained manner. This paper introduces Rationale-based Multiple Trait Scoring (RMTS), a novel approach for multi-trait essay scoring that integrates prompt-engineering-based large language models (LLMs) with fine-tuning-based essay scoring model using smaller large language models (S-LLMs). RMTS uses a LLM-based trait-wise rationale generation system where a separate LLM agent generates trait-specific rationales based on rubric guidelines, which the scoring model uses to accurately predict multi-trait scores. Extensive experiments on benchmark datasets, including ASAP, ASAP++, and Feedback Prize, show that RMTS significantly outperforms state-of-the-art models and vanilla S-LLMs in trait-specific scoring. By assisting quantitative assessment with fine-grained qualitative rationales, RMTS enhances the trait-wise reliability, providing partial explanations about essays.

📑 Paper

Rationale Behind Essay Scores: Enhancing S-LLM’s Multi-Trait Essay Scoring with Rationale Generated by LLMs
Seong Yeub Chu, Jong Woo Kim, Mun Yong Yi
Proceedings of the 2025 Conference of the North American Chapter of the Association for Computational Linguistics (NAACL '25). arXiv

⭐ Main Feature

LLM-based trait-wise rationale generation system

  • Rationale generation using LLM and rubric guideline. Image

Essay scoring with smaller LLMs

  • Incorporating a rationale with an essay for essay scoring using S-LLMs.

💻 Getting Started

Installation

scikit-learn            
scipy
tqdm
transformers==4.37.2
datasets
numpy
pandas
accelerate

How to Run (Evaluator: GPT-3.5-Turbo / Dimension: Coherence)

pip install -r requirements.txt
python main.py --model_name t5-base

🔧 Stack

  • Language: Python
  • Utilized LLMs: GPT-3.5-Turbo, Llama-3.1-8B-Instruct
  • Dependencies : Refer to "requirements.txt"
  • Dataset : ASAP & Feedback Prize

Project Structure

RMTS
├──data
│   ├── essay
│   └── feedback
├──models
│   ├── customized_modeling_bart
│   ├── customized_modeling_led
│   ├── customized_modeling_pegasus
│   ├── customized_modeling_t5
├──evaluation
└──utils

llama-emoji How to Run LLama-3.1

git lfs install
git clone https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

0