Overview - Automated Essay Scoring (AES) is one of the technologies that is changing how evaluation in human learning is carried out by enhancing the speed and uniformity of scoring. Manual grading is a meticulous and often equally laborious activity, and grade feedback is slow to come back in less preserved regions on the social map. AES uses machine learning techniques to process essay scores automatically. This paper presents a comparative study of several AES models employing the largest publicly accessible dataset of contemporary educational standards essays. The models used are: Linear Regression (Cohen’s Kappa: 0.6540), XGBoost (0.7100), LGBM (0.7210), LSTM (0.7710), and BERT (0.7806). The findings showed, that deep learning models, especially LSTM and BERT, eclipsed the rest, with higher score Estimation precision and reproducibility over the traditional methods. This study proposes a new public AES system with the goal of improving the efficiency of providing automated evaluation services to teachers while speeding up the objective informational-dependent learning feedback for the students.
The dataset used in this project is available on Kaggle. download from here https://www.kaggle.com/competitions/learning-agency-lab-automated-essay-scoring-2/data
download from this repository
Size: 24,000 argumentative essays
Score Range: 1 to 6
- Text Statistics: Word count, sentence count, average sentence length
- Text Normalization: Lowercasing, punctuation removal
- Grammar & Spelling Checks: Ensuring clarity and correctness
- Transformer-based Tokenization: Handling long essays in BERT
numpy==1.26.4
pandas==2.2.2
matplotlib==3.7.5
seaborn==0.12.2
scikit-learn==1.2.2
scipy==1.11.4
xgboost==2.0.3
lightgbm==4.2.0
tensorflow==2.15.0
torch==2.2.0
transformers==4.40.0
sentencepiece==0.2.0
nltk==3.8.1 # For text preprocessing
tqdm==4.66.2
joblib==1.4.2
h5py==3.10.0 # For saving deep learning models
Model | Cohen’s Kappa | QWK |
---|---|---|
Linear Regression | 0.6540 | 0.678 |
XGBoost | 0.7100 | 0.725 |
LightGBM | 0.7210 | 0.739 |
LSTM | 0.7710 | 0.782 |
BERT | 0.7806 | 0.795 |
![]() |