Skip to content

Custom language model ranking and probability-based scoring for IR, built from scratch and benchmarked on queries.

Notifications You must be signed in to change notification settings

martinaspeciale/custom-lm

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

2 Commits
Β 
Β 
Β 
Β 

Repository files navigation

CustomLM

CustomLM is a from-scratch implementation of a transformer-based Language Model (GPT) designed for academic exploration and experimentation. Developed as part of a course project, it focuses on tokenization strategies, architecture design, and hyperparameter analysis.

πŸ“Œ Features

  • Custom GPT Architecture
    Manual implementation of a transformer model inspired by GPT, built with PyTorch.

  • Flexible Tokenization
    Models trained on character-level, syllable-level, and word-level representations.

  • Training and Evaluation
    In-depth analysis of training/validation loss and time across hyperparameter configurations.

  • Text Generation
    Sequence generation with top-performing model variants.

πŸš€ Technologies

  • PyTorch, NLTK, datasets
  • Includes a custom syllable tokenizer and manual tokenization logic.
  • Runs in Kaggle (GPU-enabled) for training efficiency.

πŸ‘₯ Authors

Filippo Lucchesi, Francesco Pio Crispino, Martina Speciale

About

Custom language model ranking and probability-based scoring for IR, built from scratch and benchmarked on queries.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •