Skip to content

This repository contains implementations and analyses of fundamental NLP techniques and modern transformer architectures. Below you'll find Batman-style tech briefings for each project!

License

Notifications You must be signed in to change notification settings

CAI991108/Artificial-Intelligence-Utils

Repository files navigation

AI Projects Portfolio | Artificial Intelligence Utils 📚

AI Wizard
"Turning coffee into neural networks since 2025" ☕➡️🤖

Welcome to my AI project portfolio! This repository contains implementations and analyses of fundamental NLP techniques and modern transformer architectures. Below you'll find Batman-style tech briefings for each project!


🗂️ Project Catalog

1. 🔍 Word Embeddings Showdown

CBOW vs Skip-Gram vs GloVe
NLP Fundamentals with Reuters Financial News

Open in Colab PyTorch

🎯 Key Insights

  • Implemented 3 classic embedding models from scratch
  • Developed multi-modal evaluation framework:
    • KNN Semantic Clustering 👯
    • SimLex-999 Benchmark 📈
    • Vector Arithmetic for Analogies ➕➖

📉 Performance Matrix

Model KNN Clustering SimLex-999 ρ Analogy Accuracy
CBOW 🌕🌕🌗🌑🌑 0.0954 0%
SkipGram 🌕🌕🌑🌑🌑 0.0504 0%
GloVe 🌕🌑🌑🌑🌑 0.0659 0%

💡 Epiphany Moment: Even financial jargon needs bigger embeddings! (64-dim wasn't cutting it)


2. 🌀 Transformer Times

Chinese News Classification
Battling with 10 News Categories

Hugging Face WandB

🚀 Turbocharged Architecture

  • Scaled encoder layers: 2 → 8 🏗️
  • Enhanced classification head with GAP 🎯
  • Cosine decay + warmup scheduling 🔥

📊 Results Evolution

Version Accuracy F1-Score Key Improvement
Baseline 81.19% 82.55% Initial Transformer
+Preprocess 83.53% 83.34% Punctuation Ninjutsu ✂️
Final Model 84.07% 84.19% Deep Encoder Magic 🧙

Hot Take 🔥: Commas matter! But sometimes they don't... 🤷


3. 🤖 BERT Unleashed

Dual Task Dominance
Sentiment Analysis + Paraphrase Detection

BERT GPU Power

Mission Parameters

{'tasks': ['SST2', 'MRPC'],
 'model': 'bert-mini',
 'secret_sauce': 'Custom LossCallback() 🕵️',
 'hardware': 'Enough CUDA cores to fry an egg 🍳'}

📈 Performance Metrics

Task Accuracy F1-Score Prediction Prowess
SST2 82.80% 83.11% 4/5 Test Samples Correct 🎬
MRPC 75.25% 82.43% 5/5 Real-world Correct 🌍

Golden Insight 💡: Small BERTs can play big! (But they still hate irony)


🧪 Lab Environment Specs

please refer to reports accrodingly to each project:

📜 Project Reports

Project Report Link
Word Embeddings Analysis project1-2_report.pdf
News Classification project3_report.pdf
BERT Classification project4_report.pdf

📣 Future Quest Log

  • Subword embeddings for rare financial terms 💼

  • Hybrid positional encoding strategies 🧬

  • Attention visualization toolkit 👀

  • Domain-adaptive pretraining 🌐


Made with ❤️ (and probably too much caffeine) by Zijin Cai

"If debugging is removing bugs, then programming must be putting them in." - Edsger Dijkstra

About

This repository contains implementations and analyses of fundamental NLP techniques and modern transformer architectures. Below you'll find Batman-style tech briefings for each project!

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages