⛔ [DEPRECATED] Adapt Transformer-based language models to new text domains
-
Updated
Feb 21, 2024 - Jupyter Notebook
⛔ [DEPRECATED] Adapt Transformer-based language models to new text domains
We use phonetics as a feature to create a joint semantic-phonetic embedding and improve the neural machine translation between Chinese and Japanese. 🥳
This project leverages deep learning transformers to classify YouTube comments into six distinct emotions.
Hugging Face Transformers offer a powerful framework for state-of-the-art NLP, with the Pipeline API for easy inference, Tokenization for efficient preprocessing, and Quantization for optimized deployment.
Run large language models like Qwen and LLaMA locally on Android for offline, private, real-time question answering and chat — powered by ONNX Runtime.
Sesta attività di Big Data Analytics
Add a description, image, and links to the huggingface-tokenizers topic page so that developers can more easily learn about it.
To associate your repository with the huggingface-tokenizers topic, visit your repo's landing page and select "manage topics."