BanglaTranslation: Pretraining Vanilla Transformer
- Tech Stack: Transformers, Pandas, Numpy, huggingface
- Github URL: Project Link
Pre-trained a Vanilla Transformer model for English-to-Bangla translation, implementing the complete architecture from scratch.
Explored and integrated two variants of the attention mechanism — Multi-Head Attention and Multi-Query Attention — within a unified encoder-decoder framework.