GRU-based sequence-to-sequence translation model with teacher forcing and multiple optimizer comparisons.
A sequence-to-sequence neural machine translation model built in PyTorch. Uses GRU-based encoder and decoder with teacher forcing during training. Trained on 7,000 sentence pairs (≤8 words each) and benchmarked across Adam, AdamW, RMSprop, and SGD optimizers with L2 regularization, reaching 50% test accuracy.