This is largely still a work in progress, however I so far have a functional transformer model that I can use for training langauge translation models. The transformer is based on the origional transformer architerture from the paper "attention is all you need". I coded it in python using the pytorch library and I am currenly testing it with the OPUS books database from hugging face.
Another description of the image on the left.
To-do
Description of the image on the right.
To-do
Another description of the image on the left.
To-do
Final description of the image on the right.