-
Notifications
You must be signed in to change notification settings - Fork 13
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Improvements in current translation model [GSoC'23] #15
Comments
Can we use Gemini's API key and get translation done by it |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
The current model is a fine-tuned Seq2Seq model whereas my model is a transformer-based machine translation model.
One of the drawbacks of Seq2Seq modeling is that since Recurrent Neural Network (RNN) is used for constructing the encoder and the decoder, the architecture memorizes the entire sentence from the source language. Therefore, with the increase in the length of sentences, the performance would decrease.
Transformers, on the other hand, breaks the entire sentence from the source language into parts and then translates these parts making them much more efficient. It allows the decoder to see the entire input sequence all at once.
For example,
The translation for 'overgrown weeds' by:
Current model is: 'muy de la ciudad' which when translated by Google says 'very from the city'
My model is: 'Hierbas sobrecrecidas' which when translated by Google says 'overgrown grasses'
The results from my model are closer to being accurate than the current model. By fine-tuning my model, I believe the results generated will be more accurate.
The text was updated successfully, but these errors were encountered: