Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Related study on agentic translation being used to improve traditional MT systems #9

Open
enismaxim1 opened this issue Jun 11, 2024 · 2 comments
Labels
research paper(s) Comment or submission for related research

Comments

@enismaxim1
Copy link

If agentic translations can generate better results than traditional architectures (such as an end-to-end transformer that inputs a text and directly outputs a translation) -- which are often faster/cheaper to run than our approach here -- this also provides a mechanism to automatically generate training data (parallel text corpora) that can be used to further train and improve traditional algorithms. (See also this article in The Batch on using LLMs to generate training data.)

For those interested in this idea, a collaborator and I wrote a paper in April called "From LLM to NMT" demonstrating the viability of this approach. It turns out Claude 3 Opus is already a state-of-the-art LLM agent in machine translation in various languages. We then use the LLM to generate train-data for Yoruba-English translation and create a state-of-the-art translation system.

@siddhantx0
Copy link

siddhantx0 commented Jun 12, 2024 via email

@sharpHL
Copy link

sharpHL commented Jun 17, 2024

If agentic translations can generate better results than traditional architectures (such as an end-to-end transformer that inputs a text and directly outputs a translation) -- which are often faster/cheaper to run than our approach here -- this also provides a mechanism to automatically generate training data (parallel text corpora) that can be used to further train and improve traditional algorithms. (See also this article in The Batch on using LLMs to generate training data.)

For those interested in this idea, a collaborator and I wrote a paper in April called "From LLM to NMT" demonstrating the viability of this approach. It turns out Claude 3 Opus is already a state-of-the-art LLM agent in machine translation in various languages. We then use the LLM to generate train-data for Yoruba-English translation and create a state-of-the-art translation system.

good job!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
research paper(s) Comment or submission for related research
Projects
None yet
Development

No branches or pull requests

4 participants