For questions related to machine translation (MT), which is the task of translating text using a computer, machine or software.
Questions tagged [machine-translation]
50 questions
30
votes
9 answers
What is the actual quality of machine translations?
As an AI layman, till today I am confused by the promised and achieved improvements of automated translation.
My impression is: there is still a very, very far way to go. Or are there other explanations why the automated translations (offered and…

Hans-Peter Stricker
- 811
- 1
- 8
- 20
5
votes
1 answer
Is there any resource that describes in detail a naive example-based machine translation algorithm?
I'm looking to develop a machine translation tool for a constructed language. I think that the example-based approach is the most suitable because the said language is very regular and I can have a sufficient amount of parallel translations.
I…
user34314
4
votes
0 answers
Can sequence-to-sequence models be used to convert source code from one programming language to another?
Sequence-to-sequence models have achieved good performance in natural language translation. Could these models also be applied to convert source code written in one programming language to source code written in another language? Could they also be…

pfds2222
- 41
- 1
4
votes
1 answer
What are the differences between Bytenet and Wavenet?
I recently read Bytenet and Wavenet and I was curious why the first model is not as popular as the second. From my understanding, Bytenet can be seen as a seq2seq model where the encoder and the decoder are similar to Wavenet. Following the trends…

razvanc92
- 1,108
- 7
- 18
4
votes
1 answer
Is this technique image processing or computer vision?
If I use my mobile camera on a signboard or announcement board on a road or in a street (like the one attached in photo) where the message is written in Russian and my mobile shows me that message in English, would this be an image processing or…

LECS
- 171
- 4
4
votes
1 answer
How do I identify a monologue or dialogue in a conversation?
How do I identify monologues and dialogues in a conversation (or transcript) using natural language processing? How do I distinguish between the two?

Surya Shrivastava
- 41
- 1
4
votes
0 answers
What is the intuition behind the calculation of the similarity between encoder and decoder states?
Suppose that we are doing machine translation. We have a conditional language model with attention where we are are trying to predict a sequence $y_1, y_2, \dots, y_J$ from $x_1, x_2, \dots x_I$: $$P(y_1, y_2, \dots, y_{J}|x_1, x_2, \dots x_I) =…

AIguy122
- 41
- 2
3
votes
2 answers
Why can't we use Google Translate for every translation task?
Once a book is published in a language, why can't the publishers use Google Translate AI or some similar software to immediately render the book in other languages? Likewise for Wikipedia: I'm not sure I understand why we need editors for each…

Siddhartha
- 413
- 2
- 11
3
votes
1 answer
Machine Translation Transformers: Why Mask in Decoder?
I am trying to understand the purpose of masking in the decoder in the "Attention is All you Need" paper. Why wouldn't we want to rely on bidirectional context when translating? What goes wrong if masking isn't used in the decoder? Are there…

gumbelgrumbelgumbel
- 31
- 1
3
votes
1 answer
How do Transformer decoders handle arbitrary length input?
I am working through a Tensorflow Neural Machine Translation tutorial (https://www.tensorflow.org/text/tutorials/transformer) and am confused about how the decoder handles inputs when making inferences after it has been trained.
In the section where…

Dylan Larrabee
- 31
- 1
3
votes
0 answers
What is input (and shape) to K/V/Q of self-attention of EACH Decoder block of Language-translation model Transformer's tokens during Inference?
Transformer model of the original Attention paper has a decoder unit that works differently during Inference than Tranining.
I'm trying to understand the shapes used during decoder (both self-attention and enc-dec-attention blocks), but it's very…

Joe Black
- 181
- 6
3
votes
1 answer
How to do machine translation with no labeled data?
Is it be possible to train a neural network, with no parallel bilingual data, for machine translation?

Lerner Zhang
- 877
- 1
- 7
- 19
3
votes
0 answers
How to back propagate for implementation of Sequence-to-Sequence with Multi Decoders
I am proposing a modified version of Sequence-to-Sequence model with dual decoders. The problem that I am trying to solve is Neural Machine Translation into two languages at once. This is the simplified illustration for the model.
…

dee cue
- 143
- 5
2
votes
2 answers
What are the differences between BLEU and METEOR?
I am trying to understand the concept of evaluating the machine translation evaluation scores.
I understand how what BLEU score is trying to achieve. It looks into different n-grams like BLEU-1,BLEU-2, BLEU-3, BLEU-4 and try to match with the human…

Exploring
- 223
- 6
- 16
2
votes
1 answer
How is Google Translate able to convert texts of different lengths?
According to my experience with Tensorflow and many other frameworks, neural networks have to have a fixed shape for any output, but how does Google translate convert texts of different lengths?

Gavin
- 49
- 3