5

What is the reason behind the name "Transformers", for Multi Head Self-Attention-based neural networks from Attention is All You Need?

I have been googling this question for a long time, and nowhere I can find any explanation.

Leevo
  • 285
  • 1
  • 9
  • Comments are not for extended discussion; this conversation has been [moved to chat](https://chat.stackexchange.com/rooms/120926/discussion-on-question-by-leevo-why-are-transformers-called-this-way). – nbro Mar 16 '21 at 17:50
  • @nbro: Would you mind asking the authors why they have chosen the name "Transformer"? I am curious, too. – Hans-Peter Stricker May 14 '23 at 10:25

1 Answers1

2

The authors of the original paper don't provide an explanation, but I suspect it's a combination of:

brazofuerte
  • 991
  • 8
  • 24