As I searched about this two terms, I found they are somehow like each other, both try to create a vector from raw data as I understood. But, what is the difference of this two term?

- 2,626
- 1
- 19
- 31
-
Have you read https://datascience.stackexchange.com/a/54045/58034 and page 2 of: https://arxiv.org/abs/1905.06147 and https://cloud.google.com/architecture/overview-extracting-and-serving-feature-embeddings-for-machine-learning - You are allowed to answer your own question. – Rob Jul 29 '22 at 10:50
-
@Rob thank you Rob, I read both of them now, and still it is complicated, I still think their definitions are like each other, can you give me a hint? any short answer can be really helpful. – aliiiiiiiiiiiiiiiiiiiii Jul 29 '22 at 11:02
-
1Explained in the first five parts: https://developers.google.com/machine-learning/crash-course/embeddings/video-lecture or see: https://www.toptal.com/machine-learning/embeddings-in-machine-learning https://analyticsindiamag.com/machine-learning-embedding/ - embeddings *group* the representation (the data set) into categories seperated by varying distances, often with partial overlap between individual embeds. – Rob Jul 29 '22 at 12:36
2 Answers
Vector representation is a generic term used to talk about any type of feature encoding, embedding vectors are instead a special case of vector representation.
When talking about vector representation the only underlying assumption is that every variable was encoded into numerical values, without any restriction regarding the numbers or the vector itself.
Embedding vectors instead are specifically continuous vectors of fixed dimensions obtained trough matrix factorization techniques or deep learning models. They originally proposed to encode text in the Word2Vec paper, and since then they acquired more and more popularity due to the high generalization potential of the proposed method in other AI branched rather than natural language processing.

- 5,153
- 1
- 11
- 25
To the best of my knowledge:
A latent representation, or code, $\mathbb{z}$ is a learned continuous vector representation of some input $\mathbb{x}$. E.g., we might have a code of an image.
An embedding is a type of code which specifically represents a discrete variable $\mathbb{x}_\text{discrete}$. One popular example is a word embedding, where words are encoded discretely using one-hot encoding.