GRU belongs to the family of recurrent neural networks. This family of neural networks works on sequence data.
But, it is taking time for me to understand the differences between sequence length and input in the case of a GRU cell.
In the case of a CNN, the input tensor is of the form $B \times C \times H \times W$. Here $B$ is the batch size, $C$ is the number of channels, $H$ is the height of the image and $W$ is the width of the image.
We can visualize the input and outputs of CNN here.
Similarly, in the case of a GRU layer, the input tensor is of the form $B \times L \times I$. Here $B$ is the batch size, $L$ is the length of the sequence and $I$ is the number of input features.
I want to know what exactly the input and sequence lengths are. If there are any visualizations, please provide them.