In the context of Reinforcement Learning, I have seen that the policy $\pi$ (for some algorithms) is nothing but a Neural Network architecture (for example a Feedforward Neural Network).
This policy is usually annotated as $\pi_{\theta}$, suggesting the policy is parameterized by $\theta$.
Question 1: Does this mean that $\theta$ in this case would represent all the Neural Network's parameters?
Question 2: The notation $\pi_{\theta}(a_{t}|s_{t})$ can be interpreted as "the output probability of Neural Network $\pi$ with parameters $\theta$ of selecting action $a_{t}$ when being input the state $s_{t}$"?
Here I leave an example from the Hugging Face RL course which uses this kind of notation.