No, a model in RL is considered something (you can even consider it as a black box) that knows how to transition from one state to the next one, so for example, if we are playing snake, if I give you the current grid and the action where to move the head, an optimal model will have learnt the next game board, with the snake moved in the direction of the action
Then usually you can also learn the reward function with the model, but the model by itself tells you the next state you will be
Now, don't confuse it with the environment, the model learns to imitate the environment, but it's not it
After all of this, as you can see, $Q$ learning is not able to do that, it only tells you for each state and action what's the expected return, but not "if you are in this state $s_t$ and you do this action $a_t$ you'll end up at this state $s_{t+1}$"