We establish the convergence of the unified two-timescale Reinforcement Learning (RL) algorithm presented in [Angiuli et al., 2022b]. This algorithm provides solutions to Mean Field Game (MFG) or Mean Field Control (MFC) problems depending on the ratio of two learning rates, one for the value function and the other for the mean field term. We focus a setting with finite state and action spaces, discrete time and infinite horizon. The proof of convergence relies on a generalization of the two-timescale approach of [Borkar, 1997]. The accuracy of approximation to the true solutions depends on the smoothing of the policies. We then provide an numerical example illustrating the convergence. Last, we generalize our convergence result to a three-timescale RL algorithm introduced in [Angiuli et al., 2022a] to solve mixed Mean Field Control Games (MFCGs).