Current distributed routing control algorithms for dynamic networks model networks using the time evolution of density at network edges, while the routing control algorithm ensures edge density to converge to a Wardrop equilibrium, which was characterized by an equal traffic density on all used paths. We rearrange the density model to recast the problem within the framework of mean-field games. In doing that, we illustrate an extended state-space solution approach and we study the stochastic case where the density evolution is driven by a Brownian motion. Further, we investigate the case where the density evolution is perturbed by a bounded adversarial disturbance. For both the stochastic and the worst-case scenarios, we provide conditions for the density to converge to a preassigned set. Moreover, we analyze such conditions from two different perspectives, repeated games with vector payoffs and inclusion theory.