This paper demonstrates that there is a discrete-time analogue which does not require any restriction on the size of the time-step in order to preserve the exponential stability of an artificial neural network with distributed delays. The analysis exploits an appropriate Lyapunov sequence and a discrete-time system of Halanay inequalities, and also either a Young inequality or a geometric-arithmetic mean inequality, to derive several sufficient conditions on the network parameters for the exponential stability of the analogue. The sufficiency conditions are independent of the time-step, and they correspond to those that establish the exponential stability of the continuous-time network.