This paper considers a problem of distributed hypothesis testing and cooperative learning. Individual nodes in a network receive noisy local (private) observations whose distribution is parameterized by a discrete parameter (hypotheses). The conditional distributions are known locally at the nodes, but the true parameter/hypothesis is not known. We consider a social ("non-Bayesian") learning rule from previous literature, in which nodes first perform a Bayesian update of their belief (distribution estimate) of the parameter based on their local observation, communicate these updates to their neighbors, and then perform a "non-Bayesian" linear consensus using the log-beliefs of their neighbors. For this learning rule, we know that under mild assumptions, the belief of any node in any incorrect parameter converges to zero exponentially fast, and the exponential rate of learning is a characterized by the network structure and the divergences between the observations' distributions. Furthermore, any (large) deviation from this nominal rate of rejecting wrong hypothesis is known to occur with an exponentially small probability when the bounded likelihood functions is uniformly bounded. While relaxing the assumption of uniformly bounded log-likelihood functions, we show that the rates of rejecting wrong hypotheses, collectively, satisfy a large deviation principle under a mild technical condition on the log-moment generating functions of log-likelihood random variables.