Background: The Pavlovian go/no-go task is commonly used to measure individual differences in Pavlovian biases and their interaction with instrumental learning. However, prior research has found suboptimal reliability for computational model-based performance measures for this task, limiting its usefulness in individual-differences research. These studies did not make use of several strategies previously shown to enhance task-measure reliability (e.g., task gamification, hierarchical Bayesian modeling for model estimation). Here we investigated if such approaches could improve the task’s reliability. Methods: Across two experiments, we recruited two independent samples of adult participants (N=103, N=110) to complete a novel, gamified version of the Pavlovian go/no-go task multiple times over several weeks. We used hierarchical Bayesian modeling to derive reinforcement learning model-based indices of participants' task performance, and additionally to estimate the reliability of these measures. Results: In Experiment 1, we observed considerable and unexpected practice effects, with most participants reaching near-ceiling levels of performance with repeat testing. Consequently, the test-retest reliability of some model parameters was unacceptable (range: 0.379–0.973). In Experiment 2, participants completed a modified version of the task designed to lessen these practice effects. We observed greatly reduced practice effects and improved estimates of the test-retest reliability (range: 0.696–0.989). Conclusion: The results demonstrate that model-based measures of performance on the Pavlovian go/no-go task can reach levels of reliability sufficient for use in individual- differences research. However, additional investigation is necessary to validate the modified version of the task in other populations and settings.