This paper 1 considers a new variant of AMSGrad called Optimistic-AMSGrad. AMSGrad [31] is a popular adaptive gradient based optimization algorithm that is widely used in training deep neural networks. The new variant assumes that minibatch gradients in consecutive iterations have some underlying structure, which makes the gradients sequentially predictable. By exploiting the predictability and some ideas from Optimistic Online learning, the proposed algorithm can accelerate the convergence and also enjoys a tighter regret bound. We evaluate Optimistic-AMSGrad and AMSGrad in terms of various performance measures (i.e., training loss, testing loss, and classification accuracy on training/testing data), which demonstrate that Optimistic-AMSGrad improves AMSGrad. We release the code for reproducing the experiments on a github repository https://github.com/jimwang123/optimistic-amsgrad. 2