Stochastic optimal control studies the problem of sequential decision-making under uncertainty. Dynamic programming (DP) offers a principled approach to solving stochastic optimal control problems. A major drawback of DP methods, however, is that they become quickly intractable in large-scale problems. In this thesis, we show how structural results and various relaxation techniques can be used to obtain good approximations and accelerate learning. First, we propose a new provably convergent variant of Q-learning that leverages upper and lower bounds derived using information relaxation techniques to improve performance in the tabular setting. Second, we study weakly coupled DPs which are a broad class of stochastic sequential decision p...