Processing math: 100%
Skip to yearly menu bar Skip to main content


Poster

Infinite-Horizon Reinforcement Learning with Multinomial Logit Function Approximation

Mingyu Kim · Dabeen Lee


Abstract: We study model-based reinforcement learning with non-linear function approximation where the transition function of the underlying Markov decision process (MDP) is given by a multinomial logit (MNL) model. We develop a provably efficient discounted value iteration-based algorithm that works for both infinite-horizon average-reward and discounted-reward settings. For average-reward communicating MDPs, the algorithm guarantees a regret upper bound of ˜O(dDT) where d is the dimension of feature mapping, D is the diameter of the underlying MDP, and T is the horizon. For discounted-reward MDPs, our algorithm achieves ˜O(d(1γ)2T) regret where γ is the discount factor. Then we complement these upper bounds by providing several regret lower bounds. We prove a lower bound of Ω(dDT) for learning communicating MDPs of diameter D and a lower bound of Ω(d(1γ)3/2T) for learning discounted-reward MDPs with discount factor γ. Lastly, we show a regret lower bound of Ω(dH3/2K) for learning H-horizon episodic MDPs with MNL function approximation where K is the number of episodes, which improves upon the best-known lower bound for the finite-horizon setting.

Live content is unavailable. Log in and register to view live content