Entropic Regularization of Markov Decision Processes

An optimal feedback controller for a given Markov decision process (MDP) can in principle be synthesized by value or policy iteration. However, if the system dynamics and the reward function are unknown, a learning agent must discover an optimal controller via direct interaction with the environment. Such interactive data gathering commonly leads to divergence towards dangerous or uninformative regions of the state space unless additional regularization measures are taken. Prior works proposed bounding the information loss measured by the Kullback–Leibler (KL) divergence at every policy improvement step to eliminate instability in the learning dynamics. In this paper, we consider a broader family of f-divergences, and more concretely α-divergences, which inherit the beneficial property of providing the policy improvement step in closed form at the same time yielding a corresponding dual objective for policy evaluation. Such entropic proximal policy optimization view gives a unified perspective on compatible actor-critic architectures. In particular, common least-squares value function estimation coupled with advantage-weighted maximum likelihood policy improvement is shown to correspond to the Pearson χ2-divergence penalty. Other actor-critic pairs arise for various choices of the penalty-generating function f. On a concrete instantiation of our framework with the α-divergence, we carry out asymptotic analysis of the solutions for different values of α and demonstrate the effects of the divergence function choice on common standard reinforcement learning problems.


Introduction
Sequential decision-making problems under uncertainty are described by the mathematical framework of Markov decision processes (MDPs) [1]. The core problem in MDPs is to find an optimal policy-a mapping from states to actions which maximizes the expected cumulative reward collected by an agent over its lifetime. In reinforcement learning (RL), the agent is additionally assumed to have no prior knowledge about the environment dynamics and the reward function [2]. Therefore, direct policy optimization in the RL setting can be seen as a form of stochastic black-box optimization: the agent proposes a query point in the form of a policy, the environment evaluates this point by computing the expected return, after that the agent updates the proposal and the process repeats [3]. There are two conceptual steps in this scheme known as policy evaluation and policy improvement [4]. Both steps require function approximation in high-dimensional and continuous state-action spaces due to the curse of dimensionality [4]. Therefore, statistical learning approaches are employed to approximate the value function of a policy and to perform policy improvement based on the data collected from the environment.
In contrast to traditional supervised learning, in reinforcement learning, the data distribution changes with every policy update. State-of-the-art generalized policy iteration algorithms [5][6][7][8] are mindful of this covariate shift problem [9], taking active measures to account for it. To smoothen the learning dynamics, these algorithms limit the information loss between successive policy updates as measured by the KL divergence or approximations thereof [10]. In the optimization literature, such approaches are categorized as proximal (or trust region) algorithms [11].
The choice of the divergence function determines the geometry of the information manifold [12]. Recently, in particular in the area of implicit generative modeling [13], the choice of the divergence function was shown to have a dramatic effect both on the optimization performance [14] and the perceptual quality of the generated data when various f -divergences were employed [15]. In this paper, we carry over the idea of using generalized entropic proximal mappings [16] given by an f -divergence to reinforcement learning. We show that relative entropy policy search [6], framed as an instance of stochastic mirror descent [17,18] as suggested by [10], can be extended to use any divergence measure from the family of f -divergences. The resulting algorithm provides insights into the compatibility of policy and value function update rules in actor-critic architectures, which we exemplify on several instantiations of the generic f -divergence with representatives from the parametric family of α-divergences [19][20][21].

Background
This section provides the necessary background on policy gradients [3] and entropic penalties [16] for later derivations and analysis. Standard RL notation [22] is used throughout.

Policy Gradient Methods
Policy search algorithms [3] commonly use the gradient estimator of the following form [23] g =Ê t ∇ θ log π θÂ w t (1) where π θ (a|s) is a stochastic policy andÂ w t (s t , a t ) is an estimator of the advantage function at timestep t. ExpectationÊ t [. . . ] indicates an empirical average over a finite batch of samples, in an algorithm that alternates between sampling and optimization. The advantage estimateÂ w t in (1) can be obtained from an estimate of the value function [24,25], which in its turn is found by least-squares estimation. Specifically, if V w (s) denotes a parametric value function, and ifV t = ∑ ∞ k=0 γ k R t+k is taken as its rollout-based estimate, then the parameters w can be found as The advantage estimateÂ w t = ∑ ∞ k=0 γ k δ w t+k is then obtained by summing the temporal difference errors , also known as the Bellman residuals. TreatingÂ w t as fixed for the purpose of policy improvement, we can view (1) as the gradient of an advantage-weighted log-likelihood; therefore, the policy parameters θ can be found as Thus, actor-critic algorithms that use the gradient estimator (1) to update the policy can be viewed as instances of the generalized policy iteration scheme, alternating between policy evaluation (2) and policy improvement (3). In the following, we will see that the actor-critic pair (2) and (3), that combines least-squares value function fitting with linear-in-the-advantage-weighted maximum likelihood policy improvement, is just one representative from a family of such actor-critic pairs arising for different choices of the f -divergence penalty within our entropic proximal policy optimization framework.

Entropic Penalties
The term entropic penalties [16] refers to both f -divergences and Bregman divergences. In this paper, we will focus on f -divergences, leaving generalization to Bregman divergences for future work. The f -divergence [26] between two distributions P and Q with densities p and q is defined as where f is a convex function on (0, ∞) with f (1) = 0 and P is assumed to be absolutely continuous with respect to Q. For example, the KL divergence corresponds to f 1 (x) = x log x − (x − 1), with the formula also applicable to unnormalized distributions [27]. Many common divergences lie on the curve of α-divergences [19,20] defined by a special choice of the generator function [21] f The α-divergence D α = D f α will be used as the primary example of the f -divergence throughout the paper. For more details on the α-divergence and its properties, see Appendix A. Noteworthy is the symmetry of the α-divergence with respect to α = 0.5, which relates reverse divergences as D 0.5+β (p q) = D 0.5−β (q p).

Entropic Proximal Policy Optimization
Consider the average-reward RL setting [2], where the dynamics of an ergodic MDP are given by the transition density p(s |s, a). An intelligent agent can modulate the system dynamics by sampling actions a from a stochastic policy π(a|s) at every time step of the evolution of the dynamical system. The resulting modulated Markov chain with transition kernel p π (s |s) = A p(s |s, a)π(a|s)da converges to a stationary state distribution µ π (s) as time goes to infinity. This stationary state distribution induces a state-action distribution ρ π (s, a) = µ π (s)π(a|s), which corresponds to visitation frequencies of state-action pairs [1]. The goal of the agent is to steer the system dynamics to desirable states. Such objective is commonly encoded by the expectation of a random variable R : S × A → R called reward in this context. Thus, the agent seeks a policy that maximizes the expected reward J(π) = E ρ π (s,a) [R(s, a)].
In reinforcement learning, neither the reward function R nor the system dynamics p(s |s, a) are assumed to be known. Therefore, to maximize (or even evaluate) the objective J(π), the agent must sample a batch of experiences in the form of tuples (s, a, r, s ) from the dynamics and use an empirical estimateĴ =Ê t [R(s t , a t )] as a surrogate for the original objective. Since the gradient of the expected reward with respect to the policy parameters can be written as [28] ∇ θ J(π θ ) = E ρ π θ (s,a) [∇ θ log π θ (a|s)R(s, a)] with a corresponding sample-based counterpart one may be tempted to optimize a sample-based objectivê on a fixed batch of data {(s, a, r, s ) t } N t=1 till convergence. However, such an approach ignores the fact that sampling distribution ρ π θ (s, a) itself depends on the policy parameters θ; therefore, such greedy optimization aims at a wrong objective [6]. To have the correct objective, the dataset must be sampled anew after every parameter update-doing otherwise will lead to overfitting and divergence. This problem is known in statistics as the covariate shift problem [9].

Fighting Covariate Shift via Trust Regions
A principled way to account for the change in the sampling distribution at every policy update step is to construct an auxiliary local objective function that can be safely optimized till convergence. Relative entropy policy search (REPS) algorithm [6] proposes a candidate for such an objective with π 0 being the current policy under which the data samples were collected, policy π being the improvement policy that needs to be found, and η > 0 being a 'temperature' parameter that determines how much the next policy can deviate from the current one. The original formulation employs a relative entropy trust region constraint D 1 with radius ε instead of a penalty, which allows for finding the optimal temperature η as a function of the trust region radius ε. Importantly, the objective function (5) can be optimized in closed form for policy π (i.e., treating the policy itself as a variable and not its parameters, in contrast to standard policy gradients). To that end, several constraints on ρ π are added to ensure stationarity with respect to the given MDP [6]. In a similar vein, we can solve Problem (5) with respect to π for any f -divergence with a twice differentiable generator function f .

Policy Optimization with Entropic Penalties
Following the intuition of REPS, we introduce an f -divergence penalized optimization problem that the learning agent must solve at every policy iteration step The agent seeks a policy that maximizes the expected reward and does not deviate from the current policy too much. The first constraint in (6) ensures that the policy is compatible with the system dynamics, and the latter two constraints ensure that π is a proper probability distribution. Please note that π enters Problem (6) indirectly through ρ π . Since the objective has the form of free energy [29] in ρ π with an f -divergence playing the role of the usual KL, the solution can be expressed through the derivative of the convex conjugate function f * , as shown for general nonlinear problems in [16], Here, {V(s), λ, κ(s, a)} are the Lagrange dual variables corresponding to the three constraints in (6), respectively. Although we get a closed-form solution for ρ π , we still need to solve the dual optimization problem to get the optimal dual variables Remarkably, the advantage function A V (s, a) = R(s, a) + S V(s )p(s |s, a)ds − V(s) emerges automatically in the dual objective. The advantage function also appears in the penalty-free linear programming formulation of policy improvement [1], which corresponds to the zero-temperature limit η → 0 of our formulation. Thanks to the fact that the dual objective in (8) is given as an expectation with respect to ρ π 0 , it can be straightforwardly estimated from rollouts. The last constraint in (8) on the argument of f * is easy to evaluate for common α-divergences. Indeed, the convex conjugate f * α of the generator function (4) is given by Thus, the constraint on arg f * in (4) is just a linear inequality y(1 − α) < 1 for any α-divergence.

Value Function Approximation
For small grid-world problems, one can solve Problem (8) exactly for V(s). However, for larger problems or if the state space is continuous, one must resort to function approximation. Assume we plug an expressive function approximator V w (s) in (8), then vector w becomes a new vector of parameters in the dual objective. Later, it will be shown that minimizing the dual when η → ∞ is closely related to minimizing the mean squared Bellman error.

Sample-Based Algorithm for Dual Optimization
To solve Problem (8) in practice, we gather a batch of samples from policy π 0 and replace the expectation in the objective with a sample average. Please note that in principle one also needs to estimate the expectation of the future rewards S V(s )p(s |s, a)ds . However, since the probability of visiting the same state-action pair in continuous space is zero, one commonly estimates this integral from a single sample [3], which is equivalent to assuming deterministic system dynamics. Inequality constraints in (8) are linear and they must be imposed for every (s, a) pair in the dataset.

Parametric Policy Fitting
Assume Problem (8) is solved on a current batch of data sampled from π 0 and thus the optimal dual variables {V(s), λ, κ(s, a)} are given. Equation (7) allows one to evaluate the new density ρ π (s, a) on any pair (s, a) from the dataset. However, it does not yield the new policy π directly because representation (7) is variational. A common approach [3] is to assume that the policy is represented by a parameterized conditional density π θ (a|s) and fit this density to the data using maximum likelihood.
To fit a parametric density π θ (a|s) to the true solution π(a|s) given by (7), we minimize the KL divergence D 1 (ρ π ρ π θ ). Minimization of this KL is equivalent to maximization of the weighted maximum likelihoodÊ[ f * (. . . ) log ρ π θ ]. Unfortunately, distribution ρ π θ (s, a) = µ π θ (s)π θ (a|s) is in general not known because µ π θ (s) does not only depend on the policy but also on the system dynamics. Assuming the effect of policy parameters on the stationary state distribution is small [3], we arrive at the following optimization problem for fitting the policy parameters Compare our policy improvement step (10) to the commonly used advantage-weighted maximum likelihood (ML) objective (3). They look surprisingly similar (especially if f * (y) = y is a linear function), which is not a coincidence and will be systematically explained in the next sections.

Temperature Scheduling
The 'temperature' parameter η trades off reward vs divergence, as can be seen in the objective function in Problem (6). In practice, devising a schedule for η may be hard because η is sensitive to reward scaling and policy parameterization. A more intuitive way to impose the f -divergence proximity condition is by adding it as a constraint D f (ρ π ρ π 0 ) ≤ ε with a fixed ε and then treating the temperature η ≥ 0 as an optimization variable. Such formulation is easy to incorporate into the dual (8) by adding a term ηε to the objective and a constraint η ≥ 0 to the list of constraints. Constraint-based formulation was successfully used before with a KL divergence constraint [6] and with its quadratic approximation [5,7].

Practical Algorithm for Continuous State-Action Spaces
Our proposed approach for entropic proximal policy optimization is summarized in Algorithm 1. Following the generalized policy iteration scheme, we (i) collect data under a given policy, (ii) evaluate the policy by solving (8), and (iii) improve the policy by solving (10). In the following section, several instantiations of Algorithm 1 with different choices of function f will be presented and studied.
Algorithm 1: Primal-dual entropic proximal policy optimization with function approximation under current policy π θ 0 ; policy evaluation: optimize dual (8) with V(s) = V w (s) to obtain critic parameters w; policy improvement: perform weighted ML update (10) to obtain actor parameters θ; end Output: Optimal policy π θ (a|s) and the corresponding value function V w (s)

High-and Low-Temperature Limits; α-Divergences; Analytic Solutions and Asymptotics
How does the f -divergence penalty influence policy optimization? How should one choose the generator function f ? What role does the step size play in optimization? This section will try to answer these and related questions. First, two special choices of the penalty function f are presented, which reveal that the common practice of using mean squared Bellman error minimization coupled with advantage reweighted policy update is equivalent to imposing a Pearson χ 2 -divergence penalty. Second, high-and low-temperature limits are studied, on one hand revealing the special role the Pearson χ 2 -divergence plays, being the high-temperature limit of all smooth f -divergences, and on the other hand establishing a link to the linear programming formulation of policy search as the low-temperature limit of our entropic penalty-based framework.
4.1. KL Divergence (α = 1) and Pearson χ 2 -Divergence (α = 2) As can be deduced from the form of (10), great simplifications occur when f * (y) is a linear function (α = 2, see (9)) or an exponential function (α = 1). The fundamental reason for such simplifications lies in the fact that linear and exponential functions are homomorphisms with respect to addition. This allows, in particular, discovery of a closed-form solution for the dual variable λ and thus eliminate it from the optimization. Moreover, in these two special cases, the dual variables κ(s, a) can also be eliminated. They are responsible for non-negativity of probabilities: when α = 1 (KL), κ(s, a) = 0 uniformly for all η ≥ 0, when α = 2 (Pearson), κ(s, a) = 0 for sufficiently big η. Table 1 gives the corresponding empirical actor-critic optimization objective pairs. A generic primal-dual actor-critic algorithm with an α-divergence penalty performs two steps (step 1: policy evaluation) minimize wĝ α (w) (step 2: policy improvement) maximize θL α (θ) inside a policy iteration loop. It is worth comparing the explicit formulas in Table 1 to the customarily used objectives (2) and (3). To make the comparison fair, notice that (2) and (3) correspond to discounted infinite horizon formulation with discount factor γ ∈ (0, 1) whereas formulas in Table 1 are derived for the average-reward setting. In general, the difference between these two settings can be ascribed to an additional baseline that must be subtracted in the average reward setting [2]. In our derivations, the baseline corresponds to the dual variable λ, as in classical linear programming formulation of policy iteration [1], and it is automatically gets subtracted from the advantage (see (8)).

Mean Squared Error Minimization with Advantage Reweighting is Equivalent to Pearson Penalty
The baseline for α = 2 is given by the average advantage λ 2 =Ê t Â w (s t , a t ) , which also equals the average return in our setting [1,2]. Therefore, to translate the formulas from Table 1 to the discounted infinite horizon form (2) and (3), we need to remove the baseline and add discounting to the advantage; that is, set A w (s, a) = R(s, a) + γ S V w (s )p(s |s, a)ds − V w (s). Then the dual objectivê is proportional to the average squared advantage. Naive optimization of (11) leads to the family of residual gradient algorithms [30,31]. However, if the same Monte Carlo estimate of the value function is used as in (2), then (11) and (2) are exactly equivalent. The same holds for the Pearson actor and the standard policy improvement (3) provided that η =Ê t Â w (s t , a t ) . That means (12) is equivalent to (3) if the weight of the divergence penalty is equal to the expected return.

High-and Low-Temperature Limits
In the previous subsection, we established a direct correspondence between the least-squares value function fitting coupled with the advantage-weighted maximum likelihood policy parameters estimation (2) and (3) and the dual-primal pair of optimization problems (11) and (12) arising from our Algorithm 1 for the special choice of the Pearson χ 2 -divergence penalty. In this subsection, we will show that this is not a coincidence but a manifestation of the fundamental fact that the Pearson χ 2 -divergence is the quadratic approximation of any smooth f -divergence about unity.

High Temperatures: All Smooth f -Divergences Tend Towards Pearson χ 2 -Divergence
There are two ways to show the independence of the primal-dual solution (8)-(10) on the choice of the divergence penalty: either exactly solve an approximate problem or approximate the exact solution of the original problem. In the first case, the penalty is replaced with its Taylor expansion at η → ∞, which turns out to be the Pearson χ 2 -divergence, and then the derivation becomes equivalent to the natural policy gradient derivation [5]. In the second case, the exact solution (8)-(10) is expanded by Taylor: for big η, dual variables κ(s, a) can be dropped if ρ π 0 (s, a) > 0, which yields By definition of the f -divergence, the generator function f satisfies the condition f (1) = 0. Without loss of generality [32], one can impose an additional constraint f (1) = 0 for convenience. Such constraint ensures that the graph of the function f (x) lies entirely in the upper half-plane, touching the x-axis at a single point x = 1. From the definition of the convex conjugate f * = ( f ) −1 , we can deduce that f * (0) = 1 and f * (0) = 0; by rescaling, it is moreover possible to set f (1) = f * (0) = 1. These properties are automatically satisfied by the α-divergence, which can be verified by a direct computation. With this in mind, it is straightforward to see that substitution of (13) into (8) yields precisely the quadratic objectiveĝ 2 (w) from Table 1, the difference being of the second order in 1/η. To obtain the asymptotic policy update objective, one can expand (10) in the high-temperature limit η → ∞ and observe that it equalsL 2 (θ) from Table 1 with the difference being of the second order in 1/η. Therefore, it is established that the choice of the divergence function plays a minor role for big temperatures (small policy update steps). Since this is the mode in which the majority of iterative algorithms operate, our entropic proximal policy optimization point of view provides a rigorous justification for the common practice of using the mean squared Bellman error objective for value function fitting and the advantage-weighted maximum likelihood objective for policy improvement.

Low Temperatures: Linear Programming Formulation Emerges in the Limit
Setting η to a small number is equivalent to allowing large policy update steps because η is the weight of the divergence penalty in the objective function (6). Such regime is rather undesirable in reinforcement learning because of the covariate shift problem mentioned in the introduction. Problem (6) for η → 0 turns into a well-studied linear programming formulation [1,10] that can be readily applied if the model {p(s |s, a), R(s, a)} is known.
It is not straightforward to derive the asymptotics of policy evaluation (8) and policy improvement (10) for a general smooth f -divergence in the low-temperature limit η → 0 because the dual variables κ(s, a) do not disappear, in contrast to the high-temperature limit (13). However, for the KL divergence penalty (see Table 1), one can show that the policy evaluation objective g 1 (w) tends towards the supremum of the advantage g 1 (w) → sup s,a A w (s, a); the optimal policy is deterministic, π(a|s) → δ(a − arg sup b A w (s, b)), therefore L(θ) → log π θ (ā|s) with (s,ā) = arg sup s ,a A w (s , a ).

Empirical Evaluations
To develop an intuition regarding the influence of the entropic penalties on policy improvement, we first consider a simplified version of the reinforcement learning problem-namely the stochastic multi-armed bandit problem [33]. In this setting, our algorithm is closely related to the family of Exp3 algorithms [34], originally motivated by the adversarial bandit problem. Subsequently, we evaluate our approach in the standard reinforcement learning setting.

Illustrative Experiments on Stochastic Multi-Armed Bandit Problems
In the stochastic multi-armed bandit problem [33], at every time step t ∈ {1, . . . , T}, an agent chooses among K actions a ∈ A. After every choice a t = a, it receives a noisy reward R t = R(a t ) drawn from a distribution with mean Q(a). The goal of the agent is to maximize the expected total reward J = E[∑ T t=1 R t ]. Given the true values Q(a), the optimal strategy is to always choose the best action, a * t = arg max a Q(a). However, due to the lack of knowledge, the agent faces the exploration-exploitation dilemma. A generic way to encode the exploration-exploitation trade-off is by introducing a policy π t , i.e., a distribution from which the agent draws actions a t ∼ π t . Thus, the question becomes: given the current policy π t and the current estimate of action valuesQ t , what should the policy π t+1 at the next time step be? Unlike the choice of the best action under perfect information, such sampling policies are hard to derive from first principles [35].
We apply our generic Algorithm 1 to the stochastic multi-armed bandit problem to illustrate the effects of the divergence choice. The value function disappears because there is no state and no system dynamics in this problem. Therefore, the estimateQ t plays the role of the advantage, and the dual optimization (8) is performed only with respect to the remaining Lagrange multipliers. Figure 1 shows the effects of the α-divergence choice on policy updates. We consider a 10-armed bandit problem with arm values Q(a) ∼ N (0, 1) and keep the temperature fixed at η = 2 for all values of α. Several iterations starting from an initial uniform policy are shown in the figure for comparison. Extremely large positive and negative values of α result in ε-elimination and ε-greedy policies, respectively. Small values of α, in contrast, weigh actions according to their values. Policies for α < 1 are peaked and heavy-tailed, eventually turning into ε-greedy policies when α → −∞. Policies for α ≥ 1 are more uniform, but they put zero mass on bad actions, eventually turning into ε-elimination policies when α → ∞. For α ≥ 1, policy iteration may spend a lot of time in the end deciding between two best actions, whereas for α < 1 the final convergence is faster.

Effects of α on Regret
The average regret C n = nQ max − E[∑ n−1 t=0 R t ] is shown in Figure 2 for different values of α as a function of the time step n with 95% confidence error bars. The performance of the UCB algorithm [33] is also shown for comparison. The presented results are obtained in a 20-armed bandit environment where rewards have Gaussian distribution R(a) ∼ N (Q(a), 0.5). Arm values are estimated from observed rewards and the policy is updated every 20 time steps. The temperature parameter η is decreased starting from η = 1 after every policy update according to the schedule η + = βη with β = 0.8. Results are averaged over 400 runs. In general, extreme α's accumulate more regret. However, they eventually focus on a single action and flatten out. Small α's accumulate less regret, but they may keep exploring sub-optimal actions longer. Values of α ∈ [0, 2] perform comparably with UCB after around 400 steps, once reliable estimates of values have been obtained.  Figure 3 shows the average regret after a given number of time steps as a function of the divergence type α. As can be seen from the figure, smaller values of α result in lower regret. Large negative α's correspond to ε-greedy policies, which oftentimes prematurely converge to a sub-optimal action, failing to discover the optimal action for a long time if the exploration probability ε is small. Large positive α's correspond to ε-elimination policies, which may by mistake completely eliminate the best action or spend a lot of time deciding between two options in the end of learning, accumulating more regret. The optimal value of the parameter α depends on the time horizon for which the policy is being optimized. Depending on the horizon, the minimum of the curves shifts from slightly negative α's towards the range α ∈ [0, 2] with increasing time horizon.

Empirical Evaluations on Ergodic MDPs
We evaluate our policy iteration algorithm with f -divergence on standard grid-world reinforcement learning problems from OpenAI Gym [36]. The environments that terminate or have absorbing states are restarted during data collection to ensure ergodicity. Figure 4 demonstrates the learning dynamics on different environments for various choices of the divergence function. Parameter settings and other implementation details can be found in Appendix B. In summary, one can either promote risk averse behavior by choosing α < 0, which may, however, result in sub-optimal exploration, or one can promote risk seeking behavior with α > 1, which may lead to overly aggressive elimination of options. Our experiments suggest that the optimal balance should be found in the range α ∈ [0, 1]. It should be noted that the effect of the α-divergence on policy iteration is not linear and not symmetric with respect to α = 0.5, contrary to what one could have expected given the symmetry of the α-divergence as a function of α. For example, switching from α = −3 to α = −2 may have little effect on policy iteration, whereas switching from α = 3 to α = 4 may have a much more pronounced influence on the learning dynamics. Results for different values of α are split into three subplots within each row, from the more extreme α's on the left to the more refined values on the right. In all cases, more negative values α < 0 initially show faster improvement because they immediately jump to the mode and keep the exploration level low; however, after a certain number of iterations they get overtaken by moderate values α ∈ [0, 1] that weigh advantage estimates more evenly. Positive α > 1 demonstrate high variance in the learning dynamics because they clamp the probability of good actions to zero if the advantage estimates are overly pessimistic, never being able to recover from such a mistake. Large positive α's may even fail to reach the optimum altogether, as exemplified by α = 10 in the plots. The most stable and reliable α-divergences lie between the reverse KL (α = 0) and the KL (α = 1), with the Hellinger distance (α = 0.5) outperforming both on the FrozenLake environment.

Related Work
Apart from computational advantages, information-theoretic approaches provide a solid framework for describing and studying aspects of intelligent behavior [37], from autonomy [38] and curiosity [39] to bounded rationality [40] and game theory [41].
Entropic proximal mappings were introduced in [16] as a general framework for constructing approximation and smoothing schemes for optimization problem. Problem formulation (6) presented here can be considered as an application of this general theory to policy optimization in Markov decision processes. Following the recent work [10] that establishes links between popular in reinforcement learning KL-divergence-regularized policy iteration algorithms [6,7] and a well-known in optimization stochastic mirror descent algorithm [17,18], one can view our Algorithm 1 as an analog of the mirror descent with an f -divergence penalty.
Concurrent works [42,43] consider similar regularized formulations, although in the policy space instead of the state-action distribution space and in the infinite horizon discounted setting instead of the average-reward setting. The α-divergence in its entropic form, i.e., when the base measure is a uniform distribution, was used in several papers under the name Tsallis entropy [44][45][46][47], where its sparsifying effect was exploited in large discrete action spaces.
An alternative proximal reinforcement learning scheme was introduced in [48] based on the extragradient method for solving variational inequalities and leveraging operator splitting techniques. Although the idea of exploiting proximal maps and updates in the primal and dual spaces is similar to ours, regularization in [48] is applied in the value function space to smoothen generalized TD learning algorithms, whereas we study regularization in the primal space.

Conclusions
We presented a framework for deriving actor-critic algorithms as pairs of primal-dual optimization problems resulting from regularization of the standard expected return objective with so-called entropic penalties in the form of an f -divergence. Several examples with α-divergence penalties have been worked out in detail. In the limit of small policy update steps, all f -divergences with twice differentiable generator function f are approximated by the Pearson χ 2 -divergence, which was shown to yield the most commonly used in reinforcement learning pair of actor-critic updates. Thus, our framework provides a sound justification for the common practice of minimizing mean squared Bellman error in the policy evaluation step and fitting policy parameters by advantage-weighted maximum likelihood in the policy improvement step.
In the future work, incorporating non-differentiable generator functions, such as the absolute value that corresponds to the total variation distance, may provide a principled explanation for the empirical success of the algorithms not accounted for by our current smooth f -divergence framework, such as the proximal policy optimization algorithm [8]. Establishing a tighter connection between online convex optimization that employs Bregman divergences and reinforcement learning will likely yield both a deeper understanding of the optimization dynamics in RL and allow for improved practical algorithms building on the firm fundament of optimization theory. generalized to unnormalized distributions. For example, the generalized KL divergence [27] corresponds to f 1 (x) = x log x − (x − 1). The derivations in this paper benefit from employing unnormalized distributions and subsequently imposing the normalization condition as a constraint.
The α-divergence [19,20] is a one-parameter family of f -divergences generated by the α-function f α (x) with α ∈ R. The particular choice of the family of functions f α is motivated by generalization of the natural logarithm [21]. The α-logarithm log α (x) = (x α−1 − 1)/(α − 1) is a power function for α = 1 that turns into the natural logarithm for α → 1. Replacing the natural logarithm in the derivative of the KL divergence f 1 = log x by the α-logarithm and integrating f α under the condition that f α (1) = 0 yields the α-function The α-divergence generalizes the KL divergence, reverse KL divergence, Hellinger distance, Pearson χ 2 -divergence, and Neyman (reverse Pearson) χ 2 -divergence. Figure A1 displays well-known α-divergences as points on the parabola y = α(α − 1). For every divergence, there is a reverse divergence symmetric with respect to the point α = 0.5, corresponding to the Hellinger distance. The convex conjugate of f (x) is defined as f * (y) = sup x∈dom f { y, x − f (x)}, where the angle brackets y, x denote the dot product [51]. The key property ( f * ) = ( f ) −1 relating the derivatives of f * and f yields Table A1, which lists common functions f α together with their convex conjugates and derivatives. In the general case (A1), the convex conjugate and its derivative are given by Function f α is convex, non-negative, and attains minimum at x = 1 with f α (1) = 0. Function ( f * α ) is positive on its domain with ( f * α ) (0) = 1. Function f * α has the property f * α (0) = 0. The linear inequality constraint (A2) on the dom f * α follows from the requirement dom f α = (0, ∞). Another result from convex analysis crucial to our derivations is Fenchel's equality f * (y) + f (x (y)) = y, x (y) , where x (y) = arg sup x∈dom f { y, x − f (x)}. We will occasionally put the conjugation symbol at the bottom, especially for the derivative of the conjugate function f * = ( f * ) .

Appendix B
In all experiments, the temperature parameter η is exponentially decayed η i+1 = η 0 a i in each iteration i = 0, 1, . . . . The choice of η 0 and a depends on the scale of the rewards and the number of samples collected per policy update. Tables for each environment list these parameters along with the number of samples per policy update, the number of policy iteration steps, and the number of runs for averaging the results. Where applicable, environment-specific settings are also listed. (see the Tables A2-A4)