Next Article in Journal
Non-Invasive Systems and Methods Patents Review Based on Electrocardiogram for Diagnosis of Cardiovascular Diseases
Next Article in Special Issue
Outsmarting Human Design in Airline Revenue Management
Previous Article in Journal
An Effective Algorithm for Finding Shortest Paths in Tubular Spaces
Previous Article in Special Issue
Transfer Learning for Operator Selection: A Reinforcement Learning Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Long-Term Visitation Value for Deep Exploration in Sparse-Reward Reinforcement Learning

1
Meta AI Research, 4720 Forbes Avenue, Pittsburgh, PA 15213, USA
2
Department of Informatics, Technische Universität Darmstadt, Hochschulstr. 10, 64289 Darmstadt, Germany
3
Department of Electrical Engineering and Automation, Aalto University, Tietotekniikantalo, Konemiehentie 2, 02150 Espoo, Finland
*
Author to whom correspondence should be addressed.
Algorithms 2022, 15(3), 81; https://doi.org/10.3390/a15030081
Submission received: 29 January 2022 / Revised: 18 February 2022 / Accepted: 22 February 2022 / Published: 28 February 2022
(This article belongs to the Special Issue Reinforcement Learning Algorithms)

Abstract

:
Reinforcement learning with sparse rewards is still an open challenge. Classic methods rely on getting feedback via extrinsic rewards to train the agent, and in situations where this occurs very rarely the agent learns slowly or cannot learn at all. Similarly, if the agent receives also rewards that create suboptimal modes of the objective function, it will likely prematurely stop exploring. More recent methods add auxiliary intrinsic rewards to encourage exploration. However, auxiliary rewards lead to a non-stationary target for the Q-function. In this paper, we present a novel approach that (1) plans exploration actions far into the future by using a long-term visitation count, and (2) decouples exploration and exploitation by learning a separate function assessing the exploration value of the actions. Contrary to existing methods that use models of reward and dynamics, our approach is off-policy and model-free. We further propose new tabular environments for benchmarking exploration in reinforcement learning. Empirical results on classic and novel benchmarks show that the proposed approach outperforms existing methods in environments with sparse rewards, especially in the presence of rewards that create suboptimal modes of the objective function. Results also suggest that our approach scales gracefully with the size of the environment.

1. Introduction

Reinforcement learning (RL) trains agents to behave in an environment by trial and error. The agent performs actions and, in turn, the environment may provide a reward, i.e., feedback assessing the quality of the action. The goal of RL is then to learn a policy, i.e., a function producing a sequence of actions yielding the highest cumulative reward. Despite its simplicity, RL achieved impressive results, such as beating Atari videogames from pixels [1,2], or world-class champions at Chess, Go and Shogi [3]. However, a high-quality reward signal is typically necessary to learn the optimal policy, and without it, RL algorithms may perform poorly even in small environments [4].
Characteristics of high-quality rewards. One important factor that defines the quality of the reward signal is the frequency at which rewards are emitted. Frequently emitted rewards are called “dense”, in contrast to infrequent emissions which are called “sparse”. Since improving the policy relies on getting feedback via rewards, the policy cannot be improved until a reward is obtained. In situations where this occurs very rarely, the agent learns slowly or cannot learn at all. Furthermore, reinforcement signals should encourage the agent to find the best actions to solve the given task. However, the environment may also provide distracting rewards, i.e., rewards that create suboptimal modes of the objective function. In this case, the agent should be able to extensively explore the environment without prematurely converging to locally optimal solutions caused by distracting rewards.
Reward engineering. In the presence of sparse distracting rewards, efficiently exploring the environment to learn the optimal policy is challenging. Therefore, many RL algorithms rely on well-shaped reward functions such as quadratic costs. For example, in a collect-and-deliver task, we could design an additional reward function that rewards proximity to the items to be collected. These well-shaped functions help to guide the agent towards good solutions and avoid bad local optima. However, from the perspective of autonomous learning, this so-called reward engineering is unacceptable for three reasons. First, the commonly utilized reward functions heavily restrict the solution space and may prevent the agent from learning optimal behavior (especially if no solution to the task is known). Second, it is easy to misspecify the reward function and cause unexpected behavior. For instance, by getting excessively high rewards for proximity to the items the agent may never learn to deliver them. Third, reward engineering requires manual tuning of the reward function and the manual addition of constraints. This process requires expertise and the resulting reward function and constraints may not transfer to different tasks or environments.
In this paper, we address the problem of reinforcement learning with sparse rewards without relying on reward engineering or requiring prior knowledge about the environment or the task to solve. In particular, we highlight the problem of learning in the presence of rewards that create suboptimal modes of the objective function, which we call “distractors”.
The exploration-exploitation dilemma. Learning when rewards are sparse and possibly distracting challenges classic RL algorithms. Initially, the agent has no prior knowledge about the environment and needs to explore it searching for feedback. As it gathers rewards, the agent develops a better understanding of the environment and can estimate which actions are “good”, i.e., yield positive rewards, and which are not. Then, at any point, it needs to decide whether it should explore the environment searching for better rewards by executing new actions, or exploit its knowledge and execute known good actions. This “exploration-exploitation dilemma” is frequently addressed naively by dithering [1,5]. In continuous action spaces Gaussian noise is added to the action, while in discrete spaces actions are chosen ϵ -greedily, i.e., optimally with probability 1 ϵ and randomly with probability ϵ . Typically, the initial noise and ϵ are large and then decay over time. These approaches work in environments where random sequences of actions are likely to cause positive rewards. However, if rewards are sparse or are given only for specific sequences of actions, it is very unlikely that the agent will receive any feedback. In this case, the worst-case sample complexity for learning the optimal policy with dithering exploration is exponential in the number of states and actions [6,7,8]. Furthermore, if distractor rewards are easily reachable, the agent may prematurely converge to local optima. An example of this problem is depicted in Figure 1.
Deep exploration. Well-performing exploration strategies should solve long-horizon problems with sparse and possibly distracting rewards in large state-action spaces while remaining computationally tractable. This can be achieved if the agent takes several coherent actions to explore unknown states instead of just locally choosing the most promising actions. These behaviors are usually referred to as “deep exploration” and “myopic exploration”, respectively, [9]. RL literature has a long history of algorithms for deep exploration, even though the term “deep” became popular only in recent years. The first algorithms to guarantee full exploration of tabular environments date back to Kearns and Singh [10] and Brafman and Tennenholtz [11]. These algorithms learn a model of the environment, i.e., the transition and reward functions, and plan actions accordingly to a state-action visitation count to favor less-visited states. Since then, other model-based [12,13] as well as model-free algorithms [14,15,16,17] have been proposed. However, despite their strong guarantees, these algorithms either require learning the complete model of the environment, rely on optimistic initialization, or have impractically high sample complexity.
In continuous environments, intrinsic motivation and bootstrapping are the most prominent approaches. The former defines an additional intrinsic reward added to the environment extrinsic reward. If the extrinsic reward is sparse, the intrinsic reward can fill the gaps between the sparse signals, possibly giving the agent quality feedback at every timestep. This approach can be used in conjunction with any RL algorithm by just providing the modified reward signal. However, the quality of exploration strongly depends on the intrinsic reward, which may not scale gracefully with the extrinsic reward and therefore needs hand-tuning. Furthermore, combining exploration and exploitation by summing intrinsic and extrinsic rewards can result in undesired behavior due to the non-stationarity of the augmented reward function. Moreover, convergence is harder to guarantee [18], and in general, there is no agreement on the definition of the best intrinsic reward [19,20]. Bootstrapping, instead, is used to estimate the actions value posterior distribution over the environment, which then drives exploration. Because actions are selected considering the level of uncertainty associated with their value estimates, bootstrapping incentivizes experimentation with actions of highly uncertain value and, thus, induces exploration [4]. However, these methods rely on approximated posteriors and usually lack guarantees of convergence, unless either the environment model is learned or a time-dependent policy is used [4].
Deep exploration via long-term visitation value.In this paper, we present a novel approach that (1) plans exploration actions far into the future using a long-term visitation count, and (2) decouples exploration and exploitation by learning a separate function assessing the exploration value of the actions. Contrary to existing methods that use models of reward and dynamics, our approach is off-policy and model-free.
We further comprehensively benchmark our approach against existing algorithms on several environments, stressing the challenges of learning with sparse and distracting rewards. Empirical results show that the proposed approach outperforms existing algorithms, and suggest that it scales gracefully with the size of the environment.

2. Preliminaries

We start with a description of the RL framework, providing the reader with the notation used in the remainder of the paper. Subsequently, we review the most prominent exploration strategies in the literature, discuss their shortcomings, and identify open challenges.

2.1. Reinforcement Learning and Markov Decision Processes

We consider RL in an environment governed by a Markov Decision Process (MDP). An MDP is described by the tuple S , A , P , R , μ 1 , where S is the state space, A is the action space, P s | s , a defines a Markovian transition probability density between the current s and the next state s under action a, R s , a is the reward function, and μ 1 is initial distribution for state s 1 . The state and action spaces could be finite, i.e., | S | = d s and | A | = d a , or continuous, i.e., S R d s and A R d a . The former case is often called tabular MDP, as a table can be used to store and query all state-action pairs if the size of the state-action space is not too large. If the model of the environment is known, these MDPs can be solved exactly with dynamic programming. The latter, instead, requires function approximation, and algorithms typically guarantee convergence to local optima. In the remainder of the paper, we will consider only tabular MDPs and model-free RL, i.e., the model of the environment is neither known nor learned.
Given an MDP, we want to learn to act. Formally, we want to find a policy π ( a | s ) to take an appropriate action a when the agent is in state s. By following such a policy starting at initial state s 1 , we obtain a sequence of states, actions and rewards ( s t , a t , r t ) t = 1 H , where r t = R ( s t , a t ) is the reward at time t, and H is the total timesteps also called horizon, which can possibly be infinite. We refer to such sequences as trajectories or episodes. Our goal is thus to find a policy that maximizes the expected return
Q * ( s , a ) : = max π E μ π ( s ) π ( a | s ) Q π ( s , a ) ,
where Q π ( s t , a t ) : = E i = t H π ( a i + 1 | s i + 1 ) P ( s i + 1 | s i , a i ) i = t H γ i t r i ,
where μ π ( s ) is the (discounted) state distribution under π , i.e., the probability of visiting state s under π , and γ [ 0 , 1 ) is the discount factor which assigns weights to rewards at different timesteps. Q π ( s , a ) is the action-state value function (or Q-function) which is the expected return obtained by executing a in state s and then following π afterwards. In tabular MDPs, the Q-function can be stored in a table, commonly referred to as Q-table, and in stationary infinite-horizon MDPs the greedy policy π * ( a | s ) = arg max a Q * ( s , a ) is always optimal [21]. However, the Q-function of the current policy is initially unknown and has to be learned.
Q-learning. For tabular MDPs, Q-learning by Watkins and Dayan [22] was one of the most important breakthroughs in RL, and it is still at the core of recent successful algorithms. At each training step i, the agent chooses an action according to a behavior policy β ( a | s ) , i.e., a t β ( · | s t ) , and then updates the Q-table according to
Q i + 1 π ( s t , a t ) = Q i π ( s t , a t ) + η δ ( s t , a t , s t + 1 )
δ ( s t , a t , s t + 1 ) = r t + γ max a Q i π ( s t + 1 , a ) Q i π ( s t , a t ) if s t is non - terminal r t Q i π ( s t , a t ) otherwise
where η is the learning rate and δ ( s , a , s ) is the temporal difference (TD) error. The blue term is often referred to as TD target. This approach is called off-policy because the policy β ( a | s ) used to explore the environment is different from the target greedy policy π ( a | s ) .
The behavior policy and the “exploration-exploitation dilemma”. As discussed in Section 1, an important challenge in designing the behavior policy is to account for the exploration-exploitation dilemma. Exploration is a long-term endeavor where the agent tries to maximize the possibility of finding better rewards in states not yet visited. On the contrary, exploitation maximizes the expected rewards according to the current knowledge of the environment. Typically, in continuous action spaces Gaussian noise is added to the action, while in discrete spaces actions are chosen ϵ -greedily, i.e., optimally with probability 1 ϵ and randomly with probability ϵ . In this case, the exploration-exploitation dilemma is simply addressed by having larger noise and ϵ at the beginning, and then decaying them over time. This naive dithering exploration is extremely inefficient in MDPs with sparse rewards, especially if some of them are “distractors”, i.e., rewards that create suboptimal modes of the objective function. In the remainder of this section, we review more sophisticated approaches addressing exploration with sparse rewards. We first show theoretically grounded ideas for tabular or small MDP settings which are generally computationally intractable for large MDPs. We then give an overview of methods that try to apply similar principles to large domains by making approximations.

2.2. Related Work

Since finding high-value states is crucial for successful RL, a large body of work has been devoted to efficient exploration in the past years. We begin by discussing an optimal solution to exploration, then continue with confidence-bound-based approaches. Then, we discuss methods based on intrinsic motivation and describe posterior optimality value-distribution-based methods. We conclude by stating the main differences between our approach and these methods.
Optimal solution to exploration. In this paper, we consider model-free RL, i.e., the agent does not know the MDP dynamics. It is well-known that in model-free RL the optimal solution to exploration can be found by a Bayesian approach. First, assign an initial (possibly uninformative) prior distribution over possible unknown MDPs. Then, at each time step, the posterior belief distribution over possible MDPs can be computed using the Bayes’ theorem based on the current prior distribution, executed action, and made an observation. The action that optimally explores is then found by planning over possible future posterior distributions, e.g., using tree search, sufficiently far forward. Unfortunately, optimal exploration is intractable even for very small tasks [7,23,24]. The worst-case computational effort is exponential w.r.t. the planning horizon in tabular MDPs and can be even more challenging with continuous state-action spaces.
Optimism. In tabular MDPs, many of the provably efficient algorithms are based on optimism in the face of uncertainty (OFU) [25]. In OFU, the agent acts greedily w.r.t. an optimistic action value estimate composed of the value estimate and a bonus term that is proportional to the uncertainty of the value estimate. After executing an optimistic action, the agent then either experiences a high reward learning that the value of the action was indeed high, or the agent experiences a low reward and learns that the action was not optimal. After visiting a state-action pair, the exploration bonus is reduced. This approach is superior to naive approaches in that it avoids actions where low value and low information gain are possible. Under the assumption that the agent can visit every state-action pair infinitely many times, the overestimation will decrease and almost optimal behavior can be obtained [6]. Most algorithms are optimal up to a polynomial amount of states, actions, or horizon length. RL literature provides many variations of these algorithms which use bounds with varying efficacy or different simplifying assumptions, such as [6,10,11,12,26,27].
Upper confidence bound. Perhaps the best well-known OFU algorithm is the upper confidence bound (UCB1) algorithm [28]. UCB chooses actions based on a bound computed from visitation counts: the lower the count, the higher the bonus. Recently, Jin et al. [16] proved that episodic Q-learning with UCB exploration converges to a regret proportional to the square root of states, actions, and timesteps; and the square root of the cube of the episode length. Dong et al. [17] gave a similar proof for infinite horizon MDPs with sample complexity proportional to the number of states and actions. In our experiments, we compare our approach to Q-learning with UCB exploration as defined by Auer et al. [28].
Intrinsic motivation. The previously discussed algorithms are often computationally intractable, or their guarantees no longer apply in continuous state-action spaces, or when the state-action spaces are too large. Nevertheless, these provably efficient algorithms inspired more practical ones. Commonly, exploration is conducted by adding a bonus reward, also called auxiliary reward, to interesting states. This is often referred to as intrinsic motivation [29]. For example, the bonus can be similar to the UCB [15,30], thus encouraging actions of high uncertainty of low visitation count. Recent methods, instead, compute the impact of actions based on state embeddings and reward the agent for performing actions changing the environment [31].
Other forms of bonus are based on the prediction error of some quantity. For example, the agent may learn a model of the dynamics and try to predict the next state [19,20,32,33,34]. By giving a bonus proportional to the prediction error, the agent is incentivized to explore unpredictable states. Unfortunately, in the presence of noise, unpredictable states are not necessarily interesting states. Recent work addresses this issue by training an ensemble of models [35] or predicting the output of a random neural network [36].
As we will discuss in Section 3.2, our approach has similarities with the use of auxiliary reward. However, the use of auxiliary rewards can be inefficient for long-horizon exploration. Our approach addresses this issue by using a decoupled long-horizon exploration policy and, as the experiments in Section 4 show, it outperforms auxiliary-reward-based approaches [30].
Thompson sampling. Another principled well-known technique for exploration is Thompson sampling [37]. Thompson sampling samples actions from a posterior distribution which specifies the probability for each action to be optimal. Similarly to UCB-based methods, Thompson sampling is guaranteed to converge to an optimal policy in multi-armed bandit problems [38,39], and has shown strong empirical performance [40,41]. For a discussion of known shortcomings of Thompson sampling, we refer to [42,43,44].
Inspired by these successes, recent methods have followed the principle of Thompson sampling in Q-learning [4,45,46]. These methods assume that an empirical distribution over Q-functions—an ensemble of randomized Q-functions—is similar to the distribution over action optimalities used in Thompson sampling. Actions are thus sampled from such ensemble. Since the Q-functions in the ensemble become similar when updated with new samples, there is a conceptual similarity with the action optimality distribution used in Thompson sampling for which variance also decreases with new samples, while there are no general proofs for randomized Q-function approaches, Osband et al. [4] proved a bound on the Bayesian regret of an algorithm based on randomized Q-functions in tabular time-inhomogeneous MDPs with a transition kernel drawn from a Dirichlet prior, providing a starting point for more general proofs.
Other exploration methods approximating a distribution over action optimalities include for example the work of Fortunato et al. [47] and Plappert et al. [48], which apply noise to the parameters of the model. This may be interpreted as approximately inferring a posterior distribution [49], although this is not without contention [50]. Osband et al. [9] more directly approximates the posterior over Q-functions through bootstrapping [51]. The lack of a proper prior is an issue when rewards are sparse, as it causes the uncertainty of the posterior to vanish quickly. Osband et al. [52] and Osband et al. [4] try to address this by enforcing a prior via regularization or by adding randomly initialized but fixed neural network on top of each ensemble member, respectively.
Methods based on posterior distributions have yielded high performance in some tasks. However, because of the strong approximations needed to model posterior optimality distributions instead of maintaining visitation counts or explicit uncertainty estimates, these approaches may often have problems in assessing the state uncertainty correctly. In our experiments, the proposed approach outperforms state-of-the-art methods based on approximate posterior distributions, namely the algorithms proposed by Osband et al. [4,9] and D’Eramo et al. [46].
Above, we discussed state-of-the-art exploration methods that (1) use confidence bounds or distribution over action optimalities to take the exploration into account in a principled way, (2) modify the reward by adding an intrinsic reward signal to encourage exploration, and (3) approximate a posterior distribution over value functions. The main challenge that pertains to all these methods is that they do not take long-term exploration into account explicitly. Contrary to this, similarly to how value iteration propagates state values, we propose an approach that uses dynamic programming to explicitly propagate visitation information backward in time. This allows our approach to efficiently find the most promising parts of the state space that have not been visited before, and avoid getting stuck in suboptimal regions that at first appear promising.

3. Long-Term Visitation Value for Deep Exploration

In this section, we present our novel off-policy method for improving exploration with sparse and distracting rewards. We start by motivating our approach with a toy example, showing the limitation of current count-based algorithms and the need for “long-term visitation counts”. Subsequently, we formally define the proposed method and show how it solves the toy example.

3.1. Example of the Limitations of Existing Immediate Count Methods

Consider the toy problem of an agent moving in a grid, shown in Figure 2. Reward is 0 everywhere except in ( 1 , 1 ) and ( 3 , 1 ) , where it is 1 and 2, respectively. The reward is given for executing an action in the state, i.e., on transitions from reward states. The agent’s initial position is fixed in ( 1 , 3 ) and the episode ends when a reward is collected or after five steps. Any action in the prison cells ( 2 , 2 ) has only an arbitrarily small probability of success. If it fails, the agent stays in ( 2 , 2 ) . In practice, the prison cell almost completely prevents any further exploration. Despite its simplicity, this domain highlights two important challenges for exploration in RL with sparse rewards. First, there is a locally optimal policy collecting the lesser reward, which acts as a distractor. Second, the prison state is particularly dangerous as it severely hinders exploration.
Assume that the agent has already explored the environment for four episodes performing the following actions: (1) left, down, left, down, down (reward 1 collected); (2) up, down, right, right (fail), right (fail); (3) right, left, right, down, down (fail); (4) right, down, left (fail), up (fail), right (fail). The resulting state-action visitation count is shown in Figure 3a. Given this data, we initialize Q π ( s , a ) to zero, train it with Q-learning with γ = 0.99 until convergence, and then derive three greedy policies. The first simply maximizes Q π ( s , a ) . The second maximizes the behavior Q β ( s , a ) based on UCB1 [28], defined as
Q β ( s , a ) = Q π ( s , a ) + κ 2 log a j n ( s t , a j ) n ( s t , a ) ,
where n ( s , a ) is the state-action count, and κ is a scaling coefficient. Note that in classic UCB, the reward is usually bounded in [ 0 , 1 ] and no coefficient is needed. This is not usually the case for MDPs, where Q-values can be larger/smaller than 1/0. We thus need to scale the square root with κ = ( r max r min ) / ( 1 γ ) , which upper-bounds the difference between the largest and smallest possible Q-value). The third maximizes an augmented Q-function Q + ( s , a ) trained by adding the intrinsic reward based on the immediate count [30], i.e.,
r t + = r t + α n ( s t , a t ) 1 / 2 ,
where α is a scaling coefficient set to α = 0.1 as in [15,30]. Figure 3a shows the state-action count after the above trajectories, and Figure 3b–d depict the respective Q-functions (colormap) and policies (arrows). Given the state-action count, consider what the agent has done and knows so far.
  • It has executed all actions at least once in (1,3) and (2,2) (the prison cell),
  • It still did not execute “up” in (1,2),
  • It visited (1,3) once and experienced the reward of 1, and
  • It still did not execute “up” and “right” in (2,1).
In terms of exploration, the best decision would be to select new actions, such as “up” in (1,2). However, what should the agent do in states where it has already executed all actions at least once? Ideally, it should select actions driving it to unvisited states, or to states where it has not executed all actions yet. However, none of the policies above exhibits such behavior.
  • The greedy policy (Figure 3b) points to the only reward collected so far.
  • The UCB1 policy (Figure 3c) yields Q-values of much larger magnitude. Most of the actions have been executed a few times, thus their UCB is larger than their Q-value. The policy correctly selects unexecuted actions in (1,2) and (2,3) but fails (1,3). There, it also selects the least executed actions (“up” and “left”) which, however, have already been executed once. The agent should know that these actions do not bring it anywhere (the agent hits the environment boundaries and does not move) and should avoid them. However, it selects them because the policy Q β ( s , a ) is based on the immediate visitation count. Only when the count will be equal for all actions, the policy will select a random action. This behavior is clearly myopic and extremely inefficient.
  • The policy learned using the auxiliary bonus (Figure 3d) is even worse, as it acts badly not only in (1,3) but also in (1,2) and (2,3). There, it chooses “left”, which was already selected once, instead of “up” (or “right in (2,3)), which instead has not been selected yet. The reason for such behavior is that this auxiliary reward requires optimistic initialization [30]. The agent, in fact, is positively rewarded for simply executing any action, with value proportional to the visitation count. However, if the Q-function is initialized to zero, the positive feedback will make the agent believe that the action is good. This mechanism encourages the agent to execute the same action again and hinders exploration. This problem is typical of all methods using an auxiliary reward based on visitation count unless optimistic initialization of the Q-table is used [30]. More recent methods use more sophisticated rewards [16,17]. However, despite their strong theoretical convergence guarantees, for large-size problems, these methods may require an impractical number of samples and are often outperformed by standard algorithms.

3.2. The Long-Term Visitation Value Approach

Despite its simplicity, the above toy problem highlights the limitations of current exploration strategies based on the immediate count. More specifically, (1) considering the immediate count may result in shallow exploration, and (2) directly augmenting the reward to train the Q-function, i.e., coupling exploration and exploitation, can have undesired effects due to the non-stationarity of the augmented reward function. To address these limitations, we propose an approach that (1) plans exploration actions far into the future by using a long-term visitation count, and (2) decouples exploration and exploitation by learning a separate function assessing the exploration value of the actions. Contrary to existing methods that use models of reward and dynamics [13], which can be hard to come by, our approach is off-policy and model-free.
Visitation value. The key idea of our approach is to train a visitation value function using the temporal difference principle. The function, called W-function and denoted by W β ( s , a ) , approximates the cumulative sum of an intrinsic reward r W based on the visitation count, i.e.,
W β ( s t , a t ) : = E i = t β ( a t + 1 | s t + 1 ) P ( s i + 1 | s i , a i ) i = t H γ w i t r i W ,
where γ w is the visitation count discount factor. The higher the discount, the more in the future the visitation count will look. In practice, we estimate W β ( s , a ) using TD learning on the samples produced using the behavior policy β . Therefore, we use the following update
W i + 1 β ( s t , a t ) = W i β ( s t , a t ) + η δ W ( s t , a t , s t + 1 ) ,
where η is the learning rate, and δ W ( s t , a t , s t + 1 ) is the W-function TD error, which depends on the reward r W (can be either maximized or minimized). Subsequently, following the well-known upper confidence bound (UCB) algorithm [25], the behavior policy is greedy w.r.t. the sum of the Q-function and upper confidence bound, i.e.,
β ( a | s ) = arg max a Q π ( s , a ) + κ U W ( s , a ) ,
where U W ( s , a ) is an upper confidence bound based on W β ( s , a ) , and κ is the same exploration constant used by UCB1 in Equation (5). This proposed approach (1) considers long-term visitation count by employing the W-function, which is trained to maximize the cumulative—not the immediate—count. Intuitively, the W-function encourages the agent to explore state-actions that have not been visited before. Given the current state s, in fact, W β ( s , a ) specifies how much exploration we can expect on (discount-weighted) average in future states if we choose action a. Furthermore, since the Q-function is still trained greedily w.r.t. the extrinsic reward as in Q-learning, while the W-function is trained greedily on the intrinsic reward and favors less-visited states, our approach (2) effectively decouples exploration and exploitation. This allows us to more efficiently prevent underestimation of the visitation count exploration term. Below, we propose two versions of this approach based on two different rewards r W and upper bounds U W ( s , a ) .
W-function as long-term UCB. The visitation reward is the UCB of the state-action count at the current time, i.e.,
r t W = 2 log a j n ( s t , a j ) n ( s t , a t ) if s t is non - terminal 1 1 γ w 2 log a j n ( s t , a j ) n ( s t , a t ) otherwise .
This W-function represents the discount-weighted average UCB along the trajectory, i.e.,
W UCB β ( s t , a t ) = i = t H γ w i t 2 log a j n ( s i , a j ) n ( s i , a i ) ,
and its TD error is
δ W ( s t , a t , s t + 1 ) = r t W + γ w max a W UCB , i β ( s t + 1 , a ) W UCB , i β ( s t , a t ) if s t is non - terminal r t W W UCB , i β ( s t , a t ) otherwise .
Notice how we distinguish between terminal and non-terminal states for the reward in Equation (10). The visitation value of terminal states, in fact, is equal to the visitation reward alone, as denoted by the TD target in Equation (10). Consequently, the visitation value of terminal states could be significantly smaller than the one of other states, and the agent would not visit them again after the first time. This, however, may be detrimental for learning the Q-function, especially if terminal states yield high rewards. For this reason, we assume that the agent stays in terminal states forever, getting the same visitation reward at each time step. The limit of the sum of the discounted reward is the reward in Equation (10). Furthermore, given proper initialization of W UCB β and under some assumptions, the target of non-terminal states will still be higher than the one of terminal states despite the different visitation reward, if the count of the former is smaller. We will discuss this in more detail later in this section.
Another key aspect of this W-function regards the well-known overestimation bias of TD learning. Since we can only update the W-function approximately based on limited samples, it is beneficial to overestimate its target with the max operator, while it is known that in highly stochastic environments the overestimation can degrade the performance of value-based algorithm [53,54], it has been shown that underestimation does not perform well when the exploration is challenging [55], and for a wide class of environments, overestimation allows finding the optimal solution due to self-correction [56].
Assuming identical visitation over state-action pairs, the W-function becomes the discounted immediate UCB, i.e.,
W UCB β ( s t , a t ) = 1 1 γ w 2 log a j n ( s t , a j ) n ( s t , a t ) ,
thus, we can use W UCB β ( s , a ) directly in the behavior policy
β ( a t | s t ) = arg max a { Q π ( s t , a ) + κ ( 1 γ w ) W UCB β ( s t , a ) U W UCB ( s , a ) } ,
In Section 3.3, we discuss how to initialize W UCB β ( s , a ) . Finally, notice that when γ w = 0 , Equation (13) is the classic UCB, and Equation (14) is equivalent to the UCB1 policy in Equation (5).
W-function as long-term count. The visitation reward is the state-action count at the current time, i.e.,
r t W = n ( s t , a t ) if s t is non - terminal 1 1 γ w n ( s t , a t ) otherwise .
This W-function represents the discount-weighted average count along the trajectory, i.e.,
W N β ( s t , a t ) = i = t H γ w i t n ( s i , a i ) ,
and its TD error is
δ W ( s t , a t , s t + 1 ) = r t W + γ w min a W N , i β ( s t + 1 , a ) W N , i β ( s t , a t ) if s t is non - terminal r t W W N , i β ( s t , a t ) otherwise .
Once again we distinguish between terminal and non-terminal states, but in the TD error the max operator has been replaced by the min operator. Contrary to the previous case, in fact, the lower the W-function the better, as we want to incentivize the visit of lower-count states. If we would use only the immediate count as reward for terminal states, the agent would be constantly driven there. Therefore, similarly to the UCB-based reward, we assume that the agent stays in terminal states forever getting the same reward.
In the TD target, since we want to prioritize low-count state-action pairs, the min operator yields an optimistic estimate of the W-function, because the lower the W-function the more optimistic the exploration is. Assuming identical visitation over state-action pairs, the W-function becomes the discounted cumulative count, i.e.,
W N β ( s t , a t ) = 1 1 γ w n ( s t , a t ) .
We then compute the pseudocount n ^ ( s t , a t ) = ( 1 γ w ) W N β ( s t , a t ) and use it in the UCB policy
β ( a t | s t ) = arg max a { Q π ( s t , a ) + κ 2 log a j n ^ ( s t , a j ) n ^ ( s t , a ) U W N ( s , a ) } .
When the counts are zero, W N β ( s , a ) = 0 , thus we need to initialize it to zero. We discuss how to avoid numerical problem in the square root in Section 3.3. Finally notice that when γ w = 0 , Equation (19) is equivalent to the UCB1 policy in Equation (5).
Example of Long-Term Exploration with the W-function. Consider again the toy problem presented in Section 3.1. Figure 4 shows the behavior policies of Equation (14) and Equation (19) with γ w = 0.99 . The policies are identical and both drive the agent to unvisited states, or to states where it has not executed all actions yet, successfully exhibiting long-term exploration. The only difference is the magnitude of the state-action pair values (the colormap, especially in unvisited states) due to their different initialization.
  • In (1,3), both select “down” despite “up” and “left” having lower count. The agent, in fact, knows that from (1,2) it can go “down” to (1,1) where it executed only one action. Thus, it knows that there it can try new actions. On the other hand, the only thing it knows about (2,3) is that it leads to the prison state, where the agent already executed all actions. Thus, (1,2) has a higher long-term exploration value than (2,3).
  • In (1,2) both select “up”, overturning the greedy action “down” (Figure 3b). This is crucial because to fully explore the environment we need to select new actions.
  • The same happens in (2,3), where actions with count zero are selected, rather than the greedy ones.
  • None of the policies lead the agent to the prison state, because it has already been visited and all actions have been executed in there.
  • They both do not select “right” in (2,2) because it has been executed one more time than other actions. Nonetheless, the difference in the action value is minimal, as shown by the action colormap (almost uniform).

3.3. W-Function Initialization and Bounds

In UCB bandits, it is assumed that each arm/action will be executed once before actions are executed twice. In order to enforce this, we need to make sure that the upper confidence bound U W ( s , a ) is high enough so that an action cannot be executed twice before all other actions are executed once. Below, we discuss how to initialize the W-functions and set bounds in order to achieve the same behavior. For the sake of simplicity, below we drop subscripts and superscripts, i.e., we write Q ( s , a ) in place of Q π ( s , a ) , and W ( s , a ) in place of W UCB β ( s , a ) or W N β ( s , a ) .
W UCB β upper bound. To compute the initialization value we consider the worst-case scenario, that is, in state s all actions a but a ¯ have been executed. In this scenario, we want that W ( s , a ¯ ) is an upper bound of W ( s , a ) . Following Equations (12) and (14), we have
Q max + κ ( 1 γ w ) W ( s , a ¯ ) = Q max + κ ( 1 γ w ) 2 log ( | A | 1 ) + γ w max a W ( s , a ) ,
where the blue term is the TD target for non-terminal states, and 2 log ( | A | 1 ) is the immediate visitation reward for executing a ¯ in the worst-case scenario (all other | A | 1 actions have been executed, and a ^ has not been executed yet). In this scenario, W ( s , a ¯ ) is an upper bound of W ( s , a ) under the assumption of uniform count, i.e.,
( s , a ) ( s , a ¯ ) , n ( s , a ) = n ¯ ( s ) W ( s , a ¯ ) W ( s , a ) .
We can then write Equation (20) as
Q max + κ ( 1 γ w ) W ( s , a ¯ ) = Q max + κ ( 1 γ w ) ( 2 log ( | A | 1 ) + γ w W ( s , a ¯ ) ) .
In order to guarantee to select a ¯ , which has not been select yet, we need to initialize W ( s , a ¯ ) such that the following is satisfied
Q min + κ ( 1 γ w ) W ( s , a ¯ ) > Q max + κ ( 1 γ w ) ( 2 log ( | A | 1 ) + γ w W ( s , a ¯ ) ) ,
where Q min denotes the smallest possible Q-value. We get
W ( s , a ¯ ) > Q max Q min κ ( 1 γ w ) + 2 log ( | A | 1 ) + γ w W ( s , a ¯ ) W ( s , a ¯ ) > Q max Q min κ ( 1 γ w ) 2 + 2 log ( | A | 1 ) ( 1 γ w ) .
Initializing the W-function according to Equation (24) guarantees to select a ¯ .
W N β upper bound. Since this W-function represents the discounted cumulative count, it is initialized to zero. However, numerical problems may occur in the square root of Equation (19) because the pseudocount can be smaller than one or even zero. In these cases, the square root of negative numbers and the division by zero are not defined. To address these issues, we add +1 inside the logarithm and replace the square root with an upper bound when n ^ ( s , a ) = 0 . Similarly to the previous section, the bound needs to be high enough so that an action cannot be executed twice before all other actions are executed once. Following Equation (19), we need to ensure that
Q min + κ U W ( s , a ¯ ) > Q max + κ U W ( s , a ) , U W ( s , a ¯ ) > Q max Q min κ + U W ( s , a ) .
However, here assuming uniform count does not guarantee a uniform pseudocount, i.e.,
( s , a ) ( s , a ¯ ) , n ( s , a ) = n ¯ ( s ) W ( s , a ¯ ) W ( s , a ) .
To illustrate this issue, consider the chain MDP in Figure 5. The agent always starts in A and can move left, right, and down. The episode ends after eight steps, or when terminal state E is reached. The goal of this example is to explore all states uniformly. In the first episode, the agent follows the blue trajectory ( A , D , E ). In the second episode, it follows the red trajectory ( A , C , D , D , C , B , B , B ). In both trajectories, an action with count zero was always selected first. At the beginning of the third episode, all state-action pairs have been executed once, except for “left” in A, i.e., n ( s , a ) = 1 ( s , a ) ( A , l e f t ) , n ( A , l e f t ) = 0 . Thus, we need to enforce the agent to select it. However, the discounted cumulative count n ^ ( s , a ) = ( 1 γ w ) W ( s , a ) is not uniform. For example, with γ w = 0.9 , n ^ ( A , r i g h t ) = 3.8 and n ^ ( A , d o w n ) = 4.52 .
More in general, the worst-case scenario for satisfying Equation (25) is the one where (1) action a ¯ has not been executed yet, (2) an action a ˙ has been executed once and has the smallest pseudocount, and (3) all other actions have been executed once and have the highest pseudocount. In this scenario, to select a ¯ we need to guarantee U W ( s , a ¯ ) > U W ( s , a ˙ ) . Assuming that counts are uniform and no action has been executed twice, the smallest pseudocount is n ^ ( s , a ˙ ) = 1 γ w , while the highest is 1. Then, Equation (25) becomes
U W ( s , a ¯ ) > Q max Q min κ + U W ( s , a ˙ ) U W ( s , a ¯ ) > Q max Q min κ + 2 log ( 1 γ w ) + | A | 2 1 γ w .
Replacing U W ( s , a ) in Equation (19) with Equation (27) when n ( s , a ¯ ) = 0 guarantees to select a ¯ .
Example of exploration behavior under uniform count. Consider our toy problem again. This time, all state-action pairs have been executed once except for “left” in (1,3). The Q-table is optimal for visited state-action pairs, thus the greedy policy simply brings the agent to (3,1) where the highest reward lies. Figure 6 shows baseline and proposed behavior policies in two scenarios, depending on whether reward states are terminal (upper row) or not (bottom row).
  • The augmented reward policy (Figure 6b) has no interest in trying “left” in (1,3), and its value is the lowest. The agent, in fact, cannot get the visitation bonus without first executing the action. This shows once more that this passive way of exploration is highly inefficient.
  • UCB1 policy (Figure 6c) acts greedily everywhere except in (1,3), where it correctly selects the unexecuted action. However, since UCB1 considers only the immediate visitation, it has no way of propagating this information to other state-action pairs.
  • By contrast, for the proposed W-function policies (Figure 6d,e) the unexecuted state-action pair is the most interesting, and, from any state, the policy brings the agent there. Thanks to the separate W-function, we can propagate the value of unexecuted actions to other states, always prioritizing the exploration of new state-action pairs. Under uniform count n ( s , a ) = 1 ( s , a ) , then the agent acts greedily w.r.t. the Q-function.
Finally, if the count is uniform for all state-action pairs, including for “left” in (1,3), then all policies look like the greedy one (Figure 6a).

3.4. Final Remarks

In this section, we discuss the uniform count assumption used to derive the bounds, the differences between the proposed W-functions and intrinsic motivation, the benefit of using the count-based one, and potential issues in stochastic environments and continuous MDPs.
Uniform count assumption. In Section 3.3, we derived an initialization for W UCB β and a bound for W N β to guarantee that an action is not executed twice before all actions are executed once. For that, we assumed the visitation count to be uniform for all state-action pairs. We acknowledge that this assumption is not true in practice. First, the agent may need to revisit old states in order to visit new ones. Second, when an episode is over the state is reset and a new episode starts. This is common practice also for infinite horizon MDPs when usually the environment is reset after some steps. Thus, the agent’s initial state will naturally have a higher visitation count. Nonetheless, in Section 4.2.1, we empirically show that our approach allows the agent to explore the environment as uniformly as possible.
W-function vs. auxiliary rewards. Using an auxiliary reward such as an exploration bonus or an intrinsic motivation signal has similarities with the proposed approach, especially with W UCB β , but the two methods are not equivalent. Consider augmenting the reward with the same visitation reward used to train the W-function, i.e., r t + = r t + α r t W . Using TD learning, the augmented Q-function used for exploration can be rewritten as
Q + ( s t , a t ) = r t + α r t W + γ max a Q + ( s t + 1 , a ) .
Then consider the behavior policy derived from W UCB β in Equation (14), and decompose it
Q π ( s t , a t ) + α W UCB β ( s t , a t ) = r t + γ max a Q π ( s t + 1 , a ) + α r t W + α γ w max a W UCB β ( s t + 1 , a ) ,
where we considered α = κ ( 1 γ w ) . Red terms are equivalent, but blue terms are not because of the max operator, since max x { f ( x ) + g ( x ) } max x { f ( x ) } + max x { g ( x ) } .
This explicit separation between exploitation (Q) and exploration (W) has two important consequences. First, it is easy to implement the proposed behavior policy for any off-policy algorithm, as the exploration term is separate and independent from the Q-function estimates. This implies that we can propagate the exploration value independently from the action-value function. With this choice, we can select a discount factor γ w that differs from the MDP discount γ . By choosing a high exploration discount factor γ w , we do long-term exploration allowing us to find the optimal policy when exploration is crucial. By choosing a low exploration discount factor γ w , instead, we perform short-term exploration which may converge faster to a greedy policy. In the evaluation in Section 4.2, we show experiments for which the discount factors are the same, as well as experiments where they differ. Furthermore, we investigate the effect of changing the W-function discount while keeping the Q-function discount fixed.
Second, auxiliary rewards are non-stationary, as the visitation count changes at every step. This leads to a non-stationary target for the Q-function. With our approach, by decoupling exploration and exploitation, we have a stationary target for the Q-function while moving all the non-stationarity to the W-function.
Terminal states and stochasticity. The visitation rewards for training the W-functions penalize terminal state-action pairs (Equations (10) and (15)). Thus, the agent will avoid visiting them more than once. One may think this would lead to poor exploration in stochastic environments, where the same state-action pair can lead to either terminal or non-terminal states. In this case, trying the same action again in the future may yield better exploration depending on the stochastic transition function. Yet, in Section 4.2.5 we empirically show that stochasticity in the transition function does not harm our approach.
Pseudo-counts and continuous MDPs. In the formulation of our method—and later in the experiments section—we have considered only tabular MDPs, i.e., with discrete states and actions. This makes it easy to exactly count state-action pairs and accurately compute visitation rewards and value functions. We acknowledge that the use of counts may be limited to tabular MDPs, as real-world problems are often characterized by continuous states and actions. Nonetheless, pseudo-counts [57] have shown competitive performance in continuous MDPs and can be used in place of exact counts without any change to our method. Alternatively, it is possible to replace counts with density models [58] with little modifications to our method. Finally, if neural networks would be used to learn the visitation-value function W ( s , a , s ) , it may be necessary to introduce target networks similarly to deep Q-networks [1].

4. Evaluation

In this section, we evaluate the proposed method against state-of-the-art methods for exploration in model-free RL on several domains. The experiments are designed to highlight the challenges of learning with sparse rewards and are split into two parts. In the first, we present the environments and we address (1) learning when only a few states give a reward, (2) learning when a sequence of correct actions is needed to receive a reward, (3) exploring efficiently when few steps are allowed per training episode, (4) avoiding local optima and distracting rewards, and (5) avoiding dangerous states that stop the agent preventing exploration. In the second, we further investigate (6) the visitation count at convergence, (7) the empirical sample complexity, (8) the impact of the visitation value hyperparameter γ w , (9) the performance in the infinite horizon scenario, and (10) with stochastic transition function.
To ensure reproducibility, the experiments are performed over 20 and 50 fixed random seeds, for deterministic and stochastic MDPs, respectively. Pseudocode and further details of the hyperparameters are given in Appendix A. The source code is available at https://github.com/sparisi/visit-value-explore, accessed on 20 January 2022.
Baseline Algorithms. The evaluated algorithms all use Q-learning [22] and share the same hyperparameters, e.g., learning rate and discount factor. In most of the experiments, we use infinite replay memory as presented by Osband et al. [4]. In Appendix A.6 we also present an evaluation without replay memory on some domains. The algorithms all learn two separate Q-tables, a behavior one for exploration, and a target one for testing the quality of the greedy policy. The difference among the algorithms is how the behavior policy performs exploration. In this regard, we compare against the following exploration strategies: random, ϵ -greedy, bootstrapped, count-based.
For bootstrapping, we evaluate the original version proposed by Osband et al. [9], as well as the more recent versions of D’Eramo et al. [46] and Osband et al. [4]. These methods share the same core idea, i.e., to use an ensemble of Q-tables, each initialized differently and trained on different samples, to approximate a distribution over Q-values via bootstrapping. The differences are the following. Osband et al. [9] select a random Q-table at the beginning of the episode and use it until the episode is over. Osband et al. [4] further regularize the Q-tables using the squared 2 -norm to “prior tables” representing a prior over the Q-function. D’Eramo et al. [46] select the Q-table randomly within the ensemble at each step to approximate Thompson sampling, but without any prior.
For count-based algorithms, we compare against UCB1 exploration [28] on the immediate state-action count as in Equation (5), and against augmenting the reward with the exploration bonus proposed by Strehl and Littman [30] as in Equation (6). Notice that Strehl and Littman [30] apply the bonus to a model-based algorithm, but since then it has been used by many model-free algorithms [15,17]. Some recent methods use other bonuses to derive strong convergence guarantees [16,17]. However, for large-size problems, these methods may require an impractical number of samples and are outperformed by standard algorithms.

4.1. Part 1: Discounted Return and States Discovered

We use the following environments: deep sea [4,52], taxi driver [59], and four novel benchmark gridworlds. All the environments have sparse discounted rewards, and each has peculiar characteristics making it challenging to solve.
Evaluation Criteria. For each environment, we evaluate the algorithms on varying two different conditions: optimistic vs. zero initialization of the behavior Q-tables, and short vs. long horizon, for a total of four scenarios. For each of them, we report the expected discounted return of π ( a | s ) , which is greedy over Q π ( s , a ) , and the number of visited states over training samples. In all environments, the initial state is fixed.
Optimistic initialization [25] consists of initializing the Q-tables to a large value. After visiting a state-action pair, either the agent experiences a high reward confirming their belief about its quality, or the agent experiences a low reward and learns that the action was not optimal and will not execute it again in that state. By initializing the Q-tables to the upper bound r max / ( 1 γ ) we are guaranteed to explore all the state-action pairs even with just an ϵ -greedy policy. However, this requires prior knowledge of the reward bounds, and in the case of continuous states and function approximation the notion of a universally “optimistic prior” does not carry over from the tabular setting [4]. Thus, it is important that an algorithm is able to explore and learn even with a simple zero or random initialization.
The length of the horizon is also an important factor for exploration. If the agent can perform a limited number of actions during an episode, it has to carefully choose them in order to explore as efficiently as possible. Prominent examples in this regard are real-world tasks such as videogames, where the player has limited time to fulfill the objective before the game is over, or robotics, where the autonomy of the robot is limited by the hardware. Thus, it is important that an algorithm is able to explore and learn even when the training episode horizon is short. In the “short horizon” scenario, we allow the agent to explore for a number of steps slightly higher than the ones needed to find the highest reward. For example, if the highest reward is eight steps away from the initial position, the “short horizon” is ten steps. The “long horizon” scenario is instead always twice as long.
Results are first presented in plots showing the number of states discovered and discounted return against training steps averaged over 20 seeds. Due to the large number of evaluated algorithms, confidence intervals are not reported in plots. At the end of the section, a recap table summarizes the results and reports the success of an algorithm, i.e., the number of seeds the algorithm learned the optimal policy, and the percentage of states visited at the end of the learning, both with 95% confidence interval.

4.1.1. Deep Sea

In this environment (Figure 7), the agent (the ship) always starts at the top-left corner of a gridworld of size N × N , and has to descend through the grid to a chest in the bottom-right corner. At the beginning of the episode, either a treasure or a bomb appear inside the chest, with a 50-50 chance. The outcome is known to the agent, as the state consists of the ship position and a flag denoting if the bomb or the treasure has appeared. At every step, the ship descends by one step and must choose to turn left or right. If it hits the left boundary, it will descend without turning. For example, from the starting position ( 1 , 1 ) , doing “left” will move the ship to ( 2 , 1 ) , while doing “right” will move it to ( 2 , 2 ) . The episode ends when the ship reaches the bottom of the grid, and thus the horizon is fixed to N. The reward is 1 if the ship finds the treasure, 1 if it finds the bomb, and 0.01 / N if the ship goes “right” in any cell along the diagonal (denoted by sea waves). The optimal policy selects always “right” if there is the treasure, otherwise “left” at the initial state and then can act randomly.
The challenge of this environment lies in having to select the same action (“right”) N times in a row in order to receive either 1 or 1 . Doing “left” even only once prevents the ship from reaching the chest. At the same time, the agent is discouraged from doing “right”, since doing so along the diagonal yields an immediate penalty. Consequently, an ϵ -greedy policy without optimistic initialization would perform poorly, as the only short-term reward it receives would discourage it from doing “right”. A random policy is also highly unlikely to reach the chest. In particular, the probability such a policy reaches it in any given episode is ( 1 / 2 ) N . Thus, the expected number of episodes before observing either the bomb or the treasure is 2 N . Even for a moderate value of N = 50 , this is an impractical number of episodes [4].
Results.Figure 8 shows the results on a grid of depth N = 50 . Since this domain has a fixed horizon, the evaluation does not include the “short vs. long horizon” comparison. With zero initialization, only our algorithms and bootstrapping by Osband et al. [9] and Osband et al. [4] discover all states and converge to the optimal policy within steps limit. Both proposed visitation-count-based (blue and orange) outperform the other two, converging twice or more as faster. Bootstrapped exploration with random prior (green) comes second, reproducing the results reported by Osband et al. [52] and Osband et al. [4]. Results with optimistic initialization are similar, with the main difference being that also approximate Thompson sampling by D’Eramo et al. [46] (purple) converges. Bootstrapping by Osband et al. [52] (green) matches the performance of visitation-count-based with UCB (blue), and the two lines almost overlap. For more details about the percentage of successful trials and the percentage of states discovered by each algorithm with confidence intervals, we refer to Table 1.
This evaluation shows that the proposed exploration outperforms state-of-the-art bootstrapping on a domain with a fixed horizon and no local optima. In the next sections, we investigate domains with absorbing states, i.e., that can prematurely end the episode, and distractor reward. In Section 4.2.3, we also present an empirical study on sample complexity on varying the deep sea depth N. Furthermore, in Section 4.2.1, we show the state visitation count at the end of the learning for the deep sea and other domains, discussing why the proposed visitation-value-based exploration achieves new state-of-the-art results.

4.1.2. Multi-Passenger Taxi Driver

In this environment (Figure 9), the agent (taxi) starts at the top-left corner of a grid and has to pick up three persons and drop them off at a goal (flag). It can carry multiple passengers at the same time. If one, two, or all three passengers reach the goal, the agent is rewarded with 1, 3, or 15, respectively. Otherwise, the reward is 0. The state consists of the taxi position and three booleans denoting if a passenger is in the taxi. The agent can move left, right, up, or down. Black cells cannot be visited. An episode ends if the taxi goes to the flag, even without passengers. The optimal policy picks up all passengers and drops them off in 29 steps (28 steps to pick all passengers and reach the goal, plus an additional action to get the reward).
This domain is challenging because, first, it admits many locally optimal policies, depending on how many passengers reach the goal. Second, with a short horizon learning to pick up and drop off all passengers is difficult and requires good exploration.
Results. In this domain, the “short horizon” consists of 33 steps, while the “long” of 66. Note that our algorithms could learn also with a horizon of only 29 steps. However, with zero initialization other algorithms performed extremely poorly. We thus decided to give the agent 33 steps. As shown in Figure 10, bootstrapped algorithms perform significantly worse than before. None of them learned to pick all passengers if Q-tables are initialized to zero, even with “long horizon”. In particular, random prior bootstrapping (green) converged prematurely due to the presence of local optima (the algorithm does not discover new states after 750–1000 steps). The proposed algorithms (blue and orange), instead, perform extremely well, quickly discovering all states and then learning the optimal policy. Other algorithms learned to pick one or two passengers, and only the auxiliary visitation bonus (pink) learned to pick all passengers in some seeds but only with a long horizon. With optimistic initialization, most of the algorithms match the performance of our proposed ones. Unsurprisingly, ϵ -greedy (gray) learns slowly. Bonus-based exploration (pink) is also slow, either because the small bonus coefficient (see Equation (6)), or due to the issues discussed in Section 3.1. Only random exploration (light green) still cannot pick all passengers. For the percentage of successful trials and the percentage of states discovered by each algorithm with confidence intervals, we refer to Table 1.
This evaluation shows that the proposed exploration outperforms existing algorithms in the presence of local optima and that its performance is not affected by the length of the horizon. Next, we investigate what happens when we combine the challenges of the deep sea and the taxi.

4.1.3. Deep Gridworld

The first gridworld we propose is inspired by the deep sea domain. In this environment (Figure 11), the agent navigates in a 5 × 11 grid by moving left, right, up, or down. All states can be visited, but the blue corridor can be entered only from its left side. The agent can exit the corridor anytime by moving either “up” or “down”. A treasure of value 2 lies at the end of the corridor, while two treasures of value 1 serve as distractors next to the corridor entrance. The corridor is filled with puddles giving a penalty of −0.01. The agent always starts next to the entrance and the episode ends when the agent collects any of the treasures. The optimal policy consists of going always “right”.
The deep gridworld combines the challenges of both the deep sea and the taxi domains. Like the former, to discover the highest reward the agent needs to execute the action “right” multiple times in a row, receiving negative immediate rewards due to the puddles. However, doing a different action does not prevent reaching the end of the corridor, because the agent can still go back to the entrance and try again within the same episode. At the same time, the presence of the two distractor treasures results in local optima, as in the taxi driver domain.
Results. Because it is possible to exit the corridor without ending the episode, we set a longer horizon compared to the previous domains, i.e., 55 steps for the “short” scenario and 110 for the “long” one. Results shown in Figure 12 confirm previous results. The proposed exploration quickly discovers all states and learns to navigate through the corridor. By contrast, other algorithms, including bootstrapping, get stuck in local optima and learn to collect one of the two lesser treasures. Similarly to the taxi domain, with optimistic initialization, all algorithms—but random exploration (light green)—learn to navigate through the corridor. This evaluation shows that distractor rewards are challenging for existing algorithms but not for ours. Next, we increase the difficulty by adding more distractors and new special states.

4.1.4. Gridworlds

In these environments (Figure 13, Figure 14 and Figure 15), the agent navigates in a grid. Black cells cannot be visited, while any action in the prison has only an arbitrarily small probability of success. If it fails, the agent does not move. In practice, the prison almost completely prevents exploration. The reward is 0 everywhere except in treasure or penalty cells and is given for executing an action in the state, i.e., on transitions. Treasure cells (denoted by a money bag or a green number) give a bonus reward of different magnitude and end the episode. Penalty cells (denoted by a red number) give a penalty reward and end the episode. The agent also gets a small penalty of 0.01 at each step. The goal, thus, is to find the biggest treasure using as few steps as possible. The initial position is fixed such that it is far from the biggest treasure and close to the smallest one.
These domains are designed to highlight the difficulty of learning with many distractors (the smaller treasures) and, thus, of many local optima. The addition of the constant penalty at each step further discourages exploration and makes the distractors more appealing, since ending the episode will stop receiving penalties. Each domain has increasing difficulty and introduces additional challenges, as we explain below.
We start with the “toy” gridworld (Figure 13), a simple 5 × 5 grid with one reward of value 1. The focus of this domain is learning with sparse reward, without distractors or additional difficulties. The optimal policy takes only nine steps to reach the treasure, and the “short horizon” scenario ends after eleven steps.
The “prison” gridworld (Figure 14) increases the value of the furthest reward to 5, adds two distractors of value 1 and 2, and a prison cell. The first distractor is close to the initial position, while the second is close to the goal reward. The prison cell is close to the goal. The optimal policy navigates around the first distractor, then right below the prison, and finally up to the goal. The prison highlights the importance of seeking states with low long-term visitation count, rather than low immediate count. As discussed in Section 3.1, current count-based algorithms cannot deal with these kinds of states efficiently.
Finally, the “wall” gridworld (Figure 15) has a larger grid size ( 50 × 50 ), more rewards and penalties, and a wall separating the grid into two areas. The first area, where the agent starts, has six small rewards and penalties. The second area has two bigger treasures in the upper-right and bottom-right corners, of value 500 and 10,000, respectively. The optimal policy brings the agent beyond the wall and then to the bottom right corner, where the highest reward lies. The wall significantly increases the difficulty, due to the narrow passage that the agent needs to find in order to visit the second area of the grid. Learning is even harder when the horizon is short, as the agent cannot afford to lose time randomly looking for new states. To increase the chance of success of all algorithms, we set the “short horizon” to 330 steps, and 135 are needed to reach the reward of 10,000.
Results. The gridworlds confirm previous results. Without distractors (toy gridworld, Figure 16), bootstrapped algorithms (green and red) perform well, and so does using the auxiliary visitation bonus (pink). Neither, however, match ours (blue and orange, almost overlapping). Increasing the horizon helps the remainder algorithms, including random exploration (light green), except for approximate Thompson sampling (purple). In the next gridworld, however, the distractors and the prison cell substantially harm all algorithms except ours (Figure 17). Without optimistic initialization, in fact, existing algorithms cannot find the highest reward even with a long horizon, and all converge to local optima. This behavior was expected, given the study of Section 3.1. Finally, the wall gridworld results emphasize even more the superiority of our algorithms (Figure 18). With zero initialization, in fact, every other algorithm cannot go beyond the wall and find even the reward of 500. The results also stress that using the UCB reward visitation value (blue) over the count reward (orange) performs slightly best.
This evaluation strengthens the findings of previous experiments. First, it stresses how difficult it is for existing algorithms to learn with distracting rewards and a short horizon. Second, it shows that our proposed approach overcomes these challenges. Next, we investigate the algorithms sample efficiency.

4.2. Part 2: In-Depth Investigation

In this section, we present an in-depth analysis of issues associated with learning with sparse rewards, explaining in more detail why our approach outperformed existing algorithms in the previous evaluation. We start by reporting the visitation count at the end of the learning, showing that our algorithms explore the environment uniformly. We continue with a study on the impact of the visitation discount γ w , showing how it affects the performance of the W-function. Next, we present the empirical sample complexity analysis of all algorithms on the deep sea domain, showing that our approach scales gracefully with the number of states. Finally, we evaluate the algorithms in the infinite horizon setting and stochastic MDPs, evaluating the accuracy of the Q-function and the empirical sample complexity. Even in these scenarios, our approach outperforms existing algorithms.

4.2.1. Visitation Count at the End of the Learning

In Section 3.3 and Section 3.4, we discussed that the proposed behavior policies guarantee that an action is not executed twice before all other actions are executed once, under the uniform count assumption. We also acknowledged that this assumption cannot hold in practice because of state resets, or because the agent may need to revisit the same state to explore new ones. For example, in the deep sea, the agent needs to traverse diagonal states multiple times to visit every state. Nonetheless, in this section, we empirically show that our approach allows the agent to explore the environment as uniformly as possible.
Figure 19 reports the state visitation count at the end of learning in the “short-horizon zero-initialization” scenario for some domains. Only the proposed method (first column) uniformly explores all states. Recall, in fact, that episodes reset after some steps or when terminal states are reached. Thus, initial states (which are fixed) are naturally visited more often. The count in our method uniformly decreases proportionally to the distance from the initial states. This denotes that at each episode, starting from the same state, the agent followed different paths, exploring new regions of the environment uniformly.
Other algorithms suffer from distractors and local optima. In particular, UCB1 (third column) explores very slowly. The reason is that the agent myopically selects actions based on the immediate count, which makes exploration highly inefficient. By selecting the action with the lowest immediate count, the agent does not take into account where the action will lead it, i.e., if future states have already been visited or not. By contrast, our approach achieves deep exploration by taking into account the long-term visitation count of future states.
The performance of bootstrapping (second column) is peculiar. It immediately converges to the first reward found in the gridworlds, barely exploring afterward, but performs very well on the deep sea. This hints that bootstrapping may be sensitive to any kind of reward, including negative penalties. In the deep sea, in fact, diagonal states provide the only intermediate feedback besides the final reward/penalty. Coincidentally, traversing the diagonal also leads to the only positive reward. The agent may thus be guided by the reward received on the diagonal, even if it is a negative penalty. This behavior can be explained by recalling that Osband et al. [4] regularize TD learning with the 2 -distance from a prior Q-table. The agent may therefore be attracted to any state providing some kind of feedback to minimize the regularization.

4.2.2. Impact of Visitation Value Discount Factor γ w

Here, we investigate the effect of the visitation discount factor γ w on the “toy” gridworld (Figure 14). Figure 20 shows that the higher the discount, the better. As expected, with γ w = 0 the algorithms behave very similarly to UCB1. In this case, in fact, the proposed approach and UCB1 are equivalent. However, the learning curves are not exactly the same because the W-function is updated with TD learning at every step.
The better exploration of our approach is also confirmed by Figure 21, showing the count at the end of the learning. The higher γ w , the more uniform the exploration is because with a higher discount the agent will look further in the future for visitation rewards. With smaller γ w , instead, the agent rarely discovers states far from the initial position, as it does not look for future rewards. As a rule of thumb, it is appropriate to have γ w equal or larger than γ .

4.2.3. Empirical Sample Complexity on the Deep Sea

Here, we propose the same evaluation presented by Osband et al. [52] to investigate the empirical sample complexity of the proposed algorithms, and to assess how they scale to large problems. Figure 22 plots training steps to learn the optimal policy as a function of the environment size N. Only our approach (blue and orange) scales gracefully to large problem sizes. Results suggest an empirical scaling of O ( N 2.5 ) for the visitation-value-based with count reward, and even smaller for UCB reward. Bootstrapping attains an empirical complexity of O ( N 3 ) , confirming the findings of Osband et al. [4]. However, in many cases ( N = 23 , 27 , 33 , 37 , 43 , 47 , 51 , 55 ) there was one seed for which bootstrapping either did not learn within the step limit (green dashed line, empty dots) due to premature convergence, or learned after substantially more steps than the average (red dots, large error bar). Missing algorithms (random, ϵ -greedy, UCB1, approximate Thompson sampling) performed extremely poorly, often not learning within 500,000 steps even for small N, and are thus not reported.

4.2.4. Infinite Horizon Stochastic Chainworld

This evaluation investigates how the algorithms perform in a stochastic MDP with infinite horizon. In particular, we are interested in (1) the mean squared value error (MSVE) at each timestep, (2) the sample complexity when varying the number of states, and (3) how behavior policies explore and if they converge to the greedy policy.
Evaluation Criteria. The MSVE is computed between the true value function of the optimal policy V * ( s ) and the learned value function of the current policy V π t ( s ) , i.e.,
MSVE = 1 N i = 1 N V * ( s i ) V π t ( s ) 2 .
The value function is computed according to the Bellman expectation equation in matrix form, i.e., V π = ( I γ P π ) 1 R π , where I is the identity matrix, and P π and R π are the transition and reward functions induced by the policy, respectively, [56]. This error indicates how much the learned greedy policy π t deviates from the optimal policy.
Similarly, the sample complexity is defined as the number of timesteps t such that the non-stationary policy π t at time t is not ε -optimal for current state s t , i.e., V * ( s t ) V π t ( s t ) > ε [17,30]. For the behavior of exploration policies and their empirical convergence, we show how the visitation count changes over time.
MDP Characteristics. The MDP, shown in Figure 23, is a chainworld with s = 1 , , N states, three actions, and stochastic transition defined as follows. The first action moves the agent forward with probability p, backward otherwise. The second action moves the agent backward with probability p, forward otherwise. The third action keeps the agent in the current state with probability p, and randomly moves it backward or forward otherwise. The initial state is the leftmost state, and no state is terminal. This MDP is ergodic, i.e., all states are transient, positive recurrent, and aperiodic for any deterministic policy. In our experiments, we set p = 0.99 . The reward is 10 8 for doing “stay” in the initial state s 1 , 1 for doing “stay” in the last state s N , and 0 everywhere else.
Unlike previous MDPs, this has no terminal state, the horizon is infinite, and the state is never reset. Instead, to simulate infinite horizon, the agent explores the environment for one episode of 200,000 steps. To comply with the classic sample complexity definition [17,30], we use classic Q-learning without replay memory. We do not compare against bootstrapping algorithms, because they are designed to select a different behavior Q-function at every episode, and this setup has only one episode. Approximate Thompson sampling by D’Eramo et al. [46], instead, selects the Q-function at every step, and it is included in the comparison (without memory as well). All algorithms use zero Q-function initialization. Due to the stochasticity of the MDP, we increased the number of random seeds to 50.
Results.Figure 24 shows how the algorithms explore over time. Only ours (Figure 24a,b) and UCB1 (Figure 24c) explore uniformly, but UCB1 finds the last state (with the reward) later. The auxiliary rewards (Figure 24d) and approximate Thompson sampling (Figure 24e) also perform poorly, since the exploration is not uniform and the reward is found only late. ϵ -greedy exploration (Figure 24f), instead, is soon stuck in the local optimum represented by the small reward in the initial state.
The visitation count also shows that all exploration policies but ours act greedily once the reward is found. For example, UCB1 finds the reward around step 1000, and after that its visitation count increases only in the last state (and nearby states, because of the stochastic transition). This suggests that when these algorithms find the reward contained in the last state they effectively stop exploring other states. By contrast, our algorithms explore the environment for longer, yielding a more uniform visitation count. This allows the agent to learn the true Q-function for all states, achieving lower sample complexity and MSVE as shown in Figure 25 and Figure 26.
Figure 25 show the sample complexity on varying the chain size over three values of ε . In all three cases, our algorithms scale gracefully with the number of states, suggesting a complexity of O ( N 2.5 ) and strengthening the findings of Section 4.2.3. By contrast, other algorithms performed poorly. Their sample complexity has a large confidence interval, and as ε decreases the complexity increases to the point that they rarely learn an ε -optimal policy.
These results are confirmed by Figure 26 (left plots), where only our algorithms attain low sample complexity even for ε = 0 . Our algorithms are also the only ones to learn an almost perfect Q-function, with an MSVE close to zero (right plots). As anticipated, these results are explained by the uniform visitation count in Figure 24. By not converging to the greedy policy too quickly after discovering the reward, the agent keeps visiting old states and propagating to them the information about the reward, thus learning the true optimal Q-function for all states.

4.2.5. Stochastic Gridworlds

As we discussed in Section 3.4, the visitation rewards for training the W-functions penalize terminal state-action pairs. One may think that this would lead to poor exploration in stochastic environments, where the same state-action pair can lead to both terminal and non-terminal states. Here, we evaluate all algorithms again on some of the previous environments but this time with stochastic transitions and show that our W-functions still solve the MDPs and outperform existing algorithms. Each transition ( s , a ) has probability p of succeeding, 1 p / 2 of not moving the agent, and 1 p / 2 of moving the agent to a random adjacent state. In our experiments, we set p = 0.9 . Due to the stochasticity of transitions, we increased the number of random seeds from 20 to 50.
Figure 27 shows the number of states discovered and the expected discounted return computed in closed form as V π ( s 0 ) , where V π is defined according to the Bellman expectation equation V π = ( I γ P π ) 1 R π . Similarly to the deterministic setting (Figure 12, Figure 16 and Figure 17, leftmost plots), our method (blue and orange) outperform all baselines, being the only solving all MPDs within the steps limit (because of the stochasticity we decreased the learning rate from 0.5 to 0.1, and increased the learning steps). With the UCB reward (blue), our approach always learns the optimal policy. With the count reward (orange), our approach almost always learns it, and converges to a distractor only two times (out of 50) in the deep gridworld, and once in the “prison” gridworld. However, they both quickly discovered all states, as shown by top plots. The better performance of the UCB-based reward is due to the optimistic initialization of the W-function. As seen in Section 4.2.4, Figure 24a, this version of the algorithm explores the environment for longer and converges to the greedy policy later. Therefore, the agent will visit the high-reward state more often, and it is less likely to learn suboptimal policies.
Bootstrapping with a prior (green) is not affected by the stochasticity, and it performs as in the deterministic setting (it solves the “toy” gridworld but not the other two). Vanilla bootstrapping (red), instead, is heavily affected by the stochasticity, and its performance is substantially worse compared to the deterministic setting (it cannot learn even the “toy” gridworld within steps limit). UCB1 (brown) and bonus-based exploration (pink), instead, benefit from the stochasticity. In the deterministic setting they could not solve these MDPs, while here they solve the two gridworlds (even though much later than ours). This improvement is explained by the larger number of visited states (top plots), which denotes an overall better exploration. This is not surprising if we consider that a stochastic transition function naturally helps exploration.

5. Conclusions and Future Work

Effective exploration with sparse rewards is an important challenge in RL, especially when sparsity is combined with the presence of “distractors”, i.e., rewards that create suboptimal modes of the objective function. Classic algorithms relying on dithering exploration typically perform poorly, often converging to poor local optima or not learning at all. Methods based on immediate counts have strong guarantees but are empirically not sample efficient, while we showed that methods based on intrinsic auxiliary rewards require hand-tuning and are prone to suboptimal behavior. In this paper, we presented a novel approach that (1) plans exploration actions far into the future by using a long-term visitation count, and (2) decouples exploration and exploitation by learning a separate function assessing the exploration value of the actions. Contrary to existing methods that use models of reward and dynamics, our approach is off-policy and model-free. Empirical results showed that the proposed approach outperforms existing methods in environments with sparse and distracting rewards, and suggested that our approach scales gracefully with the size of the environment.
The proposed approach opens several avenues of research. First, in this work, we focused on empirical results. In the future, we will investigate the theoretical properties of the proposed approach in more detail. Second, in this work, we considered model-free RL. In the future, we will extend the proposed approach and combine it with model-based RL. Third, the experimental evaluation focused on identifying the challenges of learning with sparse and distracting rewards. In the future, we will consider more diverse tasks with continuous states and actions, and extend the proposed exploration to actor-critic methods.

Author Contributions

Conceptualization: S.P., D.T. and J.P. (Joni Pajarinen); methodology, S.P., D.T. and J.P. (Joni Pajarinen); software, S.P., D.T. and C.D.; formal analysis, S.P., D.T. and J.P. (Joni Pajarinen); investigation, S.P., D.T. and J.P. (Joni Pajarinen); resources, S.P., D.T. and C.D.; writing—original draft preparation, S.P., D.T., M.H. and J.P. (Joni Pajarinen); writing—review and editing, S.P., D.T., J.P. (Joni Pajarinen) and J.P. (Jan Peters); supervision, J.P. (Jan Peters). All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Source code at https://github.com/sparisi/visit-value-explore accessed on 20 January 2022.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Experiment Details

Below, we present the pseudocode and the hyperparameters of the algorithms used in Section 4. In all of them, we kept two separate Q-tables for the behavior policy β ( a | s ) and the target (greedy) policy π ( a | s ) . The former can be either initialized to zero or optimistically, while the latter always to zero. With the optimistic initialization, all entries of Q ( s , a ) are set to r max / ( 1 γ ) . The reason why we do not initialize Q π ( s , a ) optimistically is that if the agent does not visit all the state-action pairs, and thus never updated the corresponding Q-table entries, it would still have an optimistic belief over some states. The performance of the target policy can therefore be poor until all state-action pairs are visited. For example, in our experiments in the “prison” gridworld, the ϵ -greedy policy was finding some low-value treasures, but during the evaluation the greedy policy was not trying to collect them because it still had an optimistic belief over unvisited empty states.
In most of the pseudocode, we explicitly distinguish between Q β ( s , a ) and Q π ( s , a ) . If simply Q ( s , a ) appears, it means that both Q-tables are updated using the same equation.
In all algorithms, we evaluate π ( a | s ) , i.e., the greedy policy over Q π ( s , a ) , every 50 training steps. Each evaluation episode has a horizon 10% longer than the training one. The learning rate is η = 0.5 and the discount factor γ = 0.99 . For MDPs with stochastic transition function (Section 4.2.4 and Section 4.2.5) we used η = 0.1 . The ϵ -greedy initially has ϵ 0 = 1 , then it decays at step according to ϵ i + 1 = ξ ϵ i , where ξ is chosen such that ϵ e n d = 0.1 when the learning is over. Finally, we break ties with random selection, i.e., if two or more actions have the same max Q-value the winner is chosen randomly.

Appendix A.1. The Generic Q-Learning Scheme

In all experiments, we used Q-learning with infinite replay memory. Classic Q-learning by Watkins and Dayan [22] updates the Q-tables using only the current transition. The so-called experience replay, instead, keeps past transitions and uses them multiple times for successive updates. This procedure became popular with deep Q-learning [1] and brought substantial improvement to many RL algorithms. In our experiments, we followed the setup proposed by Osband et al. [4] and used an infinite memory, i.e., we stored all transitions, since it allowed much faster learning than classic Q-learning. This can also be seen as Dyna-Q [60] without random memory sampling. In our experiments, the size of the domains does not pose any storage issue. When storage size is an issue, it is possible to fix the memory size and sample random mini-batches as in DQN [1]. More details in Appendix A.6. Algorithm A1 describes the generic scheme of Q-learning with infinite replay memory. TD learning is used on both Q π ( s , a ) and Q β ( s , a ) , with the only difference being the initialization ( Q π ( s , a ) is always initialized to zero).
Depending on β ( a | s ) , we have different exploration strategies, i.e.,
β ( a t | s t ) unif { A } , random
β ( a t | s t ) = arg max a Q β ( s t , a ) with probability 1 ϵ , unif { A } with probability ϵ , ϵ - greedy
β ( a t | s t ) = arg max a Q β ( s t , a ) + κ 2 log a j n ( s t , a j ) n ( s t , a ) . UCB 1
Numerical stability and worst-case scenario. For UCB1 [28], a visitation count n ( s , a ) is increased after every transition (see Algorithm A2, line 7). In classic bandit problems, UCB1 is initialized by executing all actions once, i.e., with n ( s , a ) = 1 . In MDPs we cannot do that, i.e., we cannot arbitrarily set the agent in any state and execute all actions, thus n ( s , a ) = 0 . Following the W-function bound in Section 3.3, we always add +1 inside the logarithm, and bound the square root to ( Q max Q min ) / κ + 2 log | A | when n ( s , a ) = 0 . This correspond to the case where all actions but a ¯ has been executed once, and enforces the policy to choose a ¯ . In our experiments, we set Q max = r max / ( 1 γ ) and Q min = 0 .
Algorithm A1: Tabular Q-Learning with Replay Memory
Algorithms 15 00081 i001

Appendix A.2. Q-Learning with Augmented Reward

This version follows the same scheme of Algorithm A1 with ϵ -greedy exploration. The only difference is that the reward used to train the behavior policy is augmented with the exploration bonus proposed by Strehl and Littman [30]. In our experiments, α = 0.1 , as used by Strehl and Littman [30] and Bellemare et al. [15].
Algorithm A2: Tabular Q-Learning with Replay Memory and Augmented Reward
Algorithms 15 00081 i002

Appendix A.3. Q-Learning with Visitation Value

Algorithms A3 and A4 describe the novel method proposed in this paper. In our experiments κ = r max / ( 1 γ ) and γ w = 0.99 . For the chainworld in Section 4.2.4 we set γ w = 0.999 . With infinite horizon, in fact, this yielded better results by allowing the agent to explore for longer. For the stochastic gridworld in Section 4.2.5 we set γ w = 0.9 . As discussed in the results, in fact, a stochastic transition function naturally improves exploration, thus a smaller visitation discount was sufficient and yielded better results.
Algorithm A3: Tabular Q-Learning with Replay Memory and Visit. Value (UCB)
Algorithms 15 00081 i003
Algorithm A4: Tabular Q-Learning with Replay Memory and Visit. Value (Count)
Algorithms 15 00081 i004
In Algorithm A3, W UCB β ( s , a ) is initialized as Equation (24). In Algorithm A4, line 6, we bound the square root to Equation (27) when n ( s , a ) = 0 . In our experiments, with Q max = r max / ( 1 γ ) and Q min = 0 .

Appendix A.4. Bootstrapped Q-Learning

Bootstrap algorithms have an ensemble of Q-tables { Q b ( s , a ) } b = 1 B . The behavior policy is greedy over one Q-table, randomly chosen from the ensemble either at the beginning of the episode or at every step.
The behavior Q-tables are initialized randomly. After initializing each Q b ( s , a ) either optimistically or to zero, we add random noise drawn from a Gaussian N ( 0 , 1 ) . Then, each one is trained on a random mini-batch from the replay memory. In our experiments, we used batches of size 1024 and an ensemble of B = 10 behavior Q-tables. The target Q-table is still updated as in previous algorithms, i.e., using the full memory.
The above pseudocode defines the generic bootstrap approach proposed by Osband et al. [9]. In Section 4, we also compared to two slightly different versions. The first uses approximate Thompson sampling and randomly selects the behavior Q-table at every step instead of at the beginning of the episode [46]. The second keeps the sampling at the beginning of the episode, but further regularizes the TD error [4,52]. The regularization is the squared 2 -norm of the distance of the Q-tables from “prior Q-tables” Q p ( s , a ) , resulting in the following regularized TD update
δ r e g b ( s , a , s ) = δ b ( s , a , s ) + ν Q p ( s , a ) Q b ( s , a ) ,
wehere ν is the regularization coefficient which we set to ν = 0.1 (in the original paper ν = 1 , but dividing it by ten worked best in our experiments).
In theory, Q p ( s , a ) should be drawn from a distribution. In practice, the same authors fix it at the beginning of the learning, and for the deep sea domain they set it to zero. This configuration also worked best in our experiments.

Appendix A.5. Horizons

All environments in Section 4 are infinite horizon MDPs, except for the deep sea which has a finite horizon equal to its depth. However, for a practical reason, we end the episode after H steps and reset the agent to the initial state. Table A1 summarizes the training horizon H for each environment, as well as the steps needed for the optimal policy to find the highest reward. Furthermore, recall that an episode can end prematurely if a terminal state (e.g., a reward state) is reached.
Finally, notice that the agent receives the reward on state-action transition, i.e., it has to execute an action in the state with a treasure to receive the reward. This is why, for instance, the agent needs 9 steps instead of 8 to be rewarded in the 5 × 5 gridworlds.
Table A1. Time horizons H for the tabular MDPs presented in Section 4.
Table A1. Time horizons H for the tabular MDPs presented in Section 4.
Deep SeaTaxiDeep Grid.Grid. (Toy)Grid. (Prison)Grid. (Wall)
OptimalDepth291199135
Short H.Depth33551111330
Long H.Depth661102222660
Stochastic--551525-

Appendix A.6. Infinite Memory vs. No Memory

Except for the chainworld in Section 4.2.4, the evaluation in Section 4 was conducted with Q-learning with infinite replay memory, following the same setup of Osband et al. [4] as described in Section A.1. Here, we present additional results without replay memory, i.e., using classic Q-learning by Watkins and Dayan [22] which updates the Q-tables using only the current transition. We show that the proposed method benefits the most from the replay memory, but can learn even without it while other algorithms cannot. We compare ϵ -greedy exploration, our proposed visitation-value-based exploration, and bootstrapped exploration. Nevertheless, the latter needs a replay memory to randomize the Q-tables update, as each Q-table is updated using different random samples. Thus, for bootstrapped Q-learning, we compare the use of finite memory and small batches to infinite memory and large batches. Algorithm A5 describes the generic scheme of Q-learning without replay memory. The only difference from Algorithm A1 is the absence of the loop over the memory. The same modification is done to Algorithms A4 and A6 to derive the version of our algorithms without infinite memory. Bootstrapped Q-learning is the same as Algorithm A3 but with finite memory. The finite memory keeps at most 20 H steps, where H is the episode horizon, and uses mini-batches of 32 samples instead of 1024.
Algorithm A5: Classic Tabular Q-Learning
Algorithms 15 00081 i005
Algorithm A6: Tabular Bootstrapped Q-Learning with Replay Memory
Algorithms 15 00081 i006
Results.Figure A1 shows the results on two gridworlds. The “toy” one has a single reward, while the “prison” one has also distractors. We refer to Section 4.1 for their full description. In the first domain, all algorithms perform well except for ϵ -greedy exploration (green), which missed the reward in some trials. The use of infinite memory helped all algorithms, allowing bootstrap and visitation-value-based to learn substantially faster. For example, on average, without replay memory our algorithms (blue and orange) converged in 1000 steps, while with replay memory they took only 250 steps, i.e., × 4 times faster. Bootstrapped Q-learning (red), performance does not change substantially with finite and infinite memory. However, in the second domain only visitation-value-based exploration always learns, regardless of the memory. As discussed in Section 4.1, bootstrap performs poorly due to distractor rewards, and the infinite memory does not help it. For visitation-value-based exploration, the speed-up gained from the infinite memory is even larger than before, due to the higher complexity of the domain. In this case in fact, with infinite memory, it learns × 4 and × 8 times faster than without memory (blue and orange, respectively). This evaluation shows that the proposed method benefits the most from the replay memory, but can learn even without it while other algorithms cannot.
Figure A1. Comparison of the proposed exploration against ϵ -greedy and bootstrapped exploration with and without replay memory. Each line denotes the average over 20 seeds. Only exploration based on the visitation value always learns in both domains, i.e., both with and without local optima (distractor rewards), regardless of the memory.
Figure A1. Comparison of the proposed exploration against ϵ -greedy and bootstrapped exploration with and without replay memory. Each line denotes the average over 20 seeds. Only exploration based on the visitation value always learns in both domains, i.e., both with and without local optima (distractor rewards), regardless of the memory.
Algorithms 15 00081 g0a1

Appendix A.7. Recap Tables

Here, we report tables summarizing the results of Section 4.1 in terms of discovery (percentage of states discovered during learning) and success (number of times the algorithm learned the optimal policy within steps limit). These are extended versions of Table 1, which reported results only for the “zero initialization short horizon” scenario.
Table A2. Deep sea results recap.
Table A2. Deep sea results recap.
AlgorithmDiscovery (%)Success (%)
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 99.90 ± 0.01 100 ± 0
Bootstr. (Osband 2016a) 99.77 ± 0.05 100 ± 0
Zero Init.Bootstr. (D’Eramo 2019) 63.25 ± 3.31 0 ± 0
UCB1 (Auer 2002) 55.72 ± 0.34 0 ± 0
Expl. Bonus (Strehl 2008) 85.65 ± 1.0 0 ± 0
ϵ -greedy 57.74 ± 1.11 0 ± 0
Random 58.59 ± 1.35 0 ± 0
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 100 ± 0 100 ± 0
Bootstr. (Osband 2016a) 100 ± 0 100 ± 0
Opt. Init.Bootstr. (D’Eramo 2019) 100 ± 0 100 ± 0
UCB1 (Auer 2002) 76.85 ± 0.3 0 ± 0
Expl. Bonus (Strehl 2008) 98.79 ± 0.21 0 ± 0
ϵ -greedy 98.79 ± 0.20 0 ± 0
Random 58.59 ± 1.35 0 ± 0
Table A3. Taxi results recap.
Table A3. Taxi results recap.
AlgorithmDiscovery (%)Success (%)
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 69.60 ± 2.96 13.07 ± 2.96
Bootstr. (Osband 2016a) 52.44 ± 7.55 18.77 ± 9.24
Zero Init. & Short Hor.Bootstr. (D’Eramo 2019) 22.44 ± 1.81 1.9 ± 1.47
UCB1 (Auer 2002) 31.17 ± 0.70 1.53 ± 1.37
Expl. Bonus (Strehl 2008) 74.62 ± 2.24 17.6 ± 2.56
ϵ -greedy 29.64 ± 0.98 1.92 ± 1.49
Random 29.56 ± 0.98 1.92 ± 1.49
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 100 ± 0 100 ± 0
Bootstr. (Osband 2016a) 100 ± 0 100 ± 0
Opt. Init. & Short Hor.Bootstr. (D’Eramo 2019) 100 ± 0 100 ± 0
UCB1 (Auer 2002) 99.94 ± 0.08 100 ± 0
Expl. Bonus (Strehl 2008) 100 ± 0 100 ± 0
ϵ -greedy 100 ± 0 100 ± 0
Random 29.56 ± 0.98 1.92 ± 1.49
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 76.38 ± 2.31 15.69 ± 2.93
Bootstr. (Osband 2016a) 53.80 ± 6.63 29.85 ± 12.72
Zero Init. & Long Hor.Bootstr. (D’Eramo 2019) 36.34 ± 3.16 6.54 ± 2.62
UCB1 (Auer 2002) 49.66 ± 2.45 8.35 ± 1.32
Expl. Bonus (Strehl 2008) 92.83 ± 2.26 76.18 ± 16.09
ϵ -greedy 42.26 ± 2.52 7.66 ± 0.03
Random 55.89 ± 0.98 2.06 ± 1.49
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 100 ± 0 100 ± 0
Bootstr. (Osband 2016a) 100 ± 0 100 ± 0
Opt. Init. & Long Hor.Bootstr. (D’Eramo 2019) 95.45 ± 0 100 ± 0
UCB1 (Auer 2002) 98.74 ± 0.09 100 ± 0
Expl. Bonus (Strehl 2008) 100 ± 0 100 ± 0
ϵ -greedy 100 ± 0 100 ± 0
Random 55.89 ± 0.98 2.06 ± 1.49
Table A4. Deep gridworld results recap.
Table A4. Deep gridworld results recap.
AlgorithmDiscovery (%)Success (%)
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 75 ± 9.39 59.03 ± 4.2
Bootstr. (Osband 2016a) 60.82 ± 11.78 63.35 ± 6.89
Zero Init. & Short Hor.Bootstr. (D’Eramo 2019) 63.73 ± 10.35 56.85 ± 0.06
UCB1 (Auer 2002) 92.18 ± 0.36 56.88 ± 0
Expl. Bonus (Strehl 2008) 95.45 ± 1.57 69.81 ± 8.84
ϵ -greedy 74.36 ± 4.42 56.88 ± 0
Random 92.45 ± 0.64 56.88 ± 0
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 100 ± 0 100 ± 0
Bootstr. (Osband 2016a) 100 ± 0 100 ± 0
Opt. Init. & Short Hor.Bootstr. (D’Eramo 2019) 100 ± 0 100 ± 0
UCB1 (Auer 2002) 100 ± 0 100 ± 0
Expl. Bonus (Strehl 2008) 100 ± 0 100 ± 0
ϵ -greedy 100 ± 0 100 ± 0
Random 92.45 ± 0.64 56.88 ± 0
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 78.18 ± 8.97 56.88 ± 0
Bootstr. (Osband 2016a) 77.64 ± 10.02 69.81 ± 8.84
Zero Init. & Long Hor.Bootstr. (D’Eramo 2019) 71.64 ± 10.35 56.85 ± 0.06
UCB1 (Auer 2002) 95.54 ± 0.46 56.4 ± 3.12
Expl. Bonus (Strehl 2008) 95 ± 1.66 65.5 ± 7.71
ϵ -greedy 76.45 ± 4.25 56.88 ± 0
Random 92.09 ± 0.69 56.88 ± 0
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 100 ± 0 100 ± 0
Bootstr. (Osband 2016a) 100 ± 0 100 ± 0
Opt. Init. & Long Hor.Bootstr. (D’Eramo 2019) 100 ± 0 100 ± 0
UCB1 (Auer 2002) 100 ± 0 100 ± 0
Expl. Bonus (Strehl 2008) 100 ± 0 100 ± 0
ϵ -greedy 100 ± 0 100 ± 0
Random 92.09 ± 0.69 56.88 ± 0
Table A5. Gridworld (toy) results recap.
Table A5. Gridworld (toy) results recap.
AlgorithmDiscovery (%)Success (%)
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 99.8 ± 0.39 100 ± 0
Bootstr. (Osband 2016a) 99.2 ± 0.72 100 ± 0
Zero Init. & Short Hor.Bootstr. (D’Eramo 2019) 84.8 ± 4.33 40 ± 21.91
UCB1 (Auer 2002) 97.4 ± 0.99 50 ± 22.49
Expl. Bonus (Strehl 2008) 99.8 ± 0.39 100 ± 0
ϵ -greedy 97.4 ± 1.17 45 ± 22.25
Random 97.2 ± 1.15 45 ± 22.25
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 100 ± 0 100 ± 0
Bootstr. (Osband 2016a) 100 ± 0 100 ± 0
Opt. Init. & Short Hor.Bootstr. (D’Eramo 2019) 100 ± 0 100 ± 0
UCB1 (Auer 2002) 100 ± 0 100 ± 0
Expl. Bonus (Strehl 2008) 100 ± 0 100 ± 0
ϵ -greedy 100 ± 0 100 ± 0
Random 97.2 ± 1.15 45 ± 22.25
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 100 ± 0 100 ± 0
Bootstr. (Osband 2016a) 99.8 ± 0.39 100 ± 0
Zero Init. & Long Hor.Bootstr. (D’Eramo 2019) 94.6 ± 1.9 70 ± 20.49
UCB1 (Auer 2002) 100 ± 0 98.5 ± 0.22
Expl. Bonus (Strehl 2008) 100 ± 0 100 ± 0
ϵ -greedy 100 ± 0 100 ± 0
Random 100 ± 0 100 ± 0
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 100 ± 0 100 ± 0
Bootstr. (Osband 2016a) 100 ± 0 100 ± 0
Opt. Init. & Long Hor.Bootstr. (D’Eramo 2019) 100 ± 0 100 ± 0
UCB1 (Auer 2002) 100 ± 0 100 ± 0
Expl. Bonus (Strehl 2008) 100 ± 0 100 ± 0
ϵ -greedy 100 ± 0 100 ± 0
Random 100 ± 0 100 ± 0
Table A6. Gridworld (prison) results recap.
Table A6. Gridworld (prison) results recap.
AlgorithmDiscovery (%)Success (%)
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 72.39 ± 6.9 44.44 ± 14.69
Bootstr. (Osband 2016a) 64.35 ± 11.1 33.03 ± 8.54
Zero Init. & Short Hor.Bootstr. (D’Eramo 2019) 54.78 ± 6.69 25.61 ± 8.39
UCB1 (Auer 2002) 80.78 ± 2.56 32.31 ± 4.11
Expl. Bonus (Strehl 2008) 85.87 ± 3 46.96 ± 12.35
ϵ -greedy 58.38 ± 5.27 22.84 ± 2.48
Random 76.96 ± 3.08 35.36 ± 10.37
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 100 ± 0 100 ± 0
Bootstr. (Osband 2016a) 100 ± 0 100 ± 0
Opt. Init. & Short Hor.Bootstr. (D’Eramo 2019) 100 ± 0 100 ± 0
UCB1 (Auer 2002) 100 ± 0 100 ± 0
Expl. Bonus (Strehl 2008) 100 ± 0 100 ± 0
ϵ -greedy 100 ± 0 100 ± 0
Random 76.96 ± 3.08 35.36 ± 10.37
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 77.83 ± 6.93 56.04 ± 14.79
Bootstr. (Osband 2016a) 67.61 ± 9.75 44.12 ± 13.02
Zero Init. & Long Hor.Bootstr. (D’Eramo 2019) 59.57 ± 9.92 29.31 ± 4.11
UCB1 (Auer 2002) 90.17 ± 2.55 51.55 ± 10.67
Expl. Bonus (Strehl 2008) 88.91 ± 3.1 54.48 ± 11.63
ϵ -greedy 67.39 ± 5.16 24.69 ± 3.31
Random 86.09 ± 3.77 49.73 ± 11.5
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 100 ± 0 100 ± 0
Bootstr. (Osband 2016a) 100 ± 0 100 ± 0
Opt. Init. & Long Hor.Bootstr. (D’Eramo 2019) 100 ± 0 100 ± 0
UCB1 (Auer 2002) 100 ± 0 100 ± 0
Expl. Bonus (Strehl 2008) 100 ± 0 100 ± 0
ϵ -greedy 100 ± 0 100 ± 0
Random 86.09 ± 3.77 49.73 ± 11.5
Table A7. Gridworld (wall) results recap.
Table A7. Gridworld (wall) results recap.
AlgorithmDiscovery (%)Success (%)
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 9.44 ± 3.14 0.14 ± 0.06
Bootstr. (Osband 2016a) 6.35 ± 2.74 0.23 ± 0.08
Zero Init. & Short Hor.Bootstr. (D’Eramo 2019) 12.55 ± 4.45 0.18 ± 0.08
UCB1 (Auer 2002) 24.7 ± 4.45 0.46 ± 0.03
Expl. Bonus (Strehl 2008) 45.16 ± 2.29 0.52 ± 0.04
ϵ -greedy 38.15 ± 1.95 0.39 ± 0.08
Random 65.28 ± 0.45 0.59 ± 0
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 100 ± 0 100 ± 0
Bootstr. (Osband 2016a) 100 ± 0 100 ± 0
Opt. Init. & Short Hor.Bootstr. (D’Eramo 2019) 100 ± 0 100 ± 0
UCB1 (Auer 2002) 100 ± 0 100 ± 0
Expl. Bonus (Strehl 2008) 100 ± 0 100 ± 0
ϵ -greedy 100 ± 0 100 ± 0
Random 65.28 ± 0.45 0.59 ± 0
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 13.05 ± 3.83 0.2 ± 0.08
Bootstr. (Osband 2016a) 8.69 ± 3.53 0.2 ± 0.08
Zero Init. & Long Hor.Bootstr. (D’Eramo 2019) 11.64 ± 5.24 0.23 ± 0.08
UCB1 (Auer 2002) 21.44 ± 3.64 0.44 ± 0.03
Expl. Bonus (Strehl 2008) 47.75 ± 1.68 0.55 ± 0.03
ϵ -greedy 42.53 ± 1.89 0.54 ± 0.03
Random 70.51 ± 0.75 0.59 ± 0
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 100 ± 0 100 ± 0
Bootstr. (Osband 2016a) 100 ± 0 100 ± 0
Opt. Init. & Long Hor.Bootstr. (D’Eramo 2019) 100 ± 0 100 ± 0
UCB1 (Auer 2002) 100 ± 0 100 ± 0
Expl. Bonus (Strehl 2008) 100 ± 0 100 ± 0
ϵ -greedy 100 ± 0 100 ± 0
Random 70.51 ± 0.75 0.59 ± 0

References

  1. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Graves, A.; Antonoglou, I.; Wierstra, D.; Riedmiller, M. Playing Atari With Deep Reinforcement Learning. In Proceedings of the NIPS Workshop on Deep Learning, Lake Tahoe, CA, USA, 5 December 2013. [Google Scholar]
  2. Schulman, J.; Wolski, F.; Dhariwal, P.; Radford, A.; Klimov, O. Proximal policy optimization algorithms. arXiv 2017, arXiv:1707.06347. [Google Scholar]
  3. Silver, D.; Hubert, T.; Schrittwieser, J.; Antonoglou, I.; Lai, M.; Guez, A.; Lanctot, M.; Sifre, L.; Kumaran, D.; Graepel, T.; et al. Mastering chess and shogi by self-play with a general reinforcement learning algorithm. arXiv 2017, arXiv:1712.01815. [Google Scholar]
  4. Osband, I.; Roy, B.V.; Russo, D.J.; Wen, Z. Deep Exploration via Randomized Value Functions. J. Mach. Learn. Res. 2019, 20, 1–62. [Google Scholar]
  5. Lillicrap, T.P.; Hunt, J.J.; Pritzel, A.; Heess, N.; Erez, T.; Tassa, Y.; Silver, D.; Wierstra, D. Continuous control with deep reinforcement learning. In Proceedings of the International Conference on Learning Representations (ICLR), San Juan, Puerto Rico, 2 May 2016. [Google Scholar]
  6. Kakade, S. On the Sample Complexity of Reinforcement Learning. Ph.D. Thesis, University College London, London, UK, 2003. [Google Scholar]
  7. Szepesvari, C. Algorithms for Reinforcement Learning; Morgan & Claypool Publishers: Williston, VT, USA, 2010; Volume 4. [Google Scholar]
  8. Osband, I.; Van Roy, B.; Wen, Z. Generalization and Exploration via Randomized Value Functions. In Proceedings of the International Conference on Machine Learning (ICML), New York, NY, USA, 19 June 2016. [Google Scholar]
  9. Osband, I.; Blundell, C.; Pritzel, A.; Van Roy, B. Deep exploration via bootstrapped DQN. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Barcelona, Spain, 5 December 2016. [Google Scholar]
  10. Kearns, M.; Singh, S. Near-optimal reinforcement learning in polynomial time. Mach. Learn. 2002, 49, 209–232. [Google Scholar] [CrossRef] [Green Version]
  11. Brafman, R.I.; Tennenholtz, M. R-MAX—A general polynomial time algorithm for near-optimal reinforcement learning. J. Mach. Learn. Res. 2002, 3, 213–231. [Google Scholar]
  12. Jaksch, T.; Ortner, R.; Auer, P. Near-optimal regret bounds for reinforcement learning. J. Mach. Learn. Res. 2010, 11, 1563–1600. [Google Scholar]
  13. Hester, T.; Stone, P. TEXPLORE: Real-time sample-efficient reinforcement learning for robots. Mach. Learn. 2013, 90, 385–429. [Google Scholar] [CrossRef] [Green Version]
  14. Strehl, A.L.; Li, L.; Wiewiora, E.; Langford, J.; Littman, M.L. PAC model-free reinforcement learning. In Proceedings of the International Conference on Machine Learning (ICML), Pittsburgh, PA, USA, 25 June 2006; pp. 881–888. [Google Scholar]
  15. Bellemare, M.G.; Srinivasan, S.; Ostrovski, G.; Schaul, T.; Saxton, D.; Munos, R. Unifying count-based exploration and intrinsic motivation. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Barcelona, Spain, 5 December 2016; pp. 1471–1479. [Google Scholar]
  16. Jin, C.; Allen-Zhu, Z.; Bubeck, S.; Jordan, M.I. Is Q-learning provably efficient? In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 3 December 2018; pp. 4863–4873. [Google Scholar]
  17. Dong, K.; Wang, Y.; Chen, X.; Wang, L. Q-learning with UCB Exploration is Sample Efficient for Infinite-Horizon MDP. In Proceedings of the International Conference on Learning Representation (ICLR), Virtual, 26 April 2020. [Google Scholar]
  18. Kolter, J.Z.; Ng, A.Y. Near-Bayesian exploration in polynomial time. In Proceedings of the International Conference on Machine Learning (ICML), Montreal, QC, Canada, 24 June 2009; pp. 513–520. [Google Scholar]
  19. Houthooft, R.; Chen, X.; Duan, Y.; Schulman, J.; De Turck, F.; Abbeel, P. VIME: Variational information maximizing exploration. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Barcelona, Spain, 5 December 2016; pp. 1109–1117. [Google Scholar]
  20. Pathak, D.; Agrawal, P.; Efros, A.A.; Darrell, T. Curiosity-driven Exploration by Self-supervised Prediction. In Proceedings of the International Conference on Machine Learning (ICML), Sydney, NSW, Australia, 6 August 2017. [Google Scholar]
  21. Puterman, M.L. Markov Decision Processes: Discrete Stochastic Dynamic Programming; Wiley-Interscience: Chichester, UK, 1994; p. 694. [Google Scholar]
  22. Watkins, C.J.; Dayan, P. Q-learning. Mach. Learn. 1992, 8, 279–292. [Google Scholar] [CrossRef]
  23. Strens, M. A Bayesian framework for reinforcement learning. In Proceedings of the International Conference on Machine Learning (ICML), Stanford, CA, USA, 29 June 2000; pp. 943–950. [Google Scholar]
  24. Poupart, P.; Vlassis, N.; Hoey, J.; Regan, K. An analytic solution to discrete Bayesian reinforcement learning. In Proceedings of the International Conference on Machine Learning (ICML), Pittsburgh, PA, USA, 25 June 2006; pp. 697–704. [Google Scholar]
  25. Lai, T.; Robbins, H. Asymptotically Efficient Adaptive Allocation Rules. Adv. Appl. Math. 1985, 6, 4–22. [Google Scholar] [CrossRef] [Green Version]
  26. Auer, P.; Ortner, R. Logarithmic online regret bounds for undiscounted reinforcement learning. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Vancouver, BC, Canada, 3 December 2007; pp. 49–56. [Google Scholar]
  27. Dann, C.; Brunskill, E. Sample complexity of episodic fixed-horizon reinforcement learning. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 7 December 2015; pp. 2818–2826. [Google Scholar]
  28. Auer, P.; Cesa-Bianchi, N.; Fischer, P. Finite-time analysis of the multiarmed bandit problem. Mach. Learn. 2002, 47, 235–256. [Google Scholar] [CrossRef]
  29. Ryan, R.M.; Deci, E.L. Intrinsic and extrinsic motivations: Classic definitions and new directions. Contemp. Educ. Psychol. 2000, 25, 54–67. [Google Scholar] [CrossRef]
  30. Strehl, A.L.; Littman, M.L. An analysis of model-based interval estimation for Markov decision processes. J. Comput. Syst. Sci. 2008, 74, 1309–1331. [Google Scholar] [CrossRef]
  31. Raileanu, R.; Rocktäschel, T. RIDE: Rewarding Impact-Driven Exploration for Procedurally-Generated Environments. In Proceedings of the International Conference on Learning Representations (ICLR), Virtual, 26 April 2020. [Google Scholar]
  32. Stadie, B.C.; Levine, S.; Abbeel, P. Incentivizing exploration in reinforcement learning with deep predictive models. In Proceedings of the NIPS Workshop on Deep Reinforcement Learning, Montreal, QC, Canada, 7 December 2015. [Google Scholar]
  33. Schmidhuber, J. A Possibility for lmplementing Curiosity and Boredom in Model-Building Neural Controllers. In Proceedings of the International Conference on Simulation of Adaptive Behavior (SAB), Paris, France, 25 August 1991; pp. 222–227. [Google Scholar]
  34. Schmidhuber, J. Developmental robotics, optimal artificial curiosity, creativity, music, and the fine arts. Connect. Sci. 2006, 18, 173–187. [Google Scholar] [CrossRef]
  35. Pathak, D.; Gandhi, D.; Gupta, A. Self-Supervised Exploration via Disagreement. In Proceedings of the International Conference on Machine Learning (ICML), Long Beach, CA, USA, 9 June 2019. [Google Scholar]
  36. Burda, Y.; Edwards, H.; Storkey, A.; Klimov, O. Exploration by random network distillation. In Proceedings of the International Conference on Learning Representations (ICLR), New Orleans, LA, USA, 6 May 2019. [Google Scholar]
  37. Thompson, W.R. On the likelihood that one unknown probability exceeds another in view of the evidence of two samples. Biometrika 1933, 25, 285–294. [Google Scholar] [CrossRef]
  38. Kaufmann, E.; Korda, N.; Munos, R. Thompson sampling: An asymptotically optimal finite-time analysis. In Proceedings of the International Conference on Algorithmic Learning Theory (ALT), Lyon, France, 29 October 2012. [Google Scholar]
  39. Agrawal, S.; Goyal, N. Further Optimal Regret Bounds for Thompson Sampling. In Proceedings of the International Conference on Artificial Intelligence and Statistics (AISTATS), Scottsdale, AZ, USA, 29 April 2013. [Google Scholar]
  40. Scott, S.L. A modern Bayesian look at the multi-armed bandit. Appl. Stoch. Model. Bus. Ind. 2010, 26, 639–658. [Google Scholar] [CrossRef]
  41. Chapelle, O.; Li, L. An empirical evaluation of Thompson sampling. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Granada, Spain, 12 December 2011; pp. 2249–2257. [Google Scholar]
  42. Russo, D.; Van Roy, B. Learning to optimize via posterior sampling. Math. Oper. Res. 2014, 39, 1221–1243. [Google Scholar] [CrossRef] [Green Version]
  43. Russo, D.; Tse, D.; Van Roy, B. Time-sensitive bandit learning and satisficing thompson sampling. arXiv 2017, arXiv:1704.09028. [Google Scholar]
  44. Russo, D.J.; Van Roy, B.; Kazerouni, A.; Osband, I.; Wen, Z. A tutorial on Thompson sampling. Found. Trends Mach. Learn. 2018, 11, 1–96. [Google Scholar] [CrossRef]
  45. Osband, I.; Russo, D.; Van Roy, B. (More) efficient reinforcement learning via posterior sampling. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Lake Tahoe, CA, USA, 5 December 2013. [Google Scholar]
  46. D’Eramo, C.; Cini, A.; Restelli, M. Exploiting Action-Value Uncertainty to Drive Exploration in Reinforcement Learning. In Proceedings of the International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14 July 2019. [Google Scholar]
  47. Fortunato, M.; Azar, M.G.; Piot, B.; Menick, J.; Osband, I.; Graves, A.; Mnih, V.; Munos, R.; Hassabis, D.; Pietquin, O.; et al. Noisy networks for exploration. In Proceedings of the International Conference on Learning Representations (ICLR), Toulon, France, 24 April 2017. [Google Scholar]
  48. Plappert, M.; Houthooft, R.; Dhariwal, P.; Sidor, S.; Chen, R.Y.; Chen, X.; Asfour, T.; Abbeel, P.; Andrychowicz, M. Parameter Space Noise for Exploration. In Proceedings of the International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, 3 April 2018. [Google Scholar]
  49. Gal, Y.; Ghahramani, Z. Dropout as a Bayesian approximation: Representing model uncertainty in deep learning. In Proceedings of the International Conference on Machine learning (ICML), New York, NY, USA, 19 June 2016. [Google Scholar]
  50. Osband, I. Risk versus uncertainty in deep learning: Bayes, bootstrap and the dangers of dropout. In Proceedings of the NIPS Workshop on Bayesian Deep Learning, Barcelona, Spain, 5 December 2016. [Google Scholar]
  51. Efron, B. The Jackknife, the Bootstrap and Other Resampling Plans; SIAM: Philadelphia, PA, USA, 1982. [Google Scholar]
  52. Osband, I.; Aslanides, J.; Cassirer, A. Randomized prior functions for deep reinforcement learning. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 3 December 2018; pp. 8617–8629. [Google Scholar]
  53. Van Hasselt, H. Double Q-learning. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Vancouver, BC, Canada, 6 December 2010. [Google Scholar]
  54. D’Eramo, C.; Nuara, A.; Pirotta, M.; Restelli, M. Estimating the maximum expected value in continuous reinforcement learning problems. In Proceedings of the International Conference on Artificial Intelligence and Statistics (AISTATS), Fort Lauderdale, FL, USA, 20 April 2017. [Google Scholar]
  55. Tateo, D.; D’Eramo, C.; Nuara, A.; Restelli, M.; Bonarini, A. Exploiting structure and uncertainty of Bellman updates in Markov decision processes. In Proceedings of the Symposium Series on Computational Intelligence (SSCI), Honolulu, HI, USA, 27 November 2017. [Google Scholar]
  56. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; The MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  57. Tang, H.; Houthooft, R.; Foote, D.; Stooke, A.; Chen, O.X.; Duan, Y.; Schulman, J.; DeTurck, F.; Abbeel, P. #Exploration: A study of count-based exploration for deep reinforcement learning. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 24 January 2017; pp. 2753–2762. [Google Scholar]
  58. Ostrovski, G.; Bellemare, M.G.; van den Oord, A.; Munos, R. Count-based exploration with neural density models. In Proceedings of the International Conference on Machine Learning (ICML), Sydney, NSW, Australia, 6 August 2017; pp. 2721–2730. [Google Scholar]
  59. Asadi, K.; Littman, M.L. An alternative softmax operator for reinforcement learning. In Proceedings of the International Conference on Machine Learning (ICML), Sydney, NSW, Australia, 6 August 2017. [Google Scholar]
  60. Sutton, R.S. Integrated architectures for learning, planning, and reacting based on approximating dynamic programming. In Machine Learning; Elsevier: Amsterdam, The Netherlands, 1990; pp. 216–224. [Google Scholar]
Figure 1. A simple problem where naive exploration performs poorly. The agent starts in the second leftmost cell and is rewarded only for finding a treasure. It can move up/down/left/right, but if it moves up or down its position resets. Acting randomly at each step takes on average 4 N attempts to find the rightmost and most valuable reward, where N is the number of cells to the right of the agent. As no reward is given in any intermediate cell, ϵ -greedy exploration is likely to converge to the local optimum represented by the leftmost reward if ϵ decays too quickly.
Figure 1. A simple problem where naive exploration performs poorly. The agent starts in the second leftmost cell and is rewarded only for finding a treasure. It can move up/down/left/right, but if it moves up or down its position resets. Acting randomly at each step takes on average 4 N attempts to find the rightmost and most valuable reward, where N is the number of cells to the right of the agent. As no reward is given in any intermediate cell, ϵ -greedy exploration is likely to converge to the local optimum represented by the leftmost reward if ϵ decays too quickly.
Algorithms 15 00081 g001
Figure 2. Toy problem domain.
Figure 2. Toy problem domain.
Algorithms 15 00081 g002
Figure 3. Toy example (part 1). Visitation count (a) and behavior policies (bd) for the toy problem below. Cells are divided into four triangles, each one representing the count (a) or the behavior Q-value (bd) for the actions “up/down/left/right” in that cell. Arrows denote the action with the highest Q-value. None of the behavior policies exhibits long-term deep exploration. In (1,3) both UCB1 (c) and the augmented Q-table (d) myopically select the action with the smallest immediate count. The latter also acts badly in (1,2) and (2,3). There, it selects “left” which has count one, instead of “up” which has count zero.
Figure 3. Toy example (part 1). Visitation count (a) and behavior policies (bd) for the toy problem below. Cells are divided into four triangles, each one representing the count (a) or the behavior Q-value (bd) for the actions “up/down/left/right” in that cell. Arrows denote the action with the highest Q-value. None of the behavior policies exhibits long-term deep exploration. In (1,3) both UCB1 (c) and the augmented Q-table (d) myopically select the action with the smallest immediate count. The latter also acts badly in (1,2) and (2,3). There, it selects “left” which has count one, instead of “up” which has count zero.
Algorithms 15 00081 g003
Figure 4. Toy example (part 2). Visitation count (a) and behavior policies derived from the proposed W-functions (b,c). Due to the low count, the exploration upper bound dominates the greedy Q-value, and the proposed behavior policies successfully achieve long-term exploration. Both policies prioritize unexecuted actions and avoid terminal and prison states.
Figure 4. Toy example (part 2). Visitation count (a) and behavior policies derived from the proposed W-functions (b,c). Due to the low count, the exploration upper bound dominates the greedy Q-value, and the proposed behavior policies successfully achieve long-term exploration. Both policies prioritize unexecuted actions and avoid terminal and prison states.
Algorithms 15 00081 g004
Figure 5. Chainworld with uniform count.
Figure 5. Chainworld with uniform count.
Algorithms 15 00081 g005
Figure 6. Toy example (part 3). Behavior policies under uniform visitation count. In the top row, reward states (1,1) and (3,1) are terminal. In the bottom row, they are not and the MDP has an infinite horizon. All state-action pairs have been executed once, except for “left” in (1,3). Only the proposed methods (d,e) correctly identify such state-action pair as the most interesting for exploration and propagate this information to other states thanks to TD learning. When the action is executed, all policies converge to the greedy one (a).
Figure 6. Toy example (part 3). Behavior policies under uniform visitation count. In the top row, reward states (1,1) and (3,1) are terminal. In the bottom row, they are not and the MDP has an infinite horizon. All state-action pairs have been executed once, except for “left” in (1,3). Only the proposed methods (d,e) correctly identify such state-action pair as the most interesting for exploration and propagate this information to other states thanks to TD learning. When the action is executed, all policies converge to the greedy one (a).
Algorithms 15 00081 g006
Figure 7. The deep sea.
Figure 7. The deep sea.
Algorithms 15 00081 g007
Figure 8. Results on the deep sea domain averaged over 20 seeds. The proposed exploration achieves the best results, followed by bootstrapped exploration.
Figure 8. Results on the deep sea domain averaged over 20 seeds. The proposed exploration achieves the best results, followed by bootstrapped exploration.
Algorithms 15 00081 g008
Figure 9. The taxi driver.
Figure 9. The taxi driver.
Algorithms 15 00081 g009
Figure 10. Results on the taxi domain averaged over 20 seeds. The proposed algorithms outperform all others, being the only solving the task without optimistic initialization.
Figure 10. Results on the taxi domain averaged over 20 seeds. The proposed algorithms outperform all others, being the only solving the task without optimistic initialization.
Algorithms 15 00081 g010
Figure 11. The deep gridworld.
Figure 11. The deep gridworld.
Algorithms 15 00081 g011
Figure 12. Results on the deep gridworld averaged over 20 seeds. Once again, the proposed algorithms are the only solving the task without optimistic initialization.
Figure 12. Results on the deep gridworld averaged over 20 seeds. Once again, the proposed algorithms are the only solving the task without optimistic initialization.
Algorithms 15 00081 g012
Figure 13. The “toy” gridworld.
Figure 13. The “toy” gridworld.
Algorithms 15 00081 g013
Figure 14. The “prison” gridworld.
Figure 14. The “prison” gridworld.
Algorithms 15 00081 g014
Figure 15. The “wall” gridworld.
Figure 15. The “wall” gridworld.
Algorithms 15 00081 g015
Figure 16. Results on the toy gridworld averaged over 20 seeds. Bootstrapped and bonus-based exploration perform well, but cannot match the proposed one (blue and orange line overlap).
Figure 16. Results on the toy gridworld averaged over 20 seeds. Bootstrapped and bonus-based exploration perform well, but cannot match the proposed one (blue and orange line overlap).
Algorithms 15 00081 g016
Figure 17. The “prison” and distractors affect the performance of all algorithms but ours.
Figure 17. The “prison” and distractors affect the performance of all algorithms but ours.
Algorithms 15 00081 g017
Figure 18. The final gridworld emphasizes the better performance of our algorithms.
Figure 18. The final gridworld emphasizes the better performance of our algorithms.
Algorithms 15 00081 g018
Figure 19. Visitation count at the end of the learning. The seed is the same across all images. Initial states naturally have a higher count than other states. Recall that the upper portion of the deep sea and some gridworlds cells cannot be visited. Only the proposed methods explore the environment uniformly (figures show the count of UCB-based W-function. Count-based W-function performed very similarly). Other algorithms myopically focus on distractors.
Figure 19. Visitation count at the end of the learning. The seed is the same across all images. Initial states naturally have a higher count than other states. Recall that the upper portion of the deep sea and some gridworlds cells cannot be visited. Only the proposed methods explore the environment uniformly (figures show the count of UCB-based W-function. Count-based W-function performed very similarly). Other algorithms myopically focus on distractors.
Algorithms 15 00081 g019
Figure 20. Performance of the our algorithms on varying of γ w on the “toy” gridworld (Figure 14). The higher γ w , the more the agent explores, allowing to discover the reward faster.
Figure 20. Performance of the our algorithms on varying of γ w on the “toy” gridworld (Figure 14). The higher γ w , the more the agent explores, allowing to discover the reward faster.
Algorithms 15 00081 g020
Figure 21. Visitation count for W N β at the end of the learning on the same random seed (out of 20). The higher the discount, the more uniform the count is. The initial state (bottom left) has naturally higher counts because the agent always starts there. W UCB β performed very similarly.
Figure 21. Visitation count for W N β at the end of the learning on the same random seed (out of 20). The higher the discount, the more uniform the count is. The initial state (bottom left) has naturally higher counts because the agent always starts there. W UCB β performed very similarly.
Algorithms 15 00081 g021
Figure 22. Steps to learn on varying the deep sea size N = 5 , 7 , , 59 , averaged over ten runs with 500,000 steps limit. For filled dots, the run always converged, and we report the 95% confidence interval with error bars. For empty dots connected with dashed lines, the algorithm did not learn within the step limit at least once. In this case, the average is over converged runs only and we do not report any confidence interval. Using the exploration bonus (pink) learning barely converged once (most of its dots are either missing or empty). Using bootstrapping (red) all runs converged, but the results show a large confidence interval. With random prior (green) the learning rarely did not converge, and performed very similarly across the random seeds, as shown by its extremely small confidence interval. On the contrary, our method (blue, orange) is the only always converging with a confidence interval close to zero. Indeed, it outperforms all baseline by attaining the lowest sample complexity. Missing algorithms (random, ϵ -greedy, UCB1, Thompson sampling) performed poorly and are not reported.
Figure 22. Steps to learn on varying the deep sea size N = 5 , 7 , , 59 , averaged over ten runs with 500,000 steps limit. For filled dots, the run always converged, and we report the 95% confidence interval with error bars. For empty dots connected with dashed lines, the algorithm did not learn within the step limit at least once. In this case, the average is over converged runs only and we do not report any confidence interval. Using the exploration bonus (pink) learning barely converged once (most of its dots are either missing or empty). Using bootstrapping (red) all runs converged, but the results show a large confidence interval. With random prior (green) the learning rarely did not converge, and performed very similarly across the random seeds, as shown by its extremely small confidence interval. On the contrary, our method (blue, orange) is the only always converging with a confidence interval close to zero. Indeed, it outperforms all baseline by attaining the lowest sample complexity. Missing algorithms (random, ϵ -greedy, UCB1, Thompson sampling) performed poorly and are not reported.
Algorithms 15 00081 g022
Figure 23. The ergodic chainworld.
Figure 23. The ergodic chainworld.
Algorithms 15 00081 g023
Figure 24. Visitation counts for the chainworld with 27 states (x-axis) over time (y-axis). As time passes (top to bottom), the visitation count grows. Only our algorithms and UCB1 explore uniformly, producing a uniform colormap. However, UCB1 finds the last state (with the reward) much later. Once UCB1 finds the reward ((c), step 1000) the visitation count increases only in the last state. This suggests that when UCB1 finds the reward contained in the last state it effectively stops exploring other states. By contrast, our algorithms keep exploring the environment for longer. This allows the agent to learn the true Q-function for all states and explains why our algorithms achieve lower sample complexity and MSVE.
Figure 24. Visitation counts for the chainworld with 27 states (x-axis) over time (y-axis). As time passes (top to bottom), the visitation count grows. Only our algorithms and UCB1 explore uniformly, producing a uniform colormap. However, UCB1 finds the last state (with the reward) much later. Once UCB1 finds the reward ((c), step 1000) the visitation count increases only in the last state. This suggests that when UCB1 finds the reward contained in the last state it effectively stops exploring other states. By contrast, our algorithms keep exploring the environment for longer. This allows the agent to learn the true Q-function for all states and explains why our algorithms achieve lower sample complexity and MSVE.
Algorithms 15 00081 g024
Figure 25. Sample complexity on varying the chain size N = 5 , 7 , , 59 . Error bars denote 95% confidence interval. Only the proposed algorithms (blue and orange lines almost overlap) show little sensitivity to ε , as their sample complexity only slightly increases as ε decreases. By contrast, other algorithms are highly influenced by ε . Their estimate complexity has a large confidence interval, and for smaller ε they rarely learn an ε -optimal policy.
Figure 25. Sample complexity on varying the chain size N = 5 , 7 , , 59 . Error bars denote 95% confidence interval. Only the proposed algorithms (blue and orange lines almost overlap) show little sensitivity to ε , as their sample complexity only slightly increases as ε decreases. By contrast, other algorithms are highly influenced by ε . Their estimate complexity has a large confidence interval, and for smaller ε they rarely learn an ε -optimal policy.
Algorithms 15 00081 g025
Figure 26. Sample complexity against ε , and MSVE over time for 27- and 55-state chainworlds. Shaded areas denote 95% confidence interval. As shown in Figure 25, the sample complexity of our approach is barely influenced by ε . Even for ε = 0 , our algorithms attain low sample complexity, whereas other algorithms complexity is several orders of magnitude higher. Similarly, only our algorithms learn an almost perfect Q-function, achieving an MSVE close to zero.
Figure 26. Sample complexity against ε , and MSVE over time for 27- and 55-state chainworlds. Shaded areas denote 95% confidence interval. As shown in Figure 25, the sample complexity of our approach is barely influenced by ε . Even for ε = 0 , our algorithms attain low sample complexity, whereas other algorithms complexity is several orders of magnitude higher. Similarly, only our algorithms learn an almost perfect Q-function, achieving an MSVE close to zero.
Algorithms 15 00081 g026
Figure 27. Results with stochastic transition function on three of the previous MDPs. Once again, only our algorithms visit all states and learn the optimal policy within steps limit in all MDPs, and their performance is barely affected by the stochasticity.
Figure 27. Results with stochastic transition function on three of the previous MDPs. Once again, only our algorithms visit all states and learn the optimal policy within steps limit in all MDPs, and their performance is barely affected by the stochasticity.
Algorithms 15 00081 g027
Table 1. Results recap for the “zero initialization short horizon”. Only the proposed exploration strategy always discovers all states and solves the tasks in all 20 seeds.
Table 1. Results recap for the “zero initialization short horizon”. Only the proposed exploration strategy always discovers all states and solves the tasks in all 20 seeds.
AlgorithmDiscovery (%)Success (%)
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 99.90 ± 0.01 100 ± 0
Bootstr. (Osband 2016a) 99.77 ± 0.05 100 ± 0
Deep SeaBootstr. (D’Eramo 2019) 63.25 ± 3.31 0 ± 0
UCB1 (Auer 2002) 55.72 ± 0.34 0 ± 0
Expl. Bonus (Strehl 2008) 85.65 ± 1.0 0 ± 0
ϵ -greedy 57.74 ± 1.11 0 ± 0
Random 58.59 ± 1.35 0 ± 0
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 69.60 ± 2.96 13.07 ± 2.96
Bootstr. (Osband 2016a) 52.44 ± 7.55 18.77 ± 9.24
TaxiBootstr. (D’Eramo 2019) 22.44 ± 1.81 1.9 ± 1.47
UCB1 (Auer 2002) 31.17 ± 0.70 1.53 ± 1.37
Expl. Bonus (Strehl 2008) 74.62 ± 2.24 17.6 ± 2.56
ϵ -greedy 29.64 ± 0.98 1.92 ± 1.49
Random 29.56 ± 0.98 1.92 ± 1.49
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 75 ± 9.39 59.03 ± 4.2
Bootstr. (Osband 2016a) 60.82 ± 11.78 63.35 ± 6.89
Deep GridworldBootstr. (D’Eramo 2019) 63.73 ± 10.35 56.85 ± 0.06
UCB1 (Auer 2002) 92.18 ± 0.36 56.88 ± 0
Expl. Bonus (Strehl 2008) 95.45 ± 1.57 69.81 ± 8.84
ϵ -greedy 74.36 ± 4.42 56.88 ± 0
Random 92.45 ± 0.64 56.88 ± 0
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 99.8 ± 0.39 100 ± 0
Bootstr. (Osband 2016a) 99.2 ± 0.72 100 ± 0
Gridworld (Toy)Bootstr. (D’Eramo 2019) 84.8 ± 4.33 40 ± 21.91
UCB1 (Auer 2002) 97.4 ± 0.99 50 ± 22.49
Expl. Bonus (Strehl 2008) 99.8 ± 0.39 100 ± 0
ϵ -greedy 97.4 ± 1.17 45 ± 22.25
Random 97.2 ± 1.15 45 ± 22.25
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 72.39 ± 6.9 44.44 ± 14.69
Bootstr. (Osband 2016a) 64.35 ± 11.1 33.03 ± 8.54
Gridworld (Prison)Bootstr. (D’Eramo 2019) 54.78 ± 6.69 25.61 ± 8.39
UCB1 (Auer 2002) 80.78 ± 2.56 32.31 ± 4.11
Expl. Bonus (Strehl 2008) 85.87 ± 3 46.96 ± 12.35
ϵ -greedy 58.38 ± 5.27 22.84 ± 2.48
Random 76.96 ± 3.08 35.36 ± 10.37
Ours (UCB Reward) 100 ± 0 100 ± 0
Ours (Count Reward) 100 ± 0 100 ± 0
Rand. Prior (Osband 2019) 9.44 ± 3.14 0.14 ± 0.06
Bootstr. (Osband 2016a) 6.35 ± 2.74 0.23 ± 0.08
Gridworld (Wall)Bootstr. (D’Eramo 2019) 12.55 ± 4.45 0.18 ± 0.08
UCB1 (Auer 2002) 24.7 ± 4.45 0.46 ± 0.03
Expl. Bonus (Strehl 2008) 45.16 ± 2.29 0.52 ± 0.04
ϵ -greedy 38.15 ± 1.95 0.39 ± 0.08
Random 65.28 ± 0.45 0.59 ± 0
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Parisi, S.; Tateo, D.; Hensel, M.; D’Eramo, C.; Peters, J.; Pajarinen, J. Long-Term Visitation Value for Deep Exploration in Sparse-Reward Reinforcement Learning. Algorithms 2022, 15, 81. https://doi.org/10.3390/a15030081

AMA Style

Parisi S, Tateo D, Hensel M, D’Eramo C, Peters J, Pajarinen J. Long-Term Visitation Value for Deep Exploration in Sparse-Reward Reinforcement Learning. Algorithms. 2022; 15(3):81. https://doi.org/10.3390/a15030081

Chicago/Turabian Style

Parisi, Simone, Davide Tateo, Maximilian Hensel, Carlo D’Eramo, Jan Peters, and Joni Pajarinen. 2022. "Long-Term Visitation Value for Deep Exploration in Sparse-Reward Reinforcement Learning" Algorithms 15, no. 3: 81. https://doi.org/10.3390/a15030081

APA Style

Parisi, S., Tateo, D., Hensel, M., D’Eramo, C., Peters, J., & Pajarinen, J. (2022). Long-Term Visitation Value for Deep Exploration in Sparse-Reward Reinforcement Learning. Algorithms, 15(3), 81. https://doi.org/10.3390/a15030081

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop