This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Trajectory-wise Iterative Reinforcement Learning Framework for Auto-bidding

Haoming Li Shanghai Jiao Tong UniversityShanghaiChina wakkkka@sjtu.edu.cn Yusen Huo Alibaba GroupBeijingChina huoyusen.huoyusen@alibaba-inc.com Shuai Dou Alibaba GroupBeijingChina doushuai.ds@alibaba-inc.com Zhenzhe Zheng Shanghai Jiao Tong UniversityShanghaiChina zhengzhenzhe@sjtu.edu.cn Zhilin Zhang Alibaba GroupBeijingChina zhangzhilin.pt@alibaba-inc.com Chuan Yu Alibaba GroupBeijingChina yuchuan.yc@alibaba-inc.com Jian Xu Alibaba GroupBeijingChina xiyu.xj@alibaba-inc.com  and  Fan Wu Shanghai Jiao Tong UniversityShanghaiChina fwu@cs.sjtu.edu.cn
(2024)
Abstract.

In online advertising, advertisers participate in ad auctions to acquire ad opportunities, often by utilizing auto-bidding tools provided by demand-side platforms (DSPs). The current auto-bidding algorithms typically employ reinforcement learning (RL). However, due to safety concerns, most RL-based auto-bidding policies are trained in simulation, leading to a performance degradation when deployed in online environments. To narrow this gap, we can deploy multiple auto-bidding agents in parallel to collect a large interaction dataset. Offline RL algorithms can then be utilized to train a new policy. The trained policy can subsequently be deployed for further data collection, resulting in an iterative training framework, which we refer to as iterative offline RL. In this work, we identify the performance bottleneck of this iterative offline RL framework, which originates from the ineffective exploration and exploitation caused by the inherent conservatism of offline RL algorithms. To overcome this bottleneck, we propose Trajectory-wise Exploration and Exploitation (TEE), which introduces a novel data collecting and data utilization method for iterative offline RL from a trajectory perspective. Furthermore, to ensure the safety of online exploration while preserving the dataset quality for TEE, we propose Safe Exploration by Adaptive Action Selection (SEAS). Both offline experiments and real-world experiments on Alibaba display advertising platform demonstrate the effectiveness of our proposed method.

Real-time bidding, Reinforcement Learning, Offline Reinforcement Learning
journalyear: 2024copyright: acmlicensedconference: Proceedings of the ACM Web Conference 2024; May 13–17, 2024; Singapore, Singaporebooktitle: Proceedings of the ACM Web Conference 2024 (WWW ’24), May 13–17, 2024, Singapore, Singaporedoi: 10.1145/3589334.3645534isbn: 979-8-4007-0171-9/24/05ccs: Information systems Computational advertisingccs: Computing methodologies Reinforcement learning

1. Introduction

Online advertising is becoming one of the major sources of profit for Internet companies (Edelman et al., 2007). Due to the complex online advertising environments, auto-bidding tools provided by demand-side platforms (DSPs) are commonly used to bid on behalf of advertisers to optimize their advertising performance. Bidding for arriving ad impressions can be viewed as a sequential decision-making problem, and thus state-of-the-art auto-bidding algorithms leverage reinforcement learning (RL) to optimize bidding policies (Cai et al., 2017; Wu et al., 2018; He et al., 2021).

However, due to safety concerns, current RL-based auto-bidding policies are trained in simulated environments. Policies trained with simulation are shown suboptimal when deployed in the real-world system (Mou et al., 2022). Therefore, it is desirable to optimize the bidding policy by directly interacting with the online environments. Classical (online) RL algorithms iteratively switch between data collection phase and policy update phase, and typically require enormous samples (i.e., transition tuples) to achieve convergence. However, collecting data samples can be extremely time-consuming, e.g., in most RL formulations for auto-bidding (Wu et al., 2018; He et al., 2021), an RL episode corresponds to 24 hours, and thus training a policy may take a long time. Additionally, frequent updates of online policies may cause unstable performance and potentially risky bidding behaviours.

To address these issues, DSPs usually leverage a large number of auto-bidding agents as parallel workers to allow for the efficient collection of large amount of interaction data, and train an auto-bidding policy on the dataset with offline RL (Fujimoto et al., 2019; Levine et al., 2020) — a data-driven RL paradigm that aim to extract policies from large-scale pre-collected datasets. The updated policy could again be deployed for further data collection, which results in an iterative framework for data collection and policy update. We refer to the above training paradigm as iterative offline RL, as depicted in Figure 1. Iterative offline RL presents a promising solution for online policy training in industrial scenarios, and similar ideas have also been mentioned in several academic works (Mou et al., 2022; Zhang et al., 2022; Matsushima et al., 2021).

To facilitate effective policy improvement for each iteration in iterative offline RL, it is crucial for the collected datasets to encompass sufficient information regarding various states and actions. This requires the exploration policy to incorporate a degree of randomness, typically achieved by introducing noise perturbation to actions (Lillicrap et al., 2016; Matsushima et al., 2021; Mou et al., 2022). Nonetheless, employing a random exploration policy can adversely affect the performance of the trained policy. This is because the introduced randomness undermines the exploration policy’s performance, and offline RL algorithms heavily relies on the exploration policy due to the principle of conservatism (or pessimism) (Fujimoto et al., 2019; Kumar et al., 2020; Fujimoto and Gu, 2021; Kostrikov et al., 2021; Rashidinejad et al., 2021), which compels the newly updated policy to be close to the exploration policy. We demonstrate this phenomenon in Section 4, highlighting the challenges of effective exploration and exploitation for iterative offline RL.

In this work, we tackle the aforementioned challenges by adopting a trajectory perspective for both the exploration process (data collection) and the exploitation process (offline RL training). For efficient exploration, we construct exploration policies by introducing noise into the policy’s parameter space instead of the traditional action space. This choice is motivated by our key observation that the injection of parameter space noise (PSN) yields an exploration dataset with a more dispersed trajectory return distribution (please refer to Section 5.1 for details). This observation indicates that the dataset contains a considerable number of high-return trajectories, which are valuable for offline training. For effective exploitation, we propose robust trajectory weighting to fully exploit high-return trajectories in the collected dataset. Specifically, instead of uniformly sampling the dataset during training, we assign high probability weights to high-return trajectories, thereby enhancing the impact of high-quality behaviours on the training process and overcoming the conservatism problem. However, the instability of online environments leads to highly random trajectory returns that often fail to reliably reflect the trajectory qualities. To address this issue, we design a new trajectory quality indicator by approximating the expectation of the stochastic rewards, enhancing the robustness of the trajectory weighting method. We leverage PSN for online exploration and utilize robust trajectory weighting to compute sampling probabilities before offline RL training, boosting the effectiveness of exploration and exploitation in iterative offline RL.

Refer to caption
Figure 1. Iterative offline RL with Trajectory-wise Exploration and Exploitation (TEE) and Safe Exploration by Adaptive Action Selection (SEAS). Components proposed in this work are highlighted in red.

Apart from effectiveness, safety constraints must also be taken into consideration during online exploration in real-world advertising systems. Random exploration can lead to risky bidding behaviours, negatively impacting the performance of advertisers. The safety of an exploration policy is captured by a performance lower bound, which ensures that the performance drop brought by exploration is acceptable. Ensuring safe exploration often requires imposing constraints on the original exploration policy. However, existing safety-guaranteeing methods (Mou et al., 2022) often lack awareness of action qualities. While they prevent dangerous behaviours, they also restrict some high-quality actions. This may hinder the emergence of high-return trajectories during exploration, and subsequently affect the performance of the training process. To preserve data quality while ensuring safety, we propose SEAS, which dynamically determines the safe exploration action at each time step based on the cumulative rewards up to that step, as well as the predicted future return. By making adaptive decisions, SEAS preserves the quality of the collected dataset to the fullest extent, and achieves theoretically guaranteed safety at the same time.

The main contributions of this work are summarized as follows:

  • We identify and demonstrate that the performance bottleneck of the current iterative offline RL paradigm for auto-bidding algorithms mainly lies in ineffective exploration and exploitation caused by the conservatism principle of offline RL algorithms.

  • We propose TEE, a solution for effective exploration and exploitation in iterative offline RL for auto-bidding. TEE comprises two components: PSN for trajectory-wise exploration, and a novel Robust Trajectory Weighting algorithm for trajectory-wise exploitation.

  • For safe exploration in auto-bidding, we design SEAS, which adaptively decides the safe exploration action for each time step based on the cumulative rewards till that step. SEAS can guarantee provable safety and sacrifice minimal performance in policy learning when working together with TEE.

  • Extensive experiments in both simulated environments and Alibaba display advertising platform demonstrate the effectiveness of our solution in terms of trained policy’s performance, as well as the safety of the training process.

2. Related Work

Reinforcement learning for auto-bidding. Bid optimization in online advertising is a sequential decision procedure, and can be solved via reinforcement learning techniques. Cai et. al. (Cai et al., 2017) first formulated the auto-bidding problem as an MDP. Wu et. al. (Wu et al., 2018) and He et. al. (He et al., 2021) leveraged reinforcement learning to optimize bidding policies under various constraints. All of the above works train their RL bidding agent in a simulated environment. Mou et. al. (Mou et al., 2022) recognized the sim2real problem in auto-bidding, and designed an iterative offline RL framework to train the bidding policy online.

Offline RL. Offline RL (Fujimoto et al., 2019; Levine et al., 2020; Prudencio et al., 2023) (or batch RL) refers to the problem of policy optimization utilizing only previously collected data, without additional online interaction. Due to the distribution shift (Fujimoto et al., 2019) problem that arises in offline RL, most algorithms conduct conservative policy learning, which compels the learning policy to stay close to the dataset. Various algorithms achieve this by directly regularizing the actor (Fujimoto et al., 2019; Kumar et al., 2019; Peng et al., 2019; Fujimoto and Gu, 2021), learning conservative value functions (Kumar et al., 2020), or constraining the number of policy improvement steps (Brandfonbrener et al., 2021; Kostrikov et al., 2021). However, the conservatism principle causes the performance of trained policy to highly depend on the dataset quality.

Dataset coverage in offline RL. Dataset coverage is a core factor that limits the performance of offline RL algorithms. Schweighofer et. al. (Schweighofer et al., 2022) conducted extensive experiments to demonstrate that dataset coverage is essential for offline RL algorithms to learn a good policy. Prior theoretical works on offline RL (Chen and Jiang, 2019; Rashidinejad et al., 2021; Zhan et al., 2022) also relied on datasets with sufficient state-action space coverage, which is often characterized by concentrability coefficients, to produce a strong performance guarantee.

Exploration in RL. Exploration is a crucial aspect of RL as it allows the agent to gather information about the environment. Traditional exploration strategies induce novel behaviours by random perturbations of actions, such as ϵ\epsilon-greedy (Sutton and Barto, 2018) and entropy regularization (Williams, 1992). To generate meaningful behavioural patterns for hard exploration tasks, several other approaches such as intrinsic reward-based exploration (Chentanez et al., 2004; Pathak et al., 2017), count-based exploration (Bellemare et al., 2016; Ostrovski et al., 2017) and PSN (Plappert et al., 2018; Fortunato et al., 2018) have been proposed. However, most of the existing solutions have been proposed for the online RL paradigm, which is substantially different from iterative offline RL where data is collected for training offline RL algorithms.

3. Preliminaries

In this work, we consider auto-bidding with budget constraint, a sequential decision problem, where an advertiser submits bids for incoming ad impressions, aiming to maximize the total value within a fixed budget. For this problem, previous works (Zhang et al., 2014, 2016) showed that under the second price auction (10.1257/aer.97.1.242), the optimal bid bb^{*} on an impression is given by b=v/λb^{*}=v/\lambda, where vv represents the impression value (e.g. click-through rate) and λ\lambda is a scaling factor. However, determining the optimal value of λ\lambda in real time is intractable due to its dependence on values and costs of all impressions in the stream. Thus, we formulate the problem of adjusting bidding parameter λ\lambda as a Markov Decision Process (MDP), defined by a tuple (𝒮,𝒜,r,p,γ)(\mathcal{S},\mathcal{A},r,p,\gamma). In our formulation, an episode corresponds to a one-day ad campaign duration, which is divided into TT time steps. At each step tt, the advertiser observes state st𝒮s_{t}\in\mathcal{S} and takes an action at𝒜a_{t}\in\mathcal{A}. The state sts_{t} is a feature vector describing the advertising status of a campaign, which may contain time, remaining budget, budget consumption speed, and etc. The action ata_{t} is the bidding parameter λ\lambda in time step tt. The reward r(st,at)r(s_{t},a_{t}) is the total value of impressions won between time step tt and t+1t+1, and p(st+1|st,at)p(s_{t+1}|s_{t},a_{t}) denotes the transition probability of states. Both rr and pp are determined by the advertising environment. The discount factor γ[0,1]\gamma\in[0,1] accounts for the future rewards’ diminishing impact. A (deterministic) policy πΠ:𝒮𝒜\pi\in\Pi:\mathcal{S}\to\mathcal{A} is a function defining the agent’s bidding behaviour. When a policy interacts with the advertising environment over an episode, a trajectory τ={(st,at,st+1,rt)}t=0T\tau=\{(s_{t},a_{t},s_{t+1},r_{t})\}_{t=0}^{T} is generated111For simplicity, we assume all trajectories’ lengths are equal to episode length TT, though a campaign may exhaust its budget at certain time t0<Tt_{0}<T and cannot afford any impression thereafter. In such cases, we let r(st,at)=0,t0tTr(s_{t},a_{t})=0,\forall t_{0}\leq t\leq T., where the initial state s0s_{0} is drawn from a probability distribution ρ\rho.

The (discounted) return of trajectory τ\tau is R(τ)=t=0TγtrtR(\tau)=\sum_{t=0}^{T}\gamma^{t}r_{t}. The objective of RL is to maximize the expected return:

argmaxπΠJ(π)𝔼τ[R(τ)].\operatorname*{arg\,max}_{\pi\in\Pi}J(\pi)\coloneqq\mathbb{E}_{\tau}[R(\tau)].

In RL, the state value function of a policy π\pi is defined as

Vπ(s)𝔼π[u=tTγuru|st=s],s𝒮.V^{\pi}(s)\coloneqq\mathbb{E}_{\pi}[\sum_{u=t}^{T}\gamma^{u}r_{u}|s_{t}=s],\quad\forall s\in\mathcal{S}.

Similarly, the action value function is

Qπ(s,a)𝔼π[u=tTγuru|st=s,at=a],s𝒮,a𝒜.Q^{\pi}(s,a)\coloneqq\mathbb{E}_{\pi}[\sum_{u=t}^{T}\gamma^{u}r_{u}|s_{t}=s,a_{t}=a],\quad\forall s\in\mathcal{S},a\in\mathcal{A}.

Iterative offline RL follows a cyclical pattern of data collection and offline policy update, repeatedly for a total of KK iterations. Within each iteration k[K]k\in[K], an exploration policy πek\pi^{k}_{e} is constructed based on πk\pi^{k}. Then πek\pi^{k}_{e} is deployed in the advertising environment to collect interaction dataset Dk={τi}i=1ND^{k}=\{\tau_{i}\}_{i=1}^{N} containing NN trajectories. An offline RL algorithm is then used to learn a policy from DkD^{k}, producing the updated policy πk+1\pi^{k+1} for the subsequent iteration.

Safety of Bidding Policies. The performance of auto-bidding policies deployed in the advertising system must be guaranteed in either the stages of policy deployment or policy training. Therefore, safety of the RL training process is formally defined by a performance lower bound:

J(πek)(1ϵ)Js,k=1,,K,J(\pi^{k}_{e})\geq(1-\epsilon)J_{s},\forall k=1,\cdots,K,

where JsJ_{s} is the performance of a known safe policy. 222Note that the exploration policy πek\pi^{k}_{e} in every iteration kk should be ensured to be safe. To satisfy the safety constraint in the first iteration, the training process should be initialized by a known safe (but could be suboptimal) policy instead of a random policy. In practice, policies trained in a simulation (He et al., 2021) could serve as an initial policy.

4. Performance Bottleneck of iterative offline RL

Refer to caption
Figure 2. Performance of exploration policies with different noise scale, as well as the performance of policies trained with IQL on datasets collected by those exploration policies.

In this section, we present empirical observations regarding the performance bottleneck of iterative offline RL, and also introduce the idea of our proposed method. In each iteration of iterative offline RL, the data-collection policy plays a crucial role in determining the input dataset for the subsequent offline training process, which in turn influences the trained policy. The data-collection process should gather sufficient information of the underlying MDP through effective exploration.

A conventional approach for exploration in iterative offline RL (Matsushima et al., 2021) is adding noise perturbations to actions. Concretely, an exploration policy with action space noise (ASN) is constructed as πeASN(st)=π(st)+ϵt\pi_{e}^{ASN}(s_{t})=\pi(s_{t})+\epsilon_{t}, where ϵt𝒩(0,σ2I)\epsilon_{t}\sim\mathcal{N}(0,\sigma^{2}I) is sampled from a Gaussian distribution with standard deviation σ\sigma. However, a highly random policy often leads to suboptimal performance and results in a dataset consisting primarily of low-quality actions. Consequently, such datasets pose challenges for offline RL algorithms to produce high-quality policies in the subsequent training process. This is because the conservatism principle of these algorithms drives the learned policy close to the low-performing exploration policy.

We conduct experiments in a simulated environment (details provided in Appendix D) to illustrate the aforementioned issue. Based on one deterministic policy, we construct exploration policies by adding different levels of noise perturbation on the actions. These exploration policies are deployed in the environment to collect interaction datasets, which are then utilized by IQL (Kostrikov et al., 2021), an offline RL algorithm, to train new policies. Figure 2 presents the performance of exploration policies with varying noise scale, as well as performance of the trained policies. We can observe that adding a certain amount of noise boosts the trained policy by gaining more environment information, while an excessively noisy exploration policy undermines the training result. Furthermore, the optimal noise scale depends on the training algorithm, making online tuning of the noise scale impractical.

One potential method for overcoming the difficulty brought by conservatism is to manually identify high-quality behaviours from the noisy dataset and allow the learning algorithm to focus solely on these behaviours. However, evaluating the quality of individual actions within a dataset can be challenging. For one transition tuple (s,a,r,s)(s,a,r,s^{\prime}), a large reward rr does not necessarily indicate aa to be a good action, due to the influence of ss^{\prime} on future rewards. Hopefully, if a full trajectory τ\tau attains a high return, then it is reasonable to infer that this trajectory contains high-quality behaviours. In following sections, we present empirical findings to reveal that employing PSN instead of ASN for exploration leads to a dataset containing more high-return trajectories.

5. Proposed Framework

In this section, we present our novel design on iterative offline RL framework for auto-bidding, which comprises two key components: TEE and SEAS.

5.1. Trajectory-wise Exploration

We introduce Parameter Space Noise (PSN) (Plappert et al., 2018; Fortunato et al., 2018) for exploration in the iterative offline RL framework, and explain its effectiveness from a trajectory view. PSN refers to injecting noise in an RL policy’s parameter space in order to induce exploratory behaviours. It has brought performance gain on a wide range of control tasks when applied to online deep RL algorithms (Badia et al., 2020; Hessel et al., 2018). For a parameterized policy (e.g. a neural network) π(s;θ)\pi(s;\theta), where θ\theta is the parameter vector, applying additive Gaussian noise to θ\theta gives θ^=θ+ϵ\hat{\theta}=\theta+\epsilon, where ϵ𝒩(0,σ2I)\epsilon\sim\mathcal{N}(0,\sigma^{2}I). Then the exploration policy based on PSN is πePSN(st)=π(st;θ^)\pi_{e}^{PSN}(s_{t})=\pi(s_{t};\hat{\theta}). Importantly, the perturbed parameter vector θ^\hat{\theta} is only sampled at the beginning of each episode and remains fixed afterwards. This is substantially different from ASN where independent noise is added at every time step.

We now present our key observation on datasets collected by PSN by experiments in a simulated bidding environment (details are provided in Appendix D). We first construct two exploration policies with ASN and PSN based on one policy π\pi, and denote them by πeASN\pi_{e}^{ASN} and πePSN\pi_{e}^{PSN} respectively. Subsequently, two datasets DASND^{ASN} and DPSND^{PSN} of equal size are collected with πeASN\pi_{e}^{ASN} and πePSN\pi_{e}^{PSN}. 333To ensure a fair comparison, we control the noise strength of both exploration policies to guarantee that the average return of DASND^{ASN} and DPSND^{PSN} are equal. As shown in Figure 3, the return distribution of trajectories in DPSND^{PSN} is more dispersed than that of DASND^{ASN}. Specifically, DPSND^{PSN} contains high-return trajectories (e.g. trajectories with return higher than 750) which are almost absent in DASND^{ASN}.

Refer to caption
Figure 3. Comparison of trajectory return distributions of datasets collected by ASN and PSN.

The high return variance observed in PSN can be attributed to two main factors: decoupling between the exploration policy and the base policy (i.e. the original noiseless policy π\pi), and the consistency of exploration behaviours. The following example provides further explanation in the context of auto-bidding, by highlighting the incapability of ASN on generating high-return trajectories. Assume that the current policy π\pi is suboptimal in that it produces relatively low bids across most states. In this case, positive perturbations ϵt\epsilon_{t} are desirable when applying ASN. However, when πeASN(st)\pi_{e}^{ASN}(s_{t}) is lifted, π(st+1)\pi(s_{t+1}) would become low since the base policy π\pi aims to maintain the smoothness of budget consumption, thereby offsetting the effect of high bid in step tt. The counterbalancing actions of the base policy substantially prevent ASN from effectively exploring unknown areas. Another issue with ASN is that, since the perturbations ϵt\epsilon_{t} of different time steps are drawn from independent probability distributions, consistently realizing positive ϵt\epsilon_{t} throughout an entire episode is barely possible.

PSN enables us to sample a variety of policy parameters, which could be distributed to a great number of advertising campaigns in a DSP to explore different bidding behaviours. These campaigns run in parallel and collect a large dataset covering a wide range of behaviours, thus achieving large-scale trajectory-wise exploration.

5.2. Trajectory-wise Exploitation

While PSN-based exploration policies could effectively generate valuable high-return trajectories, their collected datasets still consist primarily of inferior trajectories, as depicted in Figure 3. Therefore, the trained policy’s performance is still limited by the inherent conservatism of offline RL. To alleviate this problem and fully exploit the dataset, we propose Robust Trajectory Weighting for trajectory-wise exploitation.

We consider a dataset containing multiple trajectories D={τi}i=1ND=\{\tau_{i}\}_{i=1}^{N}, where τi={(si,t,ai,t,si,t+1,ri,t)}t=0T\tau_{i}=\{(s_{i,t},a_{i,t},s_{i,t+1},r_{i,t})\}_{t=0}^{T}. Inspired by (Yue et al., 2022; Hong et al., 2023; Yue et al., 2023), instead of uniform sampling, we assign large sample probabilities to well-performing trajectories during training. A straightforward way to realize this is to assign weight wiw_{i} for trajectory τi\tau_{i} based on its return RiR(τi)R_{i}\coloneqq R(\tau_{i}), as high return typically indicates good behaviours. Nonetheless, two main issues arise when applying return-based trajectory weighting in the auto-bidding task. Firstly, due to the instability of auction environments and user feedbacks, the reward function r(st,at)r(s_{t},a_{t}) is highly stochastic, thus a high return might not necessarily be achieved by a good policy, but rather a ”lucky” trial that happens to obtain high rewards in most time steps. Secondly, trajectories within the dataset may come from different advertising campaigns, and the intrinsic characteristics of a campaign, such as its budget level and item category, also affect the trajectory return. Thus, directly comparing trajectories from different campaigns by their returns is unfair.

We address the first issue by learning a reward model r¯\bar{r} on the dataset for predicting the expected reward of a state-action pair:

r¯=argminri=1Nt=0T(r(si,t,ai,t)ri,t)2.\bar{r}=\operatorname*{arg\,min}_{r}\mathbb{\displaystyle}\sum_{i=1}^{N}\sum_{t=0}^{T}(r(s_{i,t},a_{i,t})-r_{i,t})^{2}.

The reward model could be implemented by a neural network with the above loss function. Then the original stochastic rewards ri,tr_{i,t} are replaced with their expectations r¯i,t=r¯(si,t,ai,t)\bar{r}_{i,t}=\bar{r}(s_{i,t},a_{i,t}) to calculate R¯i=t=0Tγtr¯i,t\bar{R}_{i}=\sum_{t=0}^{T}\gamma^{t}\bar{r}_{i,t}, producing more robust quality indicators.

To deal with the second issue mentioned above, we regularize the trajectory returns by subtracting the value of the initial state of the trajectory, estimated as V^=argminVi=1N(V(si,0)R¯i)2\hat{V}=\operatorname*{arg\,min}_{V}\sum_{i=1}^{N}(V(s_{i,0})-\bar{R}_{i})^{2}. The initial state typically contains information (e.g. the total budget) of the advertising campaign, therefore V^(si,0)\hat{V}(s_{i,0}) provides estimation of the expected return of the campaign behind trajectory ii. Our final indicator of trajectory quality is A^i\hat{A}_{i}:

A^i=(R¯iV^(si,0))/V^(si,0),1iN.\hat{A}_{i}=(\bar{R}_{i}-\hat{V}(s_{i,0}))/\hat{V}(s_{i,0}),\quad\forall 1\leq i\leq N.

The sample probability wi,tw_{i,t} of transition tuple (si,t,ai,t,si,t+1,ri,t)(s_{i,t},a_{i,t},s_{i,t+1},r_{i,t}) is computed according to A^i\hat{A}_{i} as follows:

wi,texp(A^i/α),1iN,0tT,w_{i,t}\propto\exp(\hat{A}_{i}/\alpha),\quad\forall 1\leq i\leq N,0\leq t\leq T,

where α+\alpha\in\mathbb{R}^{+} is a temperature parameter, and the weights should be normalized to ensure i=1Nt=0Twi,t=1\sum_{i=1}^{N}\sum_{t=0}^{T}w_{i,t}=1.

After computing weights for dataset DD, we could run any model-free offline RL algorithm (e.g. CQL, IQL) using the reweighted data sampling strategy. We provide a theoretical justification of Robust Trajectory Weighting in Appendix A, showing how it alleviates the problem brought by conservative algorithms.

5.3. Safe Exploration

Although TEE boosts the effectiveness of iterative offline RL, the safety of data-collecting policies, which is of great importance when training in real-world advertising systems, have not been considered. In this section, we propose a novel algorithm named SEAS to guarantee the safety of online exploration.

On the problem of safe exploration in auto-bidding, Mou et. al. (Mou et al., 2022) designed a method based on safety zone, to restrict the exploratory actions around a safe policy. Specifically, πe(st)clip(πe(st),πs(st)ξ,πs(st)+ξ)\pi_{e}(s_{t})\leftarrow clip(\pi_{e}(s_{t}),\pi_{s}(s_{t})-\xi,\pi_{s}(s_{t})+\xi), where πs\pi_{s} is a known safe policy. Though this method is provably safe under some assumptions on the MDP, its safety heavily relies on the radius ξ\xi which is intractable to determine. Besides, this approach lacks awareness of the quality of original exploration actions, and poses constraint on both bad and good actions, thus hurting the quality of collected datasets. SEAS mitigates these problems through an adaptive design. The aim of SEAS is to prevent the low-performing trajectories caused by the original exploration policy (e.g. πePSN\pi_{e}^{PSN}) to emerge, while preserving the high-quality ones to the fullest extent.

Algorithm 1 Safe Exploration by Adaptive Action Selection (SEAS)
1:Input: Exploration policy πe\pi_{e}, nn distinct safe policies {πsi}i=1n\{\pi_{s}^{i}\}_{i=1}^{n} and their state-action value function {Qπsi}i=1n\{Q^{\pi_{s}^{i}}\}_{i=1}^{n} , safe performance JsJ_{s} and safety coefficient ϵ(0,1)\epsilon\in(0,1)
2:Initialize: Sample initial state s0ρ(s)s_{0}\sim\rho(s), temp1temp\leftarrow 1
3:for t=0,1,,Tt=0,1,\cdots,T do
4:     aeπe(st),asπstemp(st)a_{e}\leftarrow\pi_{e}(s_{t}),a_{s}\leftarrow\pi_{s}^{temp}(s_{t})
5:     tempargmaxiQπsi(st,ae)temp\leftarrow\operatorname*{arg\,max}_{i}{Q^{\pi_{s}^{i}}(s_{t},a_{e})}, QmaxQπstemp(st,ae)Q_{max}\leftarrow Q^{\pi_{s}^{temp}}(s_{t},a_{e})
6:     if u=0t1ru+Qmax(1ϵ)Js\sum_{u=0}^{t-1}r_{u}+Q_{max}\geq(1-\epsilon)J_{s} then
7:         ataea_{t}\leftarrow a_{e}
8:     else
9:         atasa_{t}\leftarrow a_{s}
10:     end if
11:     Take action ata_{t}, observe rt,st+1r_{t},s_{t+1}
12:end for

The procedure of SEAS interacting with the environment for one episode is shown in Algorithm 1. In each step, SEAS selects between an exploratory action aea_{e} and a safe action asa_{s} according to the condition in line 6. Note that multiple safe policies are provided to the algorithm, and in each step the safe policy with maximum Q value is chosen for constructing the condition. Utilizing multiple safe policies instead of one loosens the restriction in line 6, thus better preserving exploratory actions of πe\pi_{e}. The algorithm is ”adaptive” in the sense that each action ata_{t} depends on the rewards accumulated up to time tt.

The following theorem shows that SEAS theoretically guarantees safety for any exploration policy πe\pi_{e}. Proof of Theorem 1 is provided in Appendix B.

Theorem 1.

For any policy πe\pi_{e} and any ϵ(0,1)\epsilon\in(0,1), given safe policies {πsi}i=1n\{\pi_{s}^{i}\}_{i=1}^{n} that satisfy J(πsi)Js,1inJ(\pi_{s}^{i})\geq J_{s},\forall 1\leq i\leq n, the expected return 𝔼τ[R(τ)]\mathbb{E}_{\tau}[R(\tau)] of trajectories generated by SEAS satisfies 𝔼τ[R(τ)](1ϵ)Js\mathbb{E}_{\tau}[R(\tau)]\geq(1-\epsilon)J_{s}.

The advantages of SEAS are summarized as follows: (i) SEAS needs only one hyperparameter ϵ\epsilon, which is in the definition of safety and is straightforward to set. (ii) The safety of SEAS is provable without additional assumptions on the underlying MDP. (iii) Experiments demonstrate that when functioning together with TEE, SEAS exhibits minimal performance sacrifice compared to baseline methods.

TEE and SEAS together form an iterative RL framework for auto-bidding. Appendix C provides a detailed description of the overall procedure.

6. Experiments

We provide empirical evidence for the effectiveness of our approach by both simulated experiments and real-world experiments on Alibaba display advertising platform.

6.1. Overall Performance in a Simulated Environment

A simulated advertising system is constructed for all the offline experiments in this work. Details of the setup and hyperparameters are provided in Appendix D.

We test the effectiveness of TEE and SEAS by combining them with three different offline RL algorithms. We also implement three baselines: iterative versions of IQL with and without exploration, as well as SORL (Mou et al., 2022). An expert policy is trained with TD3 (Fujimoto et al., 2018), an online RL algorithm, to serve as a performance upper bound. Methods for comparison are listed below.

  • TEE+SEAS+IQL (Kostrikov et al., 2021) / CQL (Kumar et al., 2020) / TD3BC (Fujimoto and Gu, 2021). Iterative offline RL with the proposed TEE and SEAS, using IQL / CQL / TD3BC as the offline RL algorithm.

  • IterIQL+ASN. Iterative offline RL using IQL as the offline RL algorithm. IQL is selected as a representative because it is the state-of-the-art model-free offline RL algorithm. Exploration policy πek\pi_{e}^{k} is constructed by adding ASN on πk\pi^{k}.

  • IterIQL. Iterative offline RL using IQL as the offline RL algorithm. No exploration noise is added, therefore πek=πk\pi_{e}^{k}=\pi^{k}.

  • SORL (Mou et al., 2022). SORL follows the iterative offline RL framework. The authors proposed V-CQL for offline policy training and designed an SER policy for safe and efficient exploration.

Evaluation Metrics. We evaluate the performance of a policy in terms of expected return. Besides, we check the safety of exploration policy by comparing average return in the collected dataset with the safety threshold (1ϵ)Js(1-\epsilon)J_{s}.

Refer to caption
Figure 4. Overall performance in a simulated environment.
Refer to caption
Figure 5. Safety constraint satisfaction.

Figure 4 presents the overall performance through iterations. Our proposed framework, combined with any of the three offline RL algorithms, substantially outperforms the baseline methods and achieves near-expert performance in approximately 5 iterations. Both Iterative IQL+ASN and SORL get stuck in suboptimal policies after limited performance improvements, suffering from the performance bottleneck we discussed in section 4. Additionally, Iterative IQL without exploration achieves barely any performance improvement, which demonstrates the necessity of exploration in iterative offline RL.

We present the performance of exploration policies during the training process in Figure 5. Combined with any offline RL algorithm, our framework consistently ensures the performance of exploration policies to be above the safety threshold (1ϵ)Js(1-\epsilon)J_{s}, which validates the safety guarantee ability of SEAS. We also show the results when SEAS is omitted, in which the safety constraint is violated in early stages.

6.2. Online Experiments

Table 1. Performance of our proposed method in real-world experiments.
iteration BuyCnt ROI CPA ConBdg GMV
1 +1.82% +2.64% -1.81% -0.03% +2.61%
2 +2.21% +2.94% -2.14% +0.02% +2.95%
3 +2.94% +3.33% -1.78% +1.12% +4.49%
4 +3.59% +2.44% -2.60% +0.90% +3.36%

We conduct real-world experiments on Alibaba display advertising platform. In each iteration, we utilize 20000 advertising campaigns to collect data for an epsiode, and employ IQL for offline training. Each episode contains 48 time steps, which means that the bidding parameter is adjusted every 30 minutes. After each policy update, we conduct a 10-day online A/B test using 1500 campaigns.

Evaluation Metrics. The return of the trained policy acts as our main metric of performance, and is referred to as BuyCnt in the online experiments. Additionally, we introduce several other metrics that are commonly used in the auto-bidding field.

  • BuyCut. The total value of ad impressions won by the advertiser. Our objective JJ in the RL formulation.

  • ROI. The ratio between the total revenue and the consumed budget of the advertiser.

  • CPA. Cost per aquisition, defined as the average cost for each successfully converted impression. A smaller CPA indicates a better performance of an auto-bidding policy.

  • ConBdg. The total consumed budget of the advertiser.

  • GMV. Gross Merchandise Volume, the total amount of sales over the campaign duration.

Table 1 shows the performance of our method in each iteration, compared with a static baseline policy trained by CQL(Kumar et al., 2020) on a pre-collected dataset. We can see that the BuyCnt of our policy improves steadily through iterations, and our method consistently outperforms the baseline across all metrics.

6.3. Ablation Studies

We conduct ablation studies for a deeper analysis of the how different components work with each other in our method. Specifically, we aim to answer the following questions: (1) Do trajectory-wise exploration and trajectory-wise exploitation operate in close conjunction instead of being two independent components? (2) Does the reward model effectively reduce the influence of stochastic rewards in Robust Trajectory Weighting? (3) Does SEAS achieve the theoretical safety bound in practice? (4) While achieving the same safety bound, does SEAS sacrifices less performance than other baseline safety-guaranteeing methods?

To answer these questions, we conduct extensive experiments in the simulated environment described in Section 6.1.

To answer Question 1. We develop variants of TEE to delve deeper into how trajectory-wise exploration and trajectory-wise exploitation work together. We focus on one data-collection process followed by one offline training process. We omit SEAS in this experiment, to focus solely on TEE. An IQL policy is used as base policy π\pi. Details of the variants are presented as follows.

  • TEE is implemented as described.

  • w/o T-explore removes trajectory-wise exploration (i.e. PSN). Instead, traditional ASN is used for πe\pi_{e}. To ensure a fair comparison, we control the strength of ASN to guarantee that the average return in DD are equal to that of PSN.

  • w/o T-exploit removes trajectory-wise exploitation (i.e. Robust Trajectory Weighting). After collecting DD by PSN, we train a new policy with uniform sampling.

  • w/o TEE removes TEE. ASN is used for exploration, and uniform sampling for training.

Table 2. Ablation study on TEE. The last column presents the performance improvement over base policy. Best performance of each column is marked as bolded.
Ablation Settings Budget #Improve
1500 2000 2500 3000 avg
Base policy 448.39 514.89 580.18 657.96 550.36 -
TEE 490.19±\pm13.38 569.35±\pm5.89 634.88±\pm31.88 671.68±59.69671.68\pm 59.69 591.53±\pm24.79 +7.48%
w/o T-explore 431.09±11.4431.09\pm 11.4 520.12±12.05520.12\pm 12.05 594.14±13.46594.14\pm 13.46 663.18±14.97663.18\pm 14.97 552.13±7.68552.13\pm 7.68 +0.32%
w/o T-exploit 365.53±10.55365.53\pm 10.55 414.21±29.05414.21\pm 29.05 495.89±39.82495.89\pm 39.82 571.24±51.27571.24\pm 51.27 461.72±29.96461.72\pm 29.96 -16.10%
w/o TEE 414.38±30.74414.38\pm 30.74 515.66±10.35515.66\pm 10.35 607.47±8.91607.47\pm 8.91 678.67±\pm17.53 554.04±7.13554.04\pm 7.13 +0.66%

Table 2 presents the performance of policies trained under different settings, and their performance gain over the base policy. We observe that TEE achieves significant performance improvement over the base policy under various budgets, while the absence of either component hurts its effectiveness. Interestingly, eliminating Trajectory-wise Exploitation leads to a substantial performance degradation. The reason behind this is that datasets collected by PSN contains a large fraction of undesirable actions, which are imitated by conservative algorithms. The above observation indicates that Trajectory-wise Exploration and Trajectory-wise Exploitation are inherently interconnected components, and their combination contributes substantially to the enhancement of the algorithm’s performance.

To answer Question 2. We test the effectiveness of reward model in environments with different degrees of stochasticity. We first construct simulated environments with various stochasticity by controlling the variance of impression numbers per time step. In each environment, we collect an exploratory dataset with the same policy, and use two different data sampling strategies to train a policy on the dataset: a) Our proposed robust trajectory weighting. b) Trajectory weighting with the raw rewards ri,tr_{i,t} instead of the reward model’s prediction r¯i,t\bar{r}_{i,t}. We compare the performance of trained policies, to examine how reward model benefits the trajectory weighting method.

Refer to caption
(a) Improvements of policy performance.
Figure 6. The effectiveness of reward model in environments with different degrees of stochasticity.

Degree #\# of Impr.
low 175
medium-low U[144,206]U_{[144,206]}
medium U[113,237]U_{[113,237]}
medium-high U[82,268]U_{[82,268]}
high U[50,300]U_{[50,300]}
(b) Settings of different environments.

The results of the experiments are shown in Figure 6(a), where RrobustR_{robust} denotes the return of policy trained with Robust Trajectory Weighting, RrawR_{raw} represents the return of policy trained with raw rewards, and RbaseR_{base} the return of the data-collecting policy. Figure 6(b) shows the probability distributions of the impression number, where U[a,b]U_{[a,b]} denotes a uniform distribution over [a,b][a,b]. The red line in Figure 6(a) indicates that the reward model is increasingly useful as the stochasticity of the environment intensifies. Moreover, the blue bars in the figure exhibit values lower than 1 in high stochasticity instances, which reflects that raw stochastic rewards could be misleading signals for trajectory weighting.

To answer Question 3. We fully evaluate the safety-ensuring ability of SEAS by setting different values of ϵ\epsilon and observe the rate of performance decrease 1J(πek)/Js1-J(\pi_{e}^{k})/J_{s}, where πek\pi_{e}^{k} is the policy produced by SEAS. We expect the safety constraint 1J(πek)/Jsϵ1-J(\pi_{e}^{k})/J_{s}\leq\epsilon to be satisfied. In this experiment, we take a USCB (He et al., 2021) policy as the safe policy, and obtain its QQ function through fitting a dataset collected by itself.

Table 3. Safety-ensuring ability of SEAS.
ϵ\epsilon 0.4 0.3 0.2 0.1 0.05 0.01
1J(πek)/Js1-J(\pi_{e}^{k})/J_{s} 0.202 0.137 0.039 -0.002 -0.004 -0.005

From Table 3, we observe that SEAS consistently satisfies the safety constraint over a wide range of input ϵ\epsilon. Interestingly, for small ϵ\epsilon values, the exploration policy generated by SEAS even outperforms the base policy.

To answer Question 4. We compare SEAS with two baseline safety-ensuring methods in terms of performance sacrificing, while ensuring safety to the same degree (ϵ=0.05\epsilon=0.05). One USCB (He et al., 2021) policy is leveraged as the safe policy. In this experiment, we start from an IQL policy, preserve the design of TEE, and substitute SEAS with baselines presented below. Strength of PSN is σ=0.05\sigma=0.05.

  • Small noise. We limit the strength of PSN at a low level, by setting σ=0.01\sigma=0.01.

  • Fixed range. As proposed in (Mou et al., 2022), the safe exploratory action aea_{e} is given by aeclip(πe(st),πs(st)ξ,πs(st)+ξ)a_{e}\leftarrow clip(\pi_{e}(s_{t}),\pi_{s}(s_{t})-\xi,\pi_{s}(s_{t})+\xi). The safe policy πs\pi_{s} is the same as that in SEAS, and ξ\xi is set to 0.10.1.

Table 4. Different safety ensuring methods’ impact on performance of trained policy. The third column presents the performance improvement of trained policy over base policy.
Safe Exploration Methods Return #Improve
No constraint 594.25±18.28594.25\pm 18.28 +15.16%
SEAS 572.35±\pm 20.42 +10.92%
Small noise 528.07±19.45528.07\pm 19.45 +2.34%
Fixed range 532.91±6.74532.91\pm 6.74 +3.28%

The results are presented in Table 4. We also show the result of not imposing any safety constraint, which is a performance upper bound. SEAS significantly outperforms the two baselines in terms of the average return of the trained policy. This observation suggests that the adaptive design of SEAS allows for minimal performance sacrifice.

7. Conclusion

In this work, we have presented a new iterative RL framework for auto-bidding from a trajectory perspective. We also pay particular attention on the safety of exploration in online advertisement systems, and propose SEAS for this issue. Through comprehensive experiments, our method has been shown to be effective, achieving superior results compared to other baselines. In future work, we plan to test the effectiveness of TEE in other fields such as recommender systems and healthcare, where online policy training is challenging but urgently needed.

Acknowledgements.
This work was supported in part by National Key R&D Program of China (No. 2022ZD0119100), in part by China NSF grant No. 62322206, 62132018, 62025204, U2268204, 62272307, 62372296, 61972-254, 61972252, in part by Alibaba Group through Alibaba Innovative Research Program. The opinions, findings, conclusions, and recommendations expressed in this paper are those of the authors and do not necessarily reflect the views of the funding agencies or the government.

References

  • (1)
  • Badia et al. (2020) Adrià Puigdomènech Badia, Bilal Piot, Steven Kapturowski, Pablo Sprechmann, Alex Vitvitskyi, Zhaohan Daniel Guo, and Charles Blundell. 2020. Agent57: Outperforming the atari human benchmark. In International conference on machine learning. PMLR, 507–517.
  • Bellemare et al. (2016) Marc Bellemare, Sriram Srinivasan, Georg Ostrovski, Tom Schaul, David Saxton, and Remi Munos. 2016. Unifying count-based exploration and intrinsic motivation. Advances in neural information processing systems 29 (2016).
  • Brandfonbrener et al. (2021) David Brandfonbrener, Will Whitney, Rajesh Ranganath, and Joan Bruna. 2021. Offline rl without off-policy evaluation. Advances in neural information processing systems 34 (2021), 4933–4946.
  • Cai et al. (2017) Han Cai, Kan Ren, Weinan Zhang, Kleanthis Malialis, Jun Wang, Yong Yu, and Defeng Guo. 2017. Real-time bidding by reinforcement learning in display advertising. In Proceedings of the tenth ACM international conference on web search and data mining. 661–670.
  • Chen and Jiang (2019) Jinglin Chen and Nan Jiang. 2019. Information-Theoretic Considerations in Batch Reinforcement Learning. In Proceedings of the 36th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 97). PMLR, 1042–1051. https://proceedings.mlr.press/v97/chen19e.html
  • Chentanez et al. (2004) Nuttapong Chentanez, Andrew Barto, and Satinder Singh. 2004. Intrinsically motivated reinforcement learning. Advances in neural information processing systems 17 (2004).
  • Edelman et al. (2007) Benjamin Edelman, Michael Ostrovsky, and Michael Schwarz. 2007. Internet advertising and the generalized second-price auction: Selling billions of dollars worth of keywords. American economic review 97, 1 (2007), 242–259.
  • Fortunato et al. (2018) Meire Fortunato, Mohammad Gheshlaghi Azar, Bilal Piot, Jacob Menick, Matteo Hessel, Ian Osband, Alex Graves, Volodymyr Mnih, Rémi Munos, Demis Hassabis, Olivier Pietquin, Charles Blundell, and Shane Legg. 2018. Noisy Networks For Exploration. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings.
  • Fujimoto and Gu (2021) Scott Fujimoto and Shixiang Shane Gu. 2021. A minimalist approach to offline reinforcement learning. Advances in neural information processing systems 34 (2021), 20132–20145.
  • Fujimoto et al. (2019) Scott Fujimoto, David Meger, and Doina Precup. 2019. Off-policy deep reinforcement learning without exploration. In International conference on machine learning. PMLR, 2052–2062.
  • Fujimoto et al. (2018) Scott Fujimoto, Herke van Hoof, and David Meger. 2018. Addressing Function Approximation Error in Actor-Critic Methods. In Proceedings of the 35th International Conference on Machine Learning (Proceedings of Machine Learning Research, Vol. 80), Jennifer Dy and Andreas Krause (Eds.). PMLR, 1587–1596. https://proceedings.mlr.press/v80/fujimoto18a.html
  • He et al. (2021) Yue He, Xiujun Chen, Di Wu, Junwei Pan, Qing Tan, Chuan Yu, Jian Xu, and Xiaoqiang Zhu. 2021. A unified solution to constrained bidding in online display advertising. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining. 2993–3001.
  • Hessel et al. (2018) Matteo Hessel, Joseph Modayil, Hado Van Hasselt, Tom Schaul, Georg Ostrovski, Will Dabney, Dan Horgan, Bilal Piot, Mohammad Azar, and David Silver. 2018. Rainbow: Combining improvements in deep reinforcement learning. In Proceedings of the AAAI conference on artificial intelligence, Vol. 32.
  • Hong et al. (2023) Zhang-Wei Hong, Pulkit Agrawal, Remi Tachet des Combes, and Romain Laroche. 2023. Harnessing Mixed Offline Reinforcement Learning Datasets via Trajectory Weighting. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023.
  • Kostrikov et al. (2021) Ilya Kostrikov, Ashvin Nair, and Sergey Levine. 2021. Offline Reinforcement Learning with Implicit Q-Learning. In International Conference on Learning Representations.
  • Kumar et al. (2019) Aviral Kumar, Justin Fu, Matthew Soh, George Tucker, and Sergey Levine. 2019. Stabilizing off-policy q-learning via bootstrapping error reduction. Advances in Neural Information Processing Systems 32 (2019).
  • Kumar et al. (2020) Aviral Kumar, Aurick Zhou, George Tucker, and Sergey Levine. 2020. Conservative q-learning for offline reinforcement learning. Advances in Neural Information Processing Systems 33 (2020), 1179–1191.
  • Levine et al. (2020) Sergey Levine, Aviral Kumar, George Tucker, and Justin Fu. 2020. Offline reinforcement learning: Tutorial, review, and perspectives on open problems. arXiv preprint arXiv:2005.01643 (2020).
  • Lillicrap et al. (2016) Timothy P. Lillicrap, Jonathan J. Hunt, Alexander Pritzel, Nicolas Heess, Tom Erez, Yuval Tassa, David Silver, and Daan Wierstra. 2016. Continuous control with deep reinforcement learning. In 4th International Conference on Learning Representations, ICLR 2016, San Juan, Puerto Rico, May 2-4, 2016, Conference Track Proceedings, Yoshua Bengio and Yann LeCun (Eds.).
  • Matsushima et al. (2021) Tatsuya Matsushima, Hiroki Furuta, Yutaka Matsuo, Ofir Nachum, and Shixiang Gu. 2021. Deployment-Efficient Reinforcement Learning via Model-Based Offline Optimization. In 9th International Conference on Learning Representations, ICLR 2021, Virtual Event, Austria, May 3-7, 2021.
  • Mou et al. (2022) Zhiyu Mou, Yusen Huo, Rongquan Bai, Mingzhou Xie, Chuan Yu, Jian Xu, and Bo Zheng. 2022. Sustainable Online Reinforcement Learning for Auto-bidding. In Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, NeurIPS 2022, New Orleans, LA, USA, November 28 - December 9, 2022.
  • Ostrovski et al. (2017) Georg Ostrovski, Marc G Bellemare, Aäron Oord, and Rémi Munos. 2017. Count-based exploration with neural density models. In International conference on machine learning. PMLR, 2721–2730.
  • Pathak et al. (2017) Deepak Pathak, Pulkit Agrawal, Alexei A Efros, and Trevor Darrell. 2017. Curiosity-driven exploration by self-supervised prediction. In International conference on machine learning. PMLR, 2778–2787.
  • Peng et al. (2019) Xue Bin Peng, Aviral Kumar, Grace Zhang, and Sergey Levine. 2019. Advantage-weighted regression: Simple and scalable off-policy reinforcement learning. arXiv preprint arXiv:1910.00177 (2019).
  • Plappert et al. (2018) Matthias Plappert, Rein Houthooft, Prafulla Dhariwal, Szymon Sidor, Richard Y. Chen, Xi Chen, Tamim Asfour, Pieter Abbeel, and Marcin Andrychowicz. 2018. Parameter Space Noise for Exploration. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings.
  • Prudencio et al. (2023) Rafael Figueiredo Prudencio, Marcos ROA Maximo, and Esther Luna Colombini. 2023. A survey on offline reinforcement learning: Taxonomy, review, and open problems. IEEE Transactions on Neural Networks and Learning Systems (2023).
  • Rashidinejad et al. (2021) Paria Rashidinejad, Banghua Zhu, Cong Ma, Jiantao Jiao, and Stuart Russell. 2021. Bridging offline reinforcement learning and imitation learning: A tale of pessimism. Advances in Neural Information Processing Systems 34 (2021), 11702–11716.
  • Schweighofer et al. (2022) Kajetan Schweighofer, Marius-constantin Dinu, Andreas Radler, Markus Hofmarcher, Vihang Prakash Patil, Angela Bitto-Nemling, Hamid Eghbal-zadeh, and Sepp Hochreiter. 2022. A dataset perspective on offline reinforcement learning. In Conference on Lifelong Learning Agents. PMLR, 470–517.
  • Sutton and Barto (2018) Richard S Sutton and Andrew G Barto. 2018. Reinforcement learning: An introduction. MIT press.
  • Williams (1992) Ronald J Williams. 1992. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine learning 8 (1992), 229–256.
  • Wu et al. (2018) Di Wu, Xiujun Chen, Xun Yang, Hao Wang, Qing Tan, Xiaoxun Zhang, Jian Xu, and Kun Gai. 2018. Budget constrained bidding by model-free reinforcement learning in display advertising. In Proceedings of the 27th ACM International Conference on Information and Knowledge Management. 1443–1451.
  • Yue et al. (2023) Yang Yue, Bingyi Kang, Xiao Ma, Gao Huang, Shiji Song, and Shuicheng Yan. 2023. Offline Prioritized Experience Replay. arXiv preprint arXiv:2306.05412 (2023).
  • Yue et al. (2022) Yang Yue, Bingyi Kang, Xiao Ma, Zhongwen Xu, Gao Huang, and Shuicheng Yan. 2022. Boosting offline reinforcement learning via data rebalancing. arXiv preprint arXiv:2210.09241 (2022).
  • Zhan et al. (2022) Wenhao Zhan, Baihe Huang, Audrey Huang, Nan Jiang, and Jason Lee. 2022. Offline Reinforcement Learning with Realizability and Single-policy Concentrability. In Proceedings of Thirty Fifth Conference on Learning Theory (Proceedings of Machine Learning Research, Vol. 178), Po-Ling Loh and Maxim Raginsky (Eds.). PMLR, 2730–2775. https://proceedings.mlr.press/v178/zhan22a.html
  • Zhang et al. (2022) Ruiyi Zhang, Tong Yu, Yilin Shen, and Hongxia Jin. 2022. Text-Based Interactive Recommendation via Offline Reinforcement Learning. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 36. 11694–11702.
  • Zhang et al. (2016) Weinan Zhang, Kan Ren, and Jun Wang. 2016. Optimal real-time bidding frameworks discussion. arXiv preprint arXiv:1602.01007 (2016).
  • Zhang et al. (2014) Weinan Zhang, Shuai Yuan, and Jun Wang. 2014. Optimal real-time bidding for display advertising. In Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining. 1077–1086.

Appendix A Theoretical Justification of Robust Trajectory Weighting

We formally show that under the assumption of deterministic transition and stochastic reward, applying Robust Trajectory Weighting is equivalent to training offline RL on a dataset collected by a better behaviour policy.

For a dataset D={τi}i=1ND=\{\tau_{i}\}_{i=1}^{N} where τi={(si,t,ai,t,si,t+1,ri,t)}t=0T\tau_{i}=\{(s_{i,t},a_{i,t},s_{i,t+1},r_{i,t})\}_{t=0}^{T}. Each trajectory τi\tau_{i} is collected by a different deterministic policy πi\pi_{i}, as in the case of Trajectory-wise Exploration. The behaviour policy π\pi of DD is then defined as sampling a policy from {πi}i=1N\{\pi_{i}\}_{i=1}^{N} uniformly at the start of an episode, then acting according to the sampled policy till the end of the episode. Similarly, we define a weighted behaviour policy π\pi^{\prime} as first sampling a policy from {πi}i=1N\{\pi_{i}\}_{i=1}^{N} according to probabilities {wi}i=1N\{w_{i}\}_{i=1}^{N}, then acting with it. We aim to show that J(π)J(π)J(\pi^{\prime})\geq J(\pi).

The performance J(π)J(\pi) could be expressed as expectation of value function over all possible initial states:

(1) J(π)=𝔼s0ρ[Vπ(s0)],J(π)=𝔼s0ρ[Vπ(s0)].J(\pi)=\mathbb{E}_{s_{0}\sim\rho}[V^{\pi}(s_{0})],J(\pi^{\prime})=\mathbb{E}_{s_{0}\sim\rho}[V^{\pi^{\prime}}(s_{0})].

For any initial state s0s_{0}, let Gs0={i|si,0=s0}G_{s_{0}}=\{i|s_{i,0}=s_{0}\}, Ns0=|Gs0|N_{s_{0}}=|G_{s_{0}}|, we have

(2) Vπ(s0)=iGs0Vπi(si,0)/Ns0.V^{\pi}(s_{0})=\sum_{i\in G_{s_{0}}}V^{\pi_{i}}(s_{i,0})/N_{s_{0}}.

Similarly, for weighted behaviour policy π\pi^{\prime},

(3) Vπ(s0)=iGs0wiVπi(si,0)/iGs0wi.V^{\pi^{\prime}}(s_{0})=\sum_{i\in G_{s_{0}}}w_{i}V^{\pi_{i}}(s_{i,0})/\sum_{i\in G_{s_{0}}}w_{i}.

Assuming deterministic transition and stochastic rewards of the underlying MDP, we have Vπi(si,0)=R¯iV^{\pi_{i}}(s_{i,0})=\bar{R}_{i}, where R¯i=t=0Tγtr¯i,t\bar{R}_{i}=\sum_{t=0}^{T}\gamma^{t}\bar{r}_{i,t} is the relabeled return in Robust Trajectory Weighting. Plugging it in (2) and (3) gives us

(4) Vπ(s0)=iGs0R¯i/Ns0,V^{\pi}(s_{0})=\sum_{i\in G_{s_{0}}}\bar{R}_{i}/N_{s_{0}},
(5) Vπ(s0)=iGs0wiR¯i/iGs0wi.V^{\pi^{\prime}}(s_{0})=\sum_{i\in G_{s_{0}}}w_{i}\bar{R}_{i}/\sum_{i\in G_{s_{0}}}w_{i}.

Subtracting (4) by (5),

(6) Vπ(s0)Vπ(s0)\displaystyle V^{\pi^{\prime}}(s_{0})-V^{\pi}(s_{0})
(7) =\displaystyle= iGs0wiR¯i/iGs0wiiGs0R¯i/Ns0\displaystyle\sum_{i\in G_{s_{0}}}w_{i}\bar{R}_{i}/\sum_{i\in G_{s_{0}}}w_{i}-\sum_{i\in G_{s_{0}}}\bar{R}_{i}/N_{s_{0}}
(8) =\displaystyle= 1Ns0iGs0wi(Ns0iGs0wiR¯i(iGs0wi)(iGs0R¯i))\displaystyle\frac{1}{N_{s_{0}}\sum_{i\in G_{s_{0}}}w_{i}}{(N_{s_{0}}\sum_{i\in G_{s_{0}}}w_{i}\bar{R}_{i}-(\sum_{i\in G_{s_{0}}}w_{i})(\sum_{i\in G_{s_{0}}}\bar{R}_{i}))}
(9) =\displaystyle= 1Ns0iGs0wi(i,jGs0,i<j(wiwj)(R¯iR¯j)).\displaystyle\frac{1}{N_{s_{0}}\sum_{i\in G_{s_{0}}}w_{i}}{(\sum_{i,j\in G_{s_{0}},i<j}(w_{i}-w_{j})(\bar{R}_{i}-\bar{R}_{j}))}.

In (9), each term (wiwj)(R¯iR¯j)(w_{i}-w_{j})(\bar{R}_{i}-\bar{R}_{j}) inside the summation is non-negative, because

wi=exp((R¯i/V(si,0)1)/α)/Z,w_{i}=\exp((\bar{R}_{i}/V(s_{i,0})-1)/\alpha)/Z,

where Z=i=1NwiZ=\sum_{i=1}^{N}w_{i} is a normalization term and α+\alpha\in\mathbb{R}^{+}, is non-decreasing with respect to R¯i\bar{R}_{i}. Therefore, Vπ(s0)Vπ(s0)0V^{\pi^{\prime}}(s_{0})-V^{\pi}(s_{0})\geq 0 for every initial state s0s_{0}. Then applying (1) gives J(π)J(π)J(\pi^{\prime})\geq J(\pi).

The above derivation also highlights the significance of our proposed reward model. Relabeling rewards with the reward model’s output makes Vπi(si,0)=R¯iV^{\pi_{i}}(s_{i,0})=\bar{R}_{i}, allowing us to deal with stochastic rewards.

Appendix B Proof of Theorem 1

See 1

Proof.

In each time step, SEAS takes either action aea_{e} or asa_{s}. For a trajectory τ\tau generated by SEAS, let t0t_{0} be the last step it takes exploratory action aea_{e}. For those trajectories where it never takes aea_{e}, set t0t_{0} to 0. Let tempt0temp_{t_{0}} denote the value of temptemp at step t0t_{0}. Let 𝟙(τ,s,a,i)\mathbb{1}(\tau,s,a,i) be the indicator function of st0,at0s_{t_{0}},a_{t_{0}} and tempt0temp_{t_{0}} for trajectory τ\tau, defined as follows:

𝟙(τ,s,a,i)={δ(sst0,aat0), if i=tempt00, otherwise,\mathbb{1}(\tau,s,a,i)=\left\{\begin{aligned} \delta(s-s_{t_{0}},a-a_{t_{0}}),&\text{ if }i=temp_{t_{0}}\\ 0,&\text{ otherwise,}\end{aligned}\right.

where δ()\delta(\cdot) is the Dirac delta function. Therefore

(10) τ,i=1n𝒮𝒜𝟙(τ,s,a,i)𝑑a𝑑s=1.\forall\tau,\quad\sum_{i=1}^{n}\int_{\mathcal{S}}\int_{\mathcal{A}}\mathbb{1}(\tau,s,a,i)dads=1.

Multiply both sides of (10) by R(τ)R(\tau),

τ,i=1n𝒮𝒜R(τ)𝟙(τ,s,a,i)𝑑a𝑑s=R(τ).\forall\tau,\quad\sum_{i=1}^{n}\int_{\mathcal{S}}\int_{\mathcal{A}}R(\tau)\mathbb{1}(\tau,s,a,i)dads=R(\tau).

Take expectation with respect to τ\tau on both sides,

𝔼τ[R(τ)]\displaystyle\mathbb{E}_{\tau}[R(\tau)] =i=1n𝒮𝒜𝔼τ[R(τ)𝟙(τ,s,a,i)]𝑑a𝑑s.\displaystyle=\sum_{i=1}^{n}\int_{\mathcal{S}}\int_{\mathcal{A}}\mathbb{E}_{\tau}[R(\tau)\mathbb{1}(\tau,s,a,i)]dads.

Split trajectory τ\tau by t0t_{0}, and define R(τ)=u=0t01ruR(\tau^{-})=\sum_{u=0}^{t_{0}-1}r_{u}, R(τ+)=u=t0TruR(\tau^{+})=\sum_{u={t_{0}}}^{T}r_{u}, then

𝔼τ[R(τ)]\displaystyle\mathbb{E}_{\tau}[R(\tau)] =i=1n𝒮𝒜𝔼τ[(R(τ)+R(τ+))𝟙(τ,s,a,i)]𝑑a𝑑s\displaystyle=\sum_{i=1}^{n}\int_{\mathcal{S}}\int_{\mathcal{A}}\mathbb{E}_{\tau}[(R(\tau^{-})+R(\tau^{+}))\mathbb{1}(\tau,s,a,i)]dads
=i=1n𝒮𝒜{𝔼τ[R(τ)𝟙(τ,s,a,i)]\displaystyle=\sum_{i=1}^{n}\int_{\mathcal{S}}\int_{\mathcal{A}}\{\mathbb{E}_{\tau}[R(\tau^{-})\mathbb{1}(\tau,s,a,i)]
+𝔼τ[R(τ+)𝟙(τ,s,a,i)]}dads.\displaystyle\qquad\qquad\quad+\mathbb{E}_{\tau}[R(\tau^{+})\mathbb{1}(\tau,s,a,i)]\}dads.

For trajectory τ\tau , since t0t_{0} is the last time it takes aea_{e}, all actions after t0t_{0} follows safe policy πstemp0\pi_{s}^{temp_{0}}. Therefore Eτ[R(τ+)𝟙(τ,s,a,i)]=Qπsi(s,a)𝟙(τ,s,a,i)E_{\tau}[R(\tau^{+})\mathbb{1}(\tau,s,a,i)]=Q^{\pi_{s}^{i}}(s,a)\mathbb{1}(\tau,s,a,i), then

𝔼τ[R(τ)]\displaystyle\mathbb{E}_{\tau}[R(\tau)] =i=1n𝒮𝒜{𝔼τ[R(τ)𝟙(τ,s,a,i)]\displaystyle=\sum_{i=1}^{n}\int_{\mathcal{S}}\int_{\mathcal{A}}\{\mathbb{E}_{\tau}[R(\tau^{-})\mathbb{1}(\tau,s,a,i)]
+Qπsi(s,a)𝟙(τ,s,a,i)}dads\displaystyle\qquad\qquad\quad+Q^{\pi_{s}^{i}}(s,a)\mathbb{1}(\tau,s,a,i)\}dads
=i=1n𝒮𝒜𝔼τ[(R(τ)+Qπsi(s,a))𝟙(τ,s,a,i)]𝑑a𝑑s\displaystyle=\sum_{i=1}^{n}\int_{\mathcal{S}}\int_{\mathcal{A}}\mathbb{E}_{\tau}[(R(\tau^{-})+Q^{\pi_{s}^{i}}(s,a))\mathbb{1}(\tau,s,a,i)]dads
i=1n𝒮𝒜𝔼τ[(1ϵ)Js𝟙(τ,s,a,i)]𝑑a𝑑s\displaystyle\geq\sum_{i=1}^{n}\int_{\mathcal{S}}\int_{\mathcal{A}}\mathbb{E}_{\tau}[(1-\epsilon)J_{s}\mathbb{1}(\tau,s,a,i)]dads
=𝔼τ[i=1n𝒮𝒜(1ϵ)Js𝟙(τ,s,a,i)𝑑a𝑑s]\displaystyle=\mathbb{E}_{\tau}[\sum_{i=1}^{n}\int_{\mathcal{S}}\int_{\mathcal{A}}(1-\epsilon)J_{s}\mathbb{1}(\tau,s,a,i)dads]
=𝔼τ[(1ϵ)Js]\displaystyle=\mathbb{E}_{\tau}[(1-\epsilon)J_{s}]
=(1ϵ)Js,\displaystyle=(1-\epsilon)J_{s},

where the inequality step is by line 6 in algorithm 1, and the following steps are from (10) and simple algebra. ∎

Appendix C Overall Iterative Framework

TEE and SEAS are combined to form an iterative framework for online policy training in auto-bidding. The training process is initialized with the current policy running in the bidding system, which is typically suboptimal but safe. In each iteration kk, we employ PSN for exploration based on the current policy πk\pi^{k}. We pick a large number of advertising campaigns, and independently sample parameter vectors for different campaigns’ policies. The exploration policies are input to SEAS to guarantee safety. A subset of previous policies {πκ}κ=1k1\{\pi^{\kappa}\}_{\kappa=1}^{k-1} can be utilized as safe policies for SEAS, and ϵ\epsilon is a pre-defined constant parameter through iterations. The Q functions of safe policies for SEAS could be obtained through different approaches. For policies trained by value-based or actor-critic RL algorithms, we can directly query the existing value network. Alternatively, new value networks can be fitted on the collected datasets. Specifically, for approximating the Q function of πκ\pi^{\kappa}, we can perform SARSA-style policy evaluation on dataset DκD^{\kappa}. Although this yields underestimated values due to exploration when collecting DκD^{\kappa}, it does not incur violation of the safety constraint, because Theorem 1 still holds even when the input Q functions have underestimated values. The safe exploration policies are then distributed and deployed to multiple advertising campaigns, running in parallel for a specific duration (e.g. one day) to collect an interaction dataset DkD^{k}. Subsequently, we employ an offline RL algorithm (e.g. IQL) to perform policy training on DkD^{k}, where the sampling probabilities are computed using Robust Trajectory Weighting. The resulting trained policy πk+1\pi^{k+1} becomes the input for the subsequent iteration, and the process continues iteratively.

Appendix D Details of Offline Experiments

Setup. For offline experiments, we construct a simulated advertising system mainly based on the simulated system in (Mou et al., 2022). There are 30 advertisers competing for advertising impressions. An episode corresponds to one day in simulation, which is divided into 96 time steps. The number of impressions in each time step is random, and follows a uniform distribution on [50, 300]. The budget of each advertiser follows a uniform distribution on [1500, 3000]. Before an episode starts, the number of impressions in every time step, as well as the value of each impression for each advertiser is initialized. The state of an advertiser is three-dimensional: [time,budget_consumed,budget_left][time,budget\_consumed,budget\_left]. The reward of an advertiser is the total value she wins in all ad auctions during one time step. Given current states and actions of all advertisers, the simulation returns next states and rewards. This is achieved by simulating ad auctions for all impressions in a single time step. For each impression, the system performs pre-ranking and ranking, and decides the winner of the auction. The winner gets the value of the impression, and pays for it according to the auction mechanism. We train a bidding policy for one advertiser while keeping other 29 advertisers’ policy fixed. Therefore, the policies of other bidders could be seen as part of the environment which is stationary. The trained auto-bidding policy could serve for bidders with different budgets since the information of total budget is contained in the state representation. The implementation of the pre-ranking and ranking(auction) part follows from that in (Mou et al., 2022), where more details could be found.

Parameter Settings. Datasets collected in each iteration consists of 100000 transition tuples. Strength of trajectory weighting α\alpha is set to 0.1. Safety threshold ϵ\epsilon is 0.05. PSN is implemented as factorised Gaussian noise (Fortunato et al., 2018) with σ\sigma searched from [0.01,0.03,0.05] and kept fixed during training. ASN is Gaussian noise with σ\sigma searched from [0.3,0.5,1]. In the IQL algorithm, the expectile parameter is set to 0.6 and β\beta is 1.25. Conservative factor is chosen as α=0.8\alpha=0.8 in CQL, and α=2.5\alpha=2.5 in TD3+BC. Implementation of SORL is consistent with the original paper (Mou et al., 2022) and code.