This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

An MPEC Estimator for the Sequential Search Model

Shinji Koiso, Suguru Otani koiso-shinji970@g.ecc.u-tokyo.ac.jp, Department of Economics, University of Tokyosuguru.otani@e.u-tokyo.ac.jp, Market Design Center, Department of Economics, University of Tokyo
Declarations of interest: none
(July 28, 2025)
Abstract

This paper proposes a constrained maximum likelihood estimator for sequential search models, using the MPEC (Mathematical Programming with Equilibrium Constraints) approach. This method enhances numerical accuracy while avoiding ad hoc components and errors related to equilibrium conditions. Monte Carlo simulations show that the estimator performs better in small samples, with lower bias and root-mean-squared error, though less effectively in large samples. Despite these mixed results, the MPEC approach remains valuable for identifying candidate parameters comparable to the benchmark, without relying on ad hoc look-up tables, as it generates the table through solved equilibrium constraints.
Keywords: Sequential search model, Search cost, Demand estimation, MPEC
JEL code: C50, L81, D83, M31

1 Introduction

The consumer search process, through which individuals gather information about choices, is essential to understanding decision-making behavior. This process has become increasingly observable to researchers through browsing data, which reveals the options considered before a final choice. The availability of such data has allowed for the estimation of structural models of consumer behavior, as noted by Ursu et al. (2023). The current benchmark model solves implicit functions related to reservation prices using a fixed-point approach, which is computationally demanding.

To address this, we propose an estimator based on the Mathematical Program with Equilibrium Constraints (MPEC) approach (Su and Judd 2012). MPEC is a constrained optimization problem subject to equilibrium conditions, that avoids iterative solutions to the fixed-point problem and removes approximation and estimation errors — a key issue in demand estimation (Dubé et al. 2012), dynamic programming (Su and Judd 2012), and misclassification models (Lu et al. 2014).

Monte Carlo simulations show that MPEC performs comparably, with lower bias and root-mean-squared error (RMSE) in small samples but higher bias and RMSE in larger samples, relative to the common estimation method using an ad hoc look-up table. The MPEC approach is particularly valuable for identifying parameters comparable to the benchmark, eliminating the need for the look-up table. MPEC can effectively construct this table by requiring the equilibrium constraints to be satisfied during estimation. We conclude that MPEC is a useful alternative for obtaining benchmark estimates before advancing to more complex models, especially when approximation and estimation errors from an ad hoc look-up table are unknown to researchers.

2 Weitzman’s sequential search model

2.1 Framework

We construct the sequential search model based on Weitzman (1979). A decision maker ii faces a set of boxes 𝒥={1,,J}\mathcal{J}=\{1,\cdots,J\} and box jj gives a potential reward uiju_{ij}\in\mathbb{R} independently drawn from a known distribution Fij(u)F_{ij}(u). Opening box jj takes cost cij++c_{ij}\in\mathbb{R}^{++}. An outside option is denoted as j=0j=0 with a known reward ui0u_{i0} is available at no cost. The decision maker opens boxes via sequential search steps and her goal is to maximize her expected reward net of total costs.

Suppose that the decision maker has opened a set SiS_{i} of boxes, which revealed a maximum reward value of ui=maxjSi0uiju_{i}^{*}=\max_{j\in S_{i}\cup 0}u_{ij}, and S¯i\bar{S}_{i} unopened boxes can still be opened. Her dynamic programming problem choosing whether to stop opening boxes and get payoff uiu_{i}^{*}, or to continue opening boxes is described by the following Bellman equation:

V(S¯i,ui)=max{ui,maxjS¯i{cij+Wj(S¯i,ui)}}\displaystyle V\left(\bar{S}_{i},u_{i}^{*}\right)=\max\left\{u_{i}^{*},\max_{j\in\bar{S}_{i}}\left\{-c_{ij}+W_{j}\left(\bar{S}_{i},u_{i}^{*}\right)\right\}\right\}

where Wj(S¯i,ui)W_{j}\left(\bar{S}_{i},u_{i}^{*}\right) is the expected value of continuing to open boxes and is defined as

Wj(S¯i,ui)=V(S¯i\j,ui)ui𝑑Fij(u)+uiV(S¯i\j,u)𝑑Fij(u).\displaystyle W_{j}\left(\bar{S}_{i},u_{i}^{*}\right)=V\left(\bar{S}_{i}\backslash j,u_{i}^{*}\right)\int_{-\infty}^{u_{i}^{*}}dF_{ij}(u)+\int_{u_{i}^{*}}^{\infty}V\left(\bar{S}_{i}\backslash j,u\right)dF_{ij}(u).

The reservation utility of a product zijz_{ij} is the utility level defined as

zij(uijzij)𝑑Fij(uij)=cij.\displaystyle\int_{z_{ij}}^{\infty}\left(u_{ij}-z_{ij}\right)dF_{ij}\left(u_{ij}\right)=c_{ij}.

A set of optimal decision rules, developed by Weitzman (1979), is used to characterize consumers’ optimal search and choice strategies. The rules are as follows:

  1. 1.

    Consumers know the true distribution(s) Fij(u)F_{ij}(u).

  2. 2.

    Search fully reveals the utility associated with product jj.

  3. 3.

    For each consumer i,uiji,u_{ij} is independently (across jj ) drawn from Fij(u)F_{ij}(u).

Then, the optimal search and choice decision rules are expressed as follows:

  1. 1.

    Selection Rule: The consumer searches in decreasing order of reservation utilities.

  2. 2.

    Stopping Rule: Search terminates when the maximum observed utility exceeds the reservation utility of any unsearched product.

  3. 3.

    Choice Rule: Once the consumer stops searching, she chooses the product with the highest observed utility among all searched options.

2.2 Parametrizations

Empirical economists often assume consumer ii’s utility defined as

uij\displaystyle u_{ij} =δij+εij=(ξij+μij)+εij,\displaystyle=\delta_{ij}+\varepsilon_{ij}=\left(\xi_{ij}+\mu_{ij}\right)+\varepsilon_{ij},
εij\displaystyle\quad\quad\varepsilon_{ij} i.i.dN(0,σμ),μiji.i.dN(0,σε)\displaystyle\sim_{i.i.d}N(0,\sigma_{\mu}),\quad\mu_{ij}\sim_{i.i.d}N(0,\sigma_{\varepsilon})

where δij\delta_{ij} is utility which is known by the consumer prior to search (“pre-search utility”) and εij\varepsilon_{ij} is utility that is only known by the consumer after search (“post-search taste shock”). We assume that the pre-search utility δij\delta_{ij} consists of a component ξij\xi_{ij} that can be observed by the researcher and a pre-search taste shock μij\mu_{ij} that cannot be observed by the researcher. According to Ursu et al. (2023), we need to further normalize their variance by setting σμ=σε=1\sigma_{\mu}=\sigma_{\varepsilon}=1.

Under the assumption of normally distributed post-search taste shocks, we can derive the following expression for the reservation utility:

zij=δij+m(cij)=ξij+μij+m(cij)\displaystyle z_{ij}=\delta_{ij}+m\left(c_{ij}\right)=\xi_{ij}+\mu_{ij}+m\left(c_{ij}\right)

where m(cij)m\left(c_{ij}\right) is the implicit function that solves the following equation (see Kim et al. (2010)):

cij=ϕ(m)+m×[Φ(m)1]\displaystyle c_{ij}=\phi(m)+m\times[\Phi(m)-1] (1)

with ϕ\phi and Φ\Phi denoting the standard normal pdf and cdf, respectively. Weitzman (1979) shows the existence and uniqueness of the solution of (1).

There are four primary methods to solve (1). The first method, proposed by Kim et al. (2010), involves pre-computing the mapping between mm and cc and storing it in a look-up table. The second method, suggested by Jiang et al. (2021), employs Newton’s method to compute reservation utilities by iteratively improving approximations to the root of the function:

q(m)=(1Φ(m))(ϕ(m)1Φ(m)m)c=0.q(m)=(1-\Phi(m))\left(\frac{\phi(m)}{1-\Phi(m)}-m\right)-c=0.

The third approach, introduced by Elberg et al. (2019), uses a contraction mapping defined as:

Γ(m)=c+ϕ(m)+m×Φ(m).\Gamma(m)=-c+\phi(m)+m\times\Phi(m).

The fourth method, proposed by Morozov (2023), directly estimates m(cij)m(c_{ij}).

Ursu et al. (2023) highlight limitations in each method: (1) the first method introduces errors due to linear interpolation for search cost values that do not align with grid points; (2) the second and third methods avoid interpolation errors but require iterative computation of mm and a convergence threshold, which can cause numerical errors if the threshold is too loose; and (3) the fourth method involves estimation errors for m(cij)m(c_{ij}). In practice, the second and third methods typically converge quickly, allowing for tight convergence thresholds that minimize numerical issues (Ursu et al. 2023). Similar challenges in demand estimation are addressed by the MPEC approach.

3 An MPEC estimator for the sequential search model

As a fifth method, we propose a straightforward estimator for the sequential search model, utilizing the Mathematical Programming with Equilibrium Constraints (MPEC) approach introduced by Su and Judd (2012). The MPEC estimator bypasses the need for iterative computations to find the fixed point by treating the equilibrium equations as constraints.

Let θ\theta represent the set of parameters. The MPEC estimator solves the following constrained optimization problem:

maxθ\displaystyle\max_{\theta} i𝒩logLi(θ,(zij)j𝒥,(uij)j𝒥)\displaystyle\sum_{i\in\mathcal{N}}\log L_{i}(\theta,(z_{ij})_{j\in\mathcal{J}},(u_{ij})_{j\in\mathcal{J}})
s.t.uij\displaystyle\text{s.t.}\quad u_{ij} =ξij+μij+εij\displaystyle=\xi_{ij}+\mu_{ij}+\varepsilon_{ij} (2)
zij\displaystyle z_{ij} =ξij+μij+m(cij)\displaystyle=\xi_{ij}+\mu_{ij}+m\left(c_{ij}\right)
cij\displaystyle c_{ij} =ϕ(m)+m×[Φ(m)1]\displaystyle=\phi(m)+m\times[\Phi(m)-1]

where individual likelihood Li(θ,(zij)j𝒥,(uij)j𝒥)L_{i}(\theta,(z_{ij})_{j\in\mathcal{J}},(u_{ij})_{j\in\mathcal{J}}) is derived as

Li(θ,(zij)j𝒥,(uij)j𝒥)\displaystyle L_{i}(\theta,(z_{ij})_{j\in\mathcal{J}},(u_{ij})_{j\in\mathcal{J}}) =Pr(zihmaxk𝒥{1,,h}zikhSiselection rule\displaystyle=\Pr(\underbrace{z_{ih}\geq\max_{k\in\mathcal{J}\setminus\{1,\cdots,h\}}z_{ik}\>\forall h\in S_{i}}_{\text{selection rule}}
zihmaxk=1h1uikhSimaxhSi{0}uihmaxlS¯izilstopping rule\displaystyle\cap\underbrace{z_{ih}\geq\max_{k=1}^{h-1}u_{ik}\>\forall h\in S_{i}\cap\max_{h\in S_{i}\cup\{0\}}u_{ih}\geq\max_{l\in\bar{S}_{i}}z_{il}}_{\text{stopping rule}}
uiyimaxhSi{0}uihchoice rule).\displaystyle\cap\underbrace{u_{iy_{i}}\geq\max_{h\in S_{i}\cup\{0\}}u_{ih}}_{\text{choice rule}}). (3)

A remarkable advantage of MPEC is that it does not need an ad hoc look-up table which is unknown to researchers and does not incorporate approximation and estimation error of equilibrium constraints (1), in addition to the main advantage of MPEC that it does not need to solve the fixed point problem iteratively. Detailed construction is provided in the Appendix.

4 Simulation

For comparison, we follow the parameter settings for ξij\xi_{ij} and cijc_{ij} in (2) as described in Appendix B of Ursu et al. (2023). We evaluate the MPEC approach against their kernel-smoothed frequency estimator (benchmark), which employs a look-up table—commonly used in empirical research. The same ad hoc table from Ursu et al. (2023), with a grid fineness of 0.001, is utilized. We generate 50 simulated datasets, each representing N{500,1000}N\in\{500,1000\} consumers who make sequential search and purchase decisions across four brands and an outside option (with the mean utility of the outside option normalized to zero). The utility function includes only brand intercepts, specified as (β1,β2,β3,β4)=(1.0,0.7,0.5,0.3)(\beta_{1},\beta_{2},\beta_{3},\beta_{4})=(1.0,0.7,0.5,0.3). The search cost logarithm is set at logc=3.0\log c=-3.0, and D=100D=100 draws are used for the error terms. All estimations start from an initial vector of zeros. The replication code, written in Julia for fair comparison, is available on the authors’ GitHub.

Table 1: MPEC vs ad hoc look-up (benchmark)
MPEC Look-up table
Bias RMSE Bias RMSE
β1\beta_{1} -0.179 0.250 -0.212 0.273
β2\beta_{2} -0.156 0.216 -0.168 0.237
β3\beta_{3} -0.044 0.071 -0.109 0.200
β4\beta_{4} -0.062 0.182 -0.060 0.192
logc\log c 0.214 0.250 0.248 0.304
Time 67.957 12.691
(a)
MPEC Look-up table
Bias RMSE Bias RMSE
β1\beta_{1} -0.272 0.361 -0.203 0.251
β2\beta_{2} -0.199 0.332 -0.114 0.174
β3\beta_{3} -0.173 0.255 -0.088 0.157
β4\beta_{4} -0.112 0.258 -0.036 0.159
logc\log c 0.161 0.255 0.265 0.290
Time 182.293 40.294
(b)

Note: The benchmark results closely replicate Column (4) in Table B1: Monte Carlo Simulation Results from Ursu et al. (2023). We calculate the average finish time for locally solved cases.

Table 1 presents the bias and RMSE of the estimated coefficients. Panel (a) shows that the MPEC estimator, though still biased, achieves a smaller bias and RMSE than the benchmark in small samples, aligning with the MPEC misclassification model (Lu et al. 2014). However, Panel (b) reveals worse performance in both bias and RMSE for larger samples, except for the search cost. MPEC also requires over four times the computational time and struggles with finding local optima.

Despite these seemingly discouraging results, we argue that the MPEC approach remains useful for identifying candidate parameters comparable to those from the benchmark method, which relies on an ad hoc look-up table. Furthermore, MPEC can construct this table dynamically by solving the equilibrium constraints during the estimation process.

5 Conclusion

The optimal sequential search model, based on Weitzman (1979), has been widely used in empirical research (Ursu et al. 2023). However, caution is needed regarding estimation and approximation accuracy, as the commonly used approach relies on an ad hoc look-up table.

To address these issues, we propose an MPEC estimator that bypasses the need for approximations and estimation of equilibrium constraints. Despite certain limitations, the MPEC approach proves useful for identifying parameters comparable to the benchmark while dynamically generating the look-up table during the estimation process.

Acknowledgments

This work was supported by JST ERATO Grant Number JPMJER2301, Japan.

References

  • (1)
  • Dubé et al. (2012) Dubé, Jean-Pierre, Jeremy T Fox, and Che-Lin Su, “Improving the numerical performance of static and dynamic aggregate discrete choice random coefficients demand estimation,” Econometrica, 2012, 80 (5), 2231–2267.
  • Elberg et al. (2019) Elberg, Andrés, Pedro M Gardete, Rosario Macera, and Carlos Noton, “Dynamic effects of price promotions: Field evidence, consumer search, and supply-side implications,” Quantitative Marketing and Economics, 2019, 17, 1–58.
  • Jiang et al. (2021) Jiang, Zhenling, Tat Chan, Hai Che, and Youwei Wang, “Consumer search and purchase: An empirical investigation of retargeting based on consumer online behaviors,” Marketing Science, 2021, 40 (2), 219–240.
  • Kim et al. (2010) Kim, Jun B, Paulo Albuquerque, and Bart J Bronnenberg, “Online demand under limited consumer search,” Marketing science, 2010, 29 (6), 1001–1023.
  • Lu et al. (2014) Lu, Ruichang, Yao Luo, and Ruli Xiao, “An MPEC estimator for misclassification models,” Economics Letters, 2014, 125 (2), 195–199.
  • Morozov (2023) Morozov, Ilya, “Measuring benefits from new products in markets with information frictions,” Management Science, 2023, 69 (11), 6988–7008.
  • Su and Judd (2012) Su, Che-Lin and Kenneth L Judd, “Constrained optimization approaches to estimation of structural models,” Econometrica, 2012, 80 (5), 2213–2230.
  • Ursu et al. (2023) Ursu, Raluca, Stephan Seiler, and Elisabeth Honka, “The Sequential Search Model: A Framework for Empirical Research,” Available at SSRN 4236738, 2023.
  • Weitzman (1979) Weitzman, Martin L, “Optimal Search for the Best Alternative,” Econometrica: Journal of the Econometric Society, 1979, pp. 641–654.

Appendix A Appendix (for online publication)

A.1 Crude estimator

We first introduce a crude estimator for likelihood expression (3) as the simplest approach. Define

vi,1h\displaystyle v_{i,1h} =zihmaxk𝒥{1,,h}zik\displaystyle=z_{ih}-\max_{k\in\mathcal{J}\setminus\{1,\cdots,h\}}z_{ik} (4)
vi,2h\displaystyle v_{i,2h} =zihmaxk=1h1uik\displaystyle=z_{ih}-\max_{k=1}^{h-1}u_{ik} (5)
vi,3\displaystyle v_{i,3} =maxhSi{0}uihmaxlS¯izil\displaystyle=\max_{h\in S_{i}\cup\{0\}}u_{ih}-\max_{l\in\bar{S}_{i}}z_{il} (6)
vi,4\displaystyle v_{i,4} =uiyimaxhSi{0}uih\displaystyle=u_{iy_{i}}-\max_{h\in S_{i}\cup\{0\}}u_{ih} (7)

Then, the estimation procedure is described as follows.

  • 1

    Take d=(1,,D)d=(1,\cdots,D) sets of draws of μij\mu_{ij} and εij\varepsilon_{ij} (each set of draws contains one draw of μij\mu_{ij} and one draw of εij\varepsilon_{ij}) for each consumer-product combination, i.e., D×J×ND\times J\times N sets of draws.

  • 2

    For a given guess of parameters θ\theta, compute uijdu^{d}_{ij} and zijdz^{d}_{ij} for each set of draws dd and each consumer-product combination.

  • 3

    Calculate the expressions in equations (4) to (7) for each set of draws dd and each consumer. Compute the likelihood contribution for each consumer and draw:

    Lid=[hSi𝟏{vi,1hd0}]×[hSi𝟏{vi,2hd0}]×𝟏{vi,3d0}×𝟏{vi,4d0}\displaystyle L_{i}^{d}=\left[\prod_{h\in S_{i}}\bm{1}\{v^{d}_{i,1h}\geq 0\}\right]\times\left[\prod_{h\in S_{i}}\bm{1}\{v^{d}_{i,2h}\geq 0\}\right]\times\bm{1}\{v^{d}_{i,3}\geq 0\}\times\bm{1}\{v^{d}_{i,4}\geq 0\}
  • 4

    Compute Li=1Dd=1DLidL_{i}=\frac{1}{D}\sum_{d=1}^{D}L_{i}^{d} for each consumer.

  • 5

    Compute logL=i=1Nlog(Li)\log L=\sum_{i=1}^{N}\log(L_{i}) and solve the constrained problem (2).

A.2 Kernel estimator

To improve upon the crude estimator, the kernel estimator applies a smooth kernel function to obtain the log-likelihood. Specifically, we use a multivariate scaled logistic cumulative distribution function as the kernel, resulting in the following consumer-specific likelihood contribution:

Lid=11+k=12hSiexp(ρkvi,khd)+k=12exp(ρkvi,kd),\displaystyle L_{i}^{d}=\frac{1}{1+\sum_{k=1}^{2}\sum_{h\in S_{i}}\exp(-\rho_{k}v^{d}_{i,kh})+\sum_{k=1}^{2}\exp(-\rho_{k}v^{d}_{i,k})},

where ρk\rho_{k} is a scaling parameter for each condition, to be determined by the researcher. The procedure to estimate vkv_{k} is the same as in the crude estimator. In our simulation, ρk\rho_{k} is set to 10 for N=500N=500 and 20 for N=1000N=1000 for both approaches. Further methods requiring fine-tuning are discussed in Ursu et al. (2023).