This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

A Dynamic Approach to Linear Statistical Calibration with an Application in Microwave Radiometry

Derick L. Rivers  and Edward L. Boone Corresponding author. Email: riversdl@vcu.edu Department of Statistical Sciences and Operations Research, Virginia Commonwealth University
Abstract

The problem of statistical calibration of a measuring instrument can be framed both in a statistical context as well as in an engineering context. In the first, the problem is dealt with by distinguishing between the “classical" approach and the “inverse" regression approach. Both of these models are static models and are used to estimate “exact" measurements from measurements that are affected by error. In the engineering context, the variables of interest are considered to be taken at the time at which you observe it. The Bayesian time series analysis method of Dynamic Linear Models (DLM) can be used to monitor the evolution of the measures, thus introducing an dynamic approach to statistical calibration. The research presented employs the use of Bayesian methodology to perform statistical calibration. The DLM’s framework is used to capture the time-varying parameters that maybe changing or drifting over time. Two separate DLM based models are presented in this paper. A simulation study is conducted where the two models are compared to some well known ’static’ calibration approaches in the literature from both the frequentist and Bayesian perspectives. The focus of the study is to understand how well the dynamic statistical calibration methods performs under various signal-to-noise ratios, rr. The posterior distributions of the estimated calibration points as well as the 95%95\% coverage intervals are compared by statistical summaries. These dynamic methods are applied to a microwave radiometry dataset.

1. Introduction

Calibrating measurement instruments is a important problem that engineers frequently need to address. There exist several statistical methods that address this problem that are based on a simple linear regression approach. In tradition simple linear regression the goal is to relate a known value of X to a uncertain value of Y using a linear relationship. In contrast, the statistical calibration problem seeks to utilize a simple linear regression model to relate a known value of Y to an uncertain value of X. This is why statistical calibration is sometimes called inverse regression due to its relationship to simple linear regression (Osborne 1991; Ott and Longnecker 2009). Recall in linear regression the model is given as follows:

𝐘=𝐗𝜷+ϵ{\bf Y}={\bf X}{\bm{\beta}}+{\bm{\epsilon}} (1)

where Y is a (n×1)(n\times 1) response vector, X is a (n×p)(n\times p) matrix of independent variables with p=k+1p=k+1 total model parameters, 𝜷{\bm{\beta}} is a (p×1)(p\times 1) vector of unknown fixed parameters and ϵ{\bm{\epsilon}} is a (n×1)(n\times 1) vector of uncorrelated error terms with zero mean (Myers 1990; Draper and Smith 1998; Montgomery et al. 2012). It is assumed that the value of the predictor variable X = x are nonrandom and observed with negligible error, while the nn error terms are random variables with mean zero and constant variance σ2\sigma^{2} (Myers 1990). Typically, in regression, of interest is the estimation of the parameter vector; 𝜷{\bm{\beta}}, and possibly the prediction of a future value 𝐘^i|new\hat{\bf Y}_{i|new} corresponding to a new 𝐗=xi|new{\bf X}=x_{i|new}^{\prime} value. The prediction problem is relatively straightforward, due to the fact that a future 𝐘i{\bf Y}_{i} value can be made directly by substituting xi|newx_{i|new}^{\prime} into (1) with E[ϵ]=0E[\epsilon]=0.
For the statistical calibration problem let y0y_{0} be the known observed value of the response and x0x_{0} be the corresponding regressor, x0x_{0} which is to be estimated. This problem is conducted in two stages: first measurement pairs (xi,yi)(x_{i},y_{i}) of data is observed and a simple linear regression line is fit by estimating 𝜷{\bm{\beta}}; secondly, mm observations of the response are observed, all corresponding to a single x0x_{0} (Özyurt and Erar 2003). Since y0y_{0} is fixed, inferences are different than those in a traditional regression (or prediction) problem (Osborne 1991; Eno 1999; Eno and Ye 2000).

1.   Classical Calibration Methods

Eisenhart (1939) offered the first solution to the calibration problem, and is commonly known as the classical``classical" estimator to the linear calibration problem. They assumed that the relationship between xx and yy was of a simple linear form:

E(Y|X=x)=β0+β1x.E(Y|X=x)=\beta_{0}+\beta_{1}x.

The estimated regression line for the first stage of the experiment is given by

Y^=β0^+β^1X,\hat{Y}=\hat{\beta_{0}}+\hat{\beta}_{1}X, (2)

where β0^\hat{\beta_{0}} and β1^\hat{\beta_{1}} are the least squares estimate of β0\beta_{0} and β1\beta_{1}, respectively. Using the data collected at the first stage of experimentation, Eisenhart (1939) inverts Equation (2) to estimate the unknown regressor value x0x_{0} for an observed response value y0y_{0}, by:

x^0,c=y0β^0β^1\hat{x}_{0,c}=\frac{y_{0}-\hat{\beta}_{0}}{\hat{\beta}_{1}} (3)

where x^0,c\hat{x}_{0,c} denotes the classical``classical" estimator for x0x_{0}. Since division by β^1\hat{\beta}_{1} is used there is an implicit assumption that |β^1|>0\lvert\hat{\beta}_{1}\rvert>0.
Assuming that |β^1|>0\lvert\hat{\beta}_{1}\rvert>0, Brown (1993) describes the following interval estimate corresponding to Eisenhart (1939):

y0β^0β^1(1+σ^2t2β^12Sxx)±σ^tβ^1(1+12n+(y0β^0)2+σ^2t22β^12Sxx),\frac{y_{0}-\hat{\beta}_{0}}{\hat{\beta}_{1}}\left(1+\frac{\hat{\sigma}^{2}t^{2}}{\hat{\beta}_{1}^{2}S_{xx}}\right)\pm\frac{\hat{\sigma}t}{\hat{\beta}_{1}}\left(1+\frac{1}{2n}+\frac{(y_{0}-\hat{\beta}_{0})^{2}+\hat{\sigma}^{2}t^{2}}{2\hat{\beta}_{1}^{2}S_{xx}}\right),

where

σ^=i=1n(yiβ^0β^1xi)2n2,\hat{\sigma}=\sqrt{\frac{\sum_{i=1}^{n}(y_{i}-\hat{\beta}_{0}-\hat{\beta}_{1}x_{i})^{2}}{n-2}},
Sxx=i=1n(xx¯)2,S_{xx}=\sum_{i=1}^{n}(x-\bar{x})^{2},

Krutchkoff (1967) proposed a competitive approach to Eisenhart’s (1939) classical linear calibration solution, which he called the inverse``inverse" regression calibration method and is written as:

Xi=ϕ+δYi+ϵi,X_{i}=\phi+\delta Y_{i}+\epsilon^{{}^{\prime}}_{i},

where ϕ\phi and δ\delta are the parameters in the linear relationship and ϵi\epsilon^{{}^{\prime}}_{i} are independent identically distributed measurement errors with a zero mean and finite variance. Here ϕ\phi and δ\delta are estimated via least squares. The unknown x0x_{0} can be estimated directly by substituting y0y_{0} into the fitted equation:

x^0,I=ϕ^+δ^y0.\hat{x}_{0,I}=\hat{\phi}+\hat{\delta}y_{0}. (4)

We let x^0,I\hat{x}_{0,I} denote the inverse``inverse" estimator of x0x_{0}. The 100(1α)%100(1-\alpha)\% confidence interval for E(x0,I|y0)E(x_{0,I}|y_{0}) can be written as

x0,I(y0)±tα/2σ^1n+(y0y¯)2Syyx_{0,I}(y_{0})\pm t_{\nicefrac{{\alpha}}{{2}}}\hat{\sigma}\sqrt{\frac{1}{n}+\frac{(y_{0}-\bar{y})^{2}}{S_{yy}}}

where

Syy=i=1n(yiy¯)2.S_{yy}=\sum_{i=1}^{n}(y_{i}-\bar{y})^{2}.

Krutchkoff (1967) used a simulation study, where he found that the mean squared error of estimation for x0x_{0} was uniformly less for this estimator versus the classical estimator. The inverse approach was later supported by Lwin and Maritz (1982). For criticisms of Krutchkoff’s (1967) approach such as bias see Osborne (1991).

2.   Bayesian Calibration Methods

The first noted Bayesian solution to the calibration problem was presented by Hoadley (1970). His work was motivated by the unanswered question in the Frequentist community of whether β1\beta_{1} is zero (or close to zero). Hoadley (1970) justified the use of the inverse``inverse" estimator (Krutchkoff, 1967) by considering the ususal FF-statistic to test the hypothesis that β1=0\beta_{1}=0 where F=β^12Sxx/σ^2F=\hat{\beta}_{1}^{2}S_{xx}/\hat{\sigma}^{2},

σ^2={i=1n(y1i(β^0+β^1xi))2+j=1m(y2jy¯2)2}(n+m3).\hat{\sigma}^{2}=\frac{\left\{\sum_{i=1}^{n}\left(y_{1i}-(\hat{\beta}_{0}+\hat{\beta}_{1}x_{i})\right)^{2}+\sum_{j=1}^{m}\left(y_{2j}-\bar{y}_{2}\right)^{2}\right\}}{(n+m-3)}.

The assumption made by Hoadley (1970) reflects that x0x_{0} is random and a priori independent of π(β0,β1,σ2)\pi(\beta_{0},\beta_{1},\sigma^{2}), so that the joint prior distribution of π(β0,β1,σ2,x0)π(β0,β1,σ2)π(x0)\pi(\beta_{0},\beta_{1},\sigma^{2},x_{0})\propto\pi(\beta_{0},\beta_{1},\sigma^{2})\pi(x_{0}). Hoadley (1970) first assumed that (β0,β1,σ2)(\beta_{0},\beta_{1},\sigma^{2}) had a uniform distribution,

π(β0,β1,σ2)σ2,\pi(\beta_{0},\beta_{1},\sigma^{2})\propto\sigma^{-2},

but the prior distribution for x0x_{0} was not given.
Hoadley (1970) shows for m=1m=1 (one observation at the prediction stage), that if x0x_{0} has a prior density from a Student t distribution with n3n-3 degrees of freedom, a mean of 0, and a scale parameter

σ=n+1n3,\sigma=\frac{n+1}{n-3},

the posterior distribution is

π(x0|𝐃𝐚𝐭𝐚)=tn2(x^0,I,[n+1+(x^0,I)2/RF+n2]),\pi(x_{0}|{\bf Data})=t_{n-2}\left(\hat{x}_{0,I},\left[\frac{n+1+(\hat{x}_{0,I})^{2}/R}{F+n-2}\right]\right), (5)

where x^0,I\hat{x}_{0,I} is the inverse estimator given by (4),R=FF+n2,R=\frac{F}{F+n-2} and F=β^12Sxx/σ^2F=\hat{\beta}_{1}^{2}S_{xx}/\hat{\sigma}^{2}.
Hunter and Lamboy (1981) also considered the calibration problem from a Bayesian point of view and is similar to that of Hoadley (1970) because both assume the prior distribution to be

π(β0,β1,σ2,η)σ2\pi(\beta_{0},\beta_{1},\sigma^{2},\eta)\propto\sigma^{-2}

where η=β0+β1x0\eta=\beta_{0}+\beta_{1}x_{0} which is the predicted y0y_{0}. The primary difference between their approach and the approach of Hoadley (1970) is that a priori they assume that η\eta and (β0,β1,σ2)(\beta_{0},\beta_{1},\sigma^{2}) are independent while Hoadley (1970) assumed a priori that x0x_{0} and (β0,β1,σ2)(\beta_{0},\beta_{1},\sigma^{2}) are independent.
Hunter and Lamboy (1981) uses an approximation to the posterior distribution of the unknown regressor x0x_{0} by

π(x0|𝐃𝐚𝐭𝐚)=N(x^0,c,(s11+s33)s22s122s22β^12),\pi(x_{0}|{\bf Data})=N\left(\hat{x}_{0,c},\frac{(s_{11}+s_{33})s_{22}-s_{12}^{2}}{s_{22}\hat{\beta}_{1}^{2}}\right), (6)

where

𝐒={si,j}=[s11s120s12s22000s33]=[(𝐗𝐗)1σ^2𝟎𝟎σ^2/m],{\bf S}=\{s_{i,j}\}=\left[\begin{array}[]{ccc}s_{11}&s_{12}&0\\ s_{12}&s_{22}&0\\ 0&0&s_{33}\end{array}\right]=\left[\begin{array}[]{cc}({\bf X}^{{}^{\prime}}{\bf X})^{-1}\hat{\sigma}^{2}&{\bf 0}\\ {\bf 0}&\nicefrac{{\hat{\sigma}^{2}}}{{m}}\end{array}\right],

with x^0,c\hat{x}_{0,c} being the classical estimator given in Equation (3), si,js_{i,j} denote the element of the ithi^{th} row and jthj^{th} column from variance-covariance matrix of the joint posterior density of (β0\beta_{0}, β1\beta_{1}, η\eta).
The remainder of this paper is organized as follows. Section 2 presents the development of the dynamic approaches to the statistical calibration problem. In Section 3 the results from the simulation study where the dynamics methods are evaluated along with the static approaches are presented. In Section 4 the proposed methods are applied to microwave radiometer data. In Section 5 future work and other considerations are given.

2. Dynamic Calibration Approach

Traditional calibration methods assume the regression relationship is “static” in time. In many cases this is false, for example in microwave radiometry the static nature of the relationship is known to change across time. A dynamic approach can be created by letting the regression coefficients vary through time,

yt=β0t+β1txt+ϵt,y_{t}=\beta_{0t}+\beta_{1t}x_{t}+\epsilon_{t},

where ϵtiidN[0,σt2]\epsilon_{t}\stackrel{{\scriptstyle iid}}{{\sim}}N[0,\sigma^{2}_{t}] and is known as the observationalobservational error.
The model may have different defining parameters at different times. One approach is to model β0t\beta_{0t} and β1t\beta_{1t} by using random walk type evolutions for the defining parameters, such as:

β0t\displaystyle\beta_{0t} =\displaystyle= β0(t1)+ωβ0t,\displaystyle\beta_{0(t-1)}+\omega_{\beta_{0t}},
β1t\displaystyle\beta_{1t} =\displaystyle= β1(t1)+ωβ1t,\displaystyle\beta_{1(t-1)}+\omega_{\beta_{1t}},

where ωβ0t\omega_{\beta_{0t}} and ωβ1t\omega_{\beta_{1t}} are independent zero-mean error terms with finite variances. At any time tt the calibration problem is given by:

y0t=β0t+β1tx0t+ϵt,t=1,2,,T.y_{0t}=\beta_{0t}+\beta_{1t}x_{0t}+\epsilon_{t},\hskip 30.0ptt=1,2,\dots,T.

Bayesian Dynamic Linear Models (DLMs) approach of West et al. (1985); West and Harrison (1997) can be employed to achieve this goal. Recall the DLM framework is:

Observation equation:\displaystyle\mbox{Observation equation}:\hskip 28.45274pt 𝐘t=𝐗t𝐗t𝜽t+ϵt,\displaystyle{\bf Y}_{t}={\bf X}_{t}{\bf X}_{t}{\bm{\theta}_{t}}+{\bm{\epsilon}}_{t},\hskip 25.6073pt ϵtNr[𝟎,𝐄]\displaystyle{\bm{\epsilon}}_{t}\sim N_{r}[{\bf 0,E}]
System equation:\displaystyle\mbox{System equation}:\hskip 28.45274pt 𝜽t=𝐆t𝜽t1+𝝎t,\displaystyle{\bm{\theta}_{t}}={\bf G}_{t}{\bm{\theta}_{t-1}}+{\bm{\omega}_{t}}, 𝝎tNd[𝟎,𝐖]\displaystyle{\bm{\omega}}_{t}\sim N_{d}[{\bf 0,W}]
Initial information:\displaystyle\mbox{Initial information}:\hskip 28.45274pt (𝜽0|D0)Nd[𝐦𝟎,𝐂𝟎],\displaystyle({\bm{\theta}_{0}}|D_{0})\sim N_{d}[{\bf m_{0},C_{0}}],

for some prior mean 𝐦𝟎\bf m_{0} and variance 𝐂𝟎\bf C_{0} with the vector of error terms, ϵt{\bm{\epsilon}}_{t} and 𝝎t{\bm{\omega}}_{t} independent across time and at any time.
To update the model through time West and Harrison (1997) give the following method:

  1. (a)

    Posterior distribution at t1t-1: For some mean 𝐦t1{\bf m}_{t-1} and variance 𝐂t1{\bf C}_{t-1},
              (𝜽t1|Dt1)Nd[𝐦t1,𝐂t1]({\bm{\theta}_{t-1}}|D_{t-1})\sim N_{d}[{\bf m}_{t-1},{\bf C}_{t-1}].

  2. (b)

    Prior distribution at time tt: (𝜽t|Dt1)Nd[𝐚t,𝐑t]({\bm{\theta}_{t}}|D_{t-1})\sim N_{d}[{\bf a}_{t},{\bf R}_{t}], where
              𝐚t=𝐆t𝐦t1{\bf a}_{t}={\bf G}_{t}{\bf m}_{t-1}    and    𝐑t=𝐆t𝐂t1𝐆t+𝐖{\bf R}_{t}={\bf G}_{t}{\bf C}_{t-1}{\bf G}^{{}^{\prime}}_{t}+{\bf W}.

  3. (c)

    One-step forecast: (𝐘t|Dt1)Nr[𝐟t,𝐐t]({\bf Y}_{t}|D_{t-1})\sim N_{r}[{\bf f}_{t},{\bf Q}_{t}], where
              𝐟t=𝐗t𝐚t{\bf f}_{t}={\bf X}_{t}{\bf a}_{t}    and    𝐐t=𝐗t𝐑t𝐗t+𝐄{\bf Q}_{t}={\bf X}_{t}{\bf R}_{t}{\bf X}_{t}^{{}^{\prime}}+{\bf E}.

  4. (d)

    Posterior distribution at time tt: (𝜽t|Dt)Nd[𝐦t,𝐂t]({\bm{\theta}_{t}}|D_{t})\sim N_{d}[{\bf m}_{t},{\bf C}_{t}], with
              𝐦t=𝐚t+𝐀t𝐞t{\bf m}_{t}={\bf a}_{t}+{\bf A}_{t}{\bf e}_{t}    and    𝐂t=𝐑t𝐀t𝐐t𝐀t,{\bf C}_{t}={\bf R}_{t}-{\bf A}^{{}^{\prime}}_{t}{\bf Q}_{t}{\bf A}_{t},
    where
              𝐀t=𝐐t1𝐗t𝐑t{\bf A}_{t}={\bf Q}^{-1}_{t}{\bf X}_{t}{\bf R}_{t}    and    𝐞t=𝐘t𝐟t{\bf e}_{t}={\bf Y}_{t}-{\bf f}_{t}.

The DLM framework is used to establish the evolving relationship between the fixed design matrix 𝐗t{\bf X}_{t} and 𝐘t{\bf Y}_{t} by estimating 𝜽t{\bm{\theta}}_{t}, which is a (d×n)(d\times n) matrix of time-varying regression coefficients β0t\beta_{0t} and β1t\beta_{1t}. For our calibration situation 𝐘t{\bf Y}_{t} is a (r×n)(r\times n) matrix of responses and 𝐆t{\bf G}_{t} is a known (d×dd\times d) system matrix. The error ϵt{\bm{\epsilon}}_{t} and 𝝎t{\bm{\omega}}_{t} are independent normally distributed random (r×n)(r\times n) matrices with zero mean and constant variance-covariance matrices E and W. For simplification 𝐆t{\bf G}_{t} is set equal to 𝐈(d×d){\bf I}_{(d\times d)}, 𝐄{\bf E} is set equal to σE2𝐈(r×r)\sigma^{2}_{E}{\bf I}_{(r\times r)} and 𝐖{\bf W} is σW2[𝐗t𝐗t]1\sigma^{2}_{W}\left[{\bf X}_{t}^{{}^{\prime}}{\bf X}_{t}\right]^{-1}. The past information is contained in the set D0D_{0}.
We specify a prior in the first stage of calibration for the unknown variances and derive an algorithm to draw from the posterior distribution of the unknown parameters,

π(𝜽t,σE2,σW2|𝐘t)π(𝜽t|σE2,σW2,𝐘t)π(σE2,σW2|𝐘t).\pi({\bm{\theta}}_{t},\sigma^{2}_{E},\sigma^{2}_{W}|{\bf Y}_{t})\propto\pi({\bm{\theta}}_{t}|\sigma^{2}_{E},\sigma^{2}_{W},{\bf Y}_{t})\pi(\sigma^{2}_{E},\sigma^{2}_{W}|{\bf Y}_{t}).

The second stage of the calibration experiment consists of using the joint posterior distribution π(𝜽t,σE2,σW2|𝐘t)\pi({\bm{\theta}}_{t},\sigma^{2}_{E},\sigma^{2}_{W}|{\bf Y}_{t}) to derive x0t|𝜽t,σE2,σW2x_{0t}|{\bm{\theta}}_{t},\sigma^{2}_{E},\sigma^{2}_{W} for each draw of π(σE2,σW2|𝐘t)\pi(\sigma^{2}_{E},\sigma^{2}_{W}|{\bf Y}_{t}). The estimator for the parameter of interest, x0tx_{0t}, is defined in a manner akin to Eisenhart (1939); Hunter and Lamboy (1981); Eno (1999), where

x0t=y0tβ0tβ1t.x_{0t}=\frac{y_{0t}-\beta_{0t}}{\beta_{1t}}. (7)

In the final stage of the calibration experiment, the posterior distribution summary statistics are gathered at each time point tt. The posterior median and credible intervals are taken for each tt across the draws of x0t|𝜽t,σE2,σW2x_{0t}|{\bm{\theta}}_{t},\sigma^{2}_{E},\sigma^{2}_{W}. The result of the dynamic calibration experiment is a time series of calibration distributions across time. We will be able to observe the distributional changes of the system with respect to the calibration reference.
The proposed calibration estimator is developed by first considering the joint posterior distribution π(σE2,σW2|𝐘t)\pi(\sigma^{2}_{E},\sigma^{2}_{W}|{\bf Y}_{t}). We let 𝚪{\bm{\Gamma}} denote the vector of unknown DLM dispersion parameters where 𝚪=(σE2,σW2){\bm{\Gamma}}^{\prime}=(\sigma^{2}_{E},\sigma^{2}_{W}). The prior information for the dispersion parameters is described by a prior density π(𝚪)\pi({\bm{\Gamma}}) which summarizes what is known about the variance parameters before any data are observed. Using the Bayesian inferential approach, the prior information about the parameters must be combined with information contained in the data. The information provided by the data is captured by the likelihood functions, f𝐘(𝐘t|𝜽t,σE2,σW2)f_{{\bf Y}}({\bf Y}_{t}|{\bm{\theta}}_{t},\sigma^{2}_{E},\sigma^{2}_{W}) and f𝜽(𝜽t|𝜽t1,σW2)f_{\bm{\theta}}({\bm{\theta}}_{t}|{\bm{\theta}}_{t-1},\sigma^{2}_{W}) for the observation equation and the system equation, respectively. The combined information is described by the posterior density using the Bayes theorem (Bernardo and Smith 1994) as

π(𝚪|𝐘t)f𝐘(𝐘t|𝜽t,σE2,σW2)f𝜽(𝜽t|𝜽t1,σW2)π(𝚪).\pi({\bm{\Gamma}}|{\bf Y}_{t})\propto f_{{\bf Y}}({\bf Y}_{t}|{\bm{\theta}}_{t},\sigma^{2}_{E},\sigma^{2}_{W})\cdot f_{\bm{\theta}}({\bm{\theta}}_{t}|{\bm{\theta}}_{t-1},\sigma^{2}_{W})\cdot\pi({\bm{\Gamma}}).

For our calibration problem it is believe that σE2>σW2\sigma^{2}_{E}>\sigma^{2}_{W}. To deal with the variance relationship we specify the following prior distributions:

σE2\displaystyle\sigma^{2}_{E} \displaystyle\sim Uniform(0,1)\displaystyle Uniform(0,1) (8)
σW2|σE2\displaystyle\sigma^{2}_{W}|\sigma^{2}_{E} \displaystyle\sim Uniform(0,σE2).\displaystyle Uniform(0,\sigma^{2}_{E}). (9)

Prior distributions (8) and (9) ensures the system variance to be less than the observation variance. Since these are proper prior distributions the resulting posterior distribution will also be proper.
In the first stage of calibration, the joint distribution of the observations, states, and unknown parameters is as follows:

π(𝐘1:T,𝜽0:T,σE2,σW2)\displaystyle\pi({\bf Y}_{1:T},{\bm{\theta}}_{0:T},\sigma^{2}_{E},\sigma^{2}_{W}) =\displaystyle= f𝐘(𝐘1:T|𝜽0:T,σE2,σW2)f𝜽(𝜽0:T|σW2)π(𝚪)\displaystyle f_{{\bf Y}}({\bf Y}_{1:T}|{\bm{\theta}}_{0:T},\sigma^{2}_{E},\sigma^{2}_{W})\cdot f_{\bm{\theta}}({\bm{\theta}}_{0:T}|\sigma^{2}_{W})\cdot\pi({\bm{\Gamma}})
=\displaystyle= t=1Tf𝐘(𝐘t|𝜽t,σE2)t=1Tf𝜽(𝜽t|𝜽t1,σW2)\displaystyle\prod_{t=1}^{T}f_{{\bf Y}}({\bf Y}_{t}|{\bm{\theta}}_{t},\sigma^{2}_{E})\cdot\prod_{t=1}^{T}f_{\bm{\theta}}({\bm{\theta}}_{t}|{\bm{\theta}}_{t-1},\sigma^{2}_{W})
π(𝜽0)π(σE2)π(σW2|σE2).\displaystyle\hskip 10.0pt\cdot\pi({\bm{\theta}}_{0})\cdot\pi(\sigma^{2}_{E})\cdot\pi(\sigma^{2}_{W}|\sigma^{2}_{E}).

where the likelihood for the observation equation is

f𝐘(𝐘t|𝜽t,σE2)σETexp{12σE2t=1T(𝐘t𝐗t𝜽t)2}f_{{\bf Y}}({\bf Y}_{t}|{\bm{\theta}}_{t},\sigma^{2}_{E})\propto\sigma^{-T}_{E}\mbox{exp}\left\{-\frac{1}{2\sigma^{2}_{E}}\sum_{t=1}^{T}({\bf Y}_{t}-{\bf X}_{t}{\bm{\theta}}_{t})^{2}\right\}

and the likelihood for the system equation is

f𝜽(𝜽t|𝜽t1,σW2)σWTexp{12σW2t=1T(𝜽t𝜽t1)2}.f_{{\bm{\theta}}}({\bm{\theta}}_{t}|{\bm{\theta}}_{t-1},\sigma^{2}_{W})\propto\sigma^{-T}_{W}\mbox{exp}\left\{-\frac{1}{2\sigma^{2}_{W}}\sum_{t=1}^{T}({\bm{\theta}}_{t}-{\bm{\theta}}_{t-1})^{2}\right\}.

Given the joint distribution above, the posterior distribution is

π(𝐱0t|𝜽t,𝚪,𝐘t)\pi({\bf x}_{0t}|{\bm{\theta}}_{t},{\bm{\Gamma}},{\bf Y}_{t}) (10)

where

𝐱0t=𝐲𝟎𝐭𝜽t{\bf x}_{0t}=\frac{\bf{y}^{*}_{0t}}{{\bm{\theta}}_{t}} (11)

and 𝐲0t=𝐲0ty¯t{\bf y}^{*}_{0t}={\bf y}_{0t}-\bar{y}_{t} (i.e. y¯t\bar{y}_{t} is the cumulative mean of the observations up to time tt) and 𝜽t=𝜷^1t{\bm{\theta}}_{t}=\hat{\bm{\beta}}_{1t}. Samples from the posterior distribution in Equation (10) are drawn by implementing the Sampling Importance Resampling (Albert 2007; Givens and Hoeting 2005) approach.
The development of the estimator in Equation (11) is deterministic in approach. We present a fully Bayesian approach to dynamic calibration that incorporates the uncertainty in estimation. The second dynamic calibration model is derived by Bayes’ theorem

π(𝐱𝟎𝐭|𝐘t)π(𝐱𝟎𝐭)f(𝐘t|𝐱𝟎𝐭),\pi({\bf x_{0t}}|{\bf Y}_{t})\propto\pi({\bf x_{0t}})f({\bf Y}_{t}|{\bf x_{0t}}),

where π(𝐱𝟎𝐭|𝐘t)\pi({\bf x_{0t}}|{\bf Y}_{t}) is the posterior distribution for 𝐱0t{\bf x}_{0t}. The prior belief for the calibration values is denoted as π(𝐱𝟎𝐭)\pi({\bf x_{0t}}) with the f(𝐘t|𝐱𝟎𝐭)f({\bf Y}_{t}|{\bf x_{0t}}) denoting the likelihood function.
The objective of any Bayesian approach is to obtain the posterior distribution from which inferences can be made. Here the desired posterior is

π(𝐱𝟎𝐭|𝐘t)\pi({\bf x_{0t}}|{\bf Y}_{t}) (12)

which must be dynamic through time. We determine the posterior distribution (12) in a similiar manner as described above in Equations (10) and (11). In the first stage of the calibration experiment the data is scaled and centered, therefore setting the yy-intercept equal to zero and the reference measurements centered at zero. Centering of the data is used to reduce the parameter space. The posterior distribution can be thought of as:

π(𝐳𝟎𝐭|𝐘t)π(𝐳𝟎𝐭)f(𝐘t|𝐳𝟎𝐭),\pi({\bf z_{0t}}|{\bf Y}^{*}_{t})\propto\pi({\bf z_{0t}})f({\bf Y}^{*}_{t}|{\bf z_{0t}}), (13)

with 𝐳𝟎𝐭{\bf z_{0t}} representing the transformed calibrated value at time tt and 𝐘t=𝐘tY¯t{\bf Y}^{*}_{t}={\bf Y}_{t}-\bar{Y}_{t}, where Y¯t\bar{Y}_{t} is the cumulative mean of the observations. Given this information aprioria~priori we define the prior distribution

π(𝐳𝟎𝐭)=N(0,1).\pi({\bf z_{0t}})=N(0,1).

The posterior density in Equation (13) is defined as

π(𝐳𝟎𝐭|𝐘t)exp{12[σYt2t=1T(𝝃t𝐳0t)2+𝐳0t2]}\pi({\bf z_{0t}}|{\bf Y}^{*}_{t})\propto\mbox{exp}\left\{-\frac{1}{2}\left[\sigma^{-2}_{Y_{t}}\sum_{t=1}^{T}({\bm{\xi}}_{t}-{\bf z}_{0t})^{2}+{\bf z}^{2}_{0t}\right]\right\} (14)

where 𝝃t=𝐘0t/𝜽t{\bm{\xi}}_{t}=\nicefrac{{{\bf Y}^{*}_{0t}}}{{\bm{\theta}_{t}}}. Applying Bayes theorem and completing the square, the posterior distribution is

π(𝐳𝟎𝐭|𝐘t)N(μz0t,σz0t2),\pi({\bf z_{0t}}|{\bf Y}^{*}_{t})\sim N(\mu_{z_{0t}},\sigma^{2}_{z_{0t}}), (15)

with

μz0t\displaystyle\mu_{z_{0t}} =\displaystyle= 𝝃t1+σYt2,\displaystyle\frac{{\bm{\xi}}_{t}}{1+\sigma^{2}_{Y_{t}}},
σz0t2\displaystyle\sigma^{2}_{z_{0t}} =\displaystyle= 11+σYt2\displaystyle\frac{1}{1+\sigma^{2}_{Y_{t}}}

and

σYt2=tr(𝐐t).\sigma^{2}_{Y_{t}}=\mbox{tr}({\bf Q}_{t}).

where tr( . ) denotes trace of the one-step forecast variance-covariance matrix. We derive the posterior in Equation (12) by drawing from Equation (15) and transforming the data back to the original scale as so:

𝐱0t=X¯+𝐳0tσX,{\bf x}_{0t}=\bar{X}+{\bf z}_{0t}\sigma_{X}, (16)

where X¯\bar{X} is the mean of the reference measurements vector and σX\sigma_{X} is the standard deviation of the reference measurements vector.
The dynamic calibration algorithm is developed for both of the approaches using R (R Development Core Team, 2013) and is conducted as below.

Algorithm: Dynamic Calibration 1. Generate MM proposal samples for (σE2,σW2)(\sigma^{2}_{E},\sigma^{2}_{W}) from π(σE2)\pi(\sigma^{2}_{E}) and π(σW2|σE2)\pi(\sigma^{2}_{W}|\sigma^{2}_{E}); 2. Calibration data are fit using the DLM framework for each of the MM proposal samples (σE2(m),σW2(m))(\sigma^{2(m)}_{E},\sigma^{2(m)}_{W}), with the prior moments for (𝜽0|D0)({\bm{\theta}_{0}}|D_{0}) as 𝐦𝟎=𝟏𝐝\bf m_{0}=1_{d} and 𝐂𝟎=𝟏𝟎𝟎𝐈(𝐝×𝐝)\bf C_{0}=100I_{(d\times d)}, where 𝟏𝐝{\bf 1_{d}} is a dd-dimensional vector of ones.
  1. a.

    Data are scaled and shifted such that i=1rxi=0\sum^{r}_{i=1}x_{i}=0, 1ni=1rxi2=1\frac{1}{n}\sum^{r}_{i=1}x^{2}_{i}=1 and yy-intercept =0=0, where yt=yty¯ty^{*}_{t}=y_{t}-\bar{y}_{t} for all tt (i.e. y¯t\bar{y}_{t} is the cumulative mean up to time tt);

  2. b.

    Estimate 𝜽t(m)|σE2(m),σW2(m){\bm{\theta}}^{(m)}_{t}|\sigma^{2(m)}_{E},\sigma^{2(m)}_{W} for the mthm^{th} proposal sample is calculated for allt\mbox{for all}~t;

  3. c.

    Estimate x0t(m)|𝜽t(m),σE2(m),σW2(m)x^{(m)}_{0t}|{\bm{\theta}}^{(m)}_{t},\sigma^{2(m)}_{E},\sigma^{2(m)}_{W} for the mthm^{th} proposal sample is calculated for allt\mbox{for all}~t, using either Equation (11) or drawing from Equation (15);

  4. d.

    Calculate log-likelihood density weights, log[f(𝚪(m))]log[f({\bm{\Gamma}}^{(m)})], for each (σE2(m),σW2(m))(\sigma^{2(m)}_{E},\sigma^{2(m)}_{W}) pair

  • 3.

    Sampling Importance Resampling (SIR) is used to simulate samples of x0t|𝜽t,σE2,σW2x_{0t}|{\bm{\theta}}_{t},\sigma^{2}_{E},\sigma^{2}_{W} by accepting a subset of N=1,000N=1,000 from the proposal density to be distributed according to the posterior density π(𝚪|𝐘t)\pi({\bm{\Gamma}}|{\bf Y}_{t}) with candidate density π(𝚪)\pi({\bm{\Gamma}}).

    1. a.

      Calculate the standardized importance weights, w(𝚪(1)),,w(𝚪(M))w({{\bm{\Gamma}}^{(1)}}),\dots,w({{\bm{\Gamma}}^{(M)}}) , where w(𝚪(m))=log[f(𝚪(m))]log[g(𝚪(m))]w({{\bm{\Gamma}^{(m)}}})=log[f({\bm{\Gamma}}^{(m)})]-log[g({\bm{\Gamma}}^{(m)})] for the mthm^{th} proposal sample;

    2. b.

      Sample NN calibrated time series from the MM proposal values with replacement given probabilities p(𝚪(m))p({{\bm{\Gamma}}^{(m)}}) where

      p(𝚪(m))=ew(𝚪(m))j=1Mew(𝚪(j)).p({{\bm{\Gamma}}^{(m)}})=\frac{e^{w({{\bm{\Gamma}^{(m)}}})}}{\sum_{j=1}^{M}e^{w({{\bm{\Gamma}^{(j)}}})}}.
  • 4.

    Rescale calibrated time series to original scale by Equation (16) and take summary statistics (i.e. medians and credible sets) across each time tt .

  • 3. Simulation Study

    A simulation study, mirroring the microwave radiometer example in Section 4, considers the performance of the proposed dynamic calibration approaches to the static approaches discussed in Section 1. For notation, the calibration methods are labelled as follows:

    1. 1.

      MD1M_{D1} is the first deterministic dynamic calibration model given in Equation (11);

    2. 2.

      MD2M_{D2} is the Bayesian dynamic calibration model given by Equation (15);

    3. 3.

      MF1M_{F1} is the classical``classical" approach of Eisenhart (1939) defined in Equation (3);

    4. 4.

      MF2M_{F2} is the inverse``inverse" approach of Krutchkoff (1967) defined in Equation (4);

    5. 5.

      MB1M_{B1} is Hoadley (1970) Bayesian approach as defined in Equation (5);

    6. 6.

      MB2M_{B2} is Hunter & Lamboy (1981) Bayesian approach as defined in Equation (6).

    Note that static methods MF1M_{F1}, MF2M_{F2}, MB1M_{B1}, and MB2M_{B2} require that model fitting and the calibration take place after all the data has been collected. This is in contrast to the dynamic methods that both fit the model and generate calibrated values each point through time and hence provide a near real time calibration. In order to assess the performance of the calibration methods 100 datasets were randomly generated according to

    𝐘t=𝐗𝜽t+ϵt,{\bf Y}_{t}={\bf X}{\bm{\theta}}_{t}+{\bm{\epsilon}}_{t}, (17)

    where 𝐗{\bf X} is a known fixed design matrix of reference values. The number of references measurements used in the study was two and five. The reference values at the first stage of the simulation study were equally spaced, covering the interval [20,100][20,100]. For the two reference case, the fixed design matrix is

    𝐗=[1201100]{\bf X}=\left[\begin{array}[]{cc}1&20\\ 1&100\end{array}\right]

    and for the five reference case the design matrix is

    𝐗=[1201401601801100].{\bf X}=\left[\begin{array}[]{cc}1&20\\ 1&40\\ 1&60\\ 1&80\\ 1&100\end{array}\right].

    The vector of regression parameters, 𝜽t{\bm{\theta}}_{t}, are randomly drawn from a multivariate normal distribution with mean vector [12.74340.02655][12.7434~0.02655]^{{}^{\prime}} and variance-covariance matrix, 𝚺=σW2[𝐗𝐗]1{\bm{\Sigma}}=\sigma^{2}_{W}\left[{\bf X}^{{}^{\prime}}{\bf X}\right]^{-1} for t=1,,Tt=1,\dots,T, where T=1000T=1000. For each tt, the random multivariate error vector is

    ϵtNr[𝟎,σE2𝐈]{\bm{\epsilon}}_{t}\sim N_{r}[{\bf 0},\sigma^{2}_{E}{\bf I}]

    where the errors are mutually independent. The relationship of the values for σE2\sigma^{2}_{E} and σW2\sigma^{2}_{W} will be explained later.
    The dynamic and static calibration methods are evaluated for three distinct system fluctuations, gtg_{t}, on the regression slope calculated in the first stage of calibration. The value gtg_{t} is added to β1t\beta_{1t}, therefore making Equation (17)

    yjt=β0t+(β1t+gt)xj+ϵt,t=1,,Ty_{jt}=\beta_{0t}+(\beta_{1t}+g_{t})x_{j}+\epsilon_{t},\quad t=1,\dots,T\\

    for the jthj^{th} calibration references. The three scenarios for the fluctuations gtg_{t} are as follows:

    1. 1.

      a constant zero (gt=0g_{t}=0) for all tt, representing a stable system;

    2. 2.

      a stable system with abrupt shifts (gt=ai)g_{t}=a_{i}) in system, with i=1nai=0\sum_{i=1}^{n}a_{i}=0; and

    3. 3.

      a constant sinusoidal fluctuation (gt=0.1sin(0.025t)g_{t}=0.1\mbox{sin}(0.025t)) for all tt.

    Figure 1) explains the relationship of gtg_{t} across time.

    Refer to caption
    Refer to caption
    Refer to caption
    Figure 1: Three distinct gain fluctuations: (a) gt=0g_{t}=0; (b) gt=aig_{t}=a_{i} with i=1nai=0\sum_{i=1}^{n}a_{i}=0; (c) gt=0.1sin(0.025t)g_{t}=0.1\mbox{sin}(0.025t)

    The magnitude and relationship of the variance pair (σE2,σW2)(\sigma^{2}_{E},\sigma^{2}_{W}) influence the DLM and hence to study this influence we set the variances to reflect various signal-to-noise ratios. The true values for σE2\sigma^{2}_{E} and σW2\sigma^{2}_{W} used in the simulation study are (0.0001,0.001,0.01)(0.0001,0.001,0.01) and (0.00001,0.00005)(0.00001,0.00005), respectively. Petris et al. (2009) define the signal-to-noise ratio rr as follows:

    r=Observation VarianceSystem Variance=σE2σW2.r=\frac{\mbox{Observation Variance}}{\mbox{System Variance}}=\frac{\sigma^{2}_{E}}{\sigma^{2}_{W}}.

    The signal-to-noise ratios rr in the simulation study were examined in two sets. First, rr is set equal to 10, 100, and 1000. Next, the ratio rr was set to equal 2, 20, and 200. The variety of rr values allow us to examine the methods under different levels of noise. Each simulation is repeated 100 times for both the 2- and 5-point calibration models, thus providing us with 36 possible models for examination from the settings of rr.
    After the data was fit with each of the methods we considered the following measures for assessing the performance of the dynamic methods compared to the familiar static approaches: (1) average mean square error; (2) average coverage probability; and (3) average interval width. For each of the simulated data sets, the mean squared error (MSEMSE) is calculated as

    MSE=1Tt=1T(x^0tx0t)2.MSE=\frac{1}{T}\sum_{t=1}^{T}(\hat{x}_{0t}-x_{0t})^{2}.

    The MSEMSE are averaged across the 100 simulated data sets thus deriving an average mean squared error (AvMSEAvMSE) as

    AvMSE=1100j=1100MSEj.AvMSE=\frac{1}{100}\sum_{j=1}^{100}MSE_{j}.

    The coverage probability based on the 95%95\% coverage interval is estimated for all of the calibration methods. The coverage interval for the dynamic and static Bayesian approaches is the 95%95\% credible interval and the 95%95\% confidence interval is used for the frequentist methods. Note, that for credible intervals x0tLx_{0t}^{L} is the 0.025 posterior quantile for x0tx_{0t}, and x0tUx_{0t}^{U} is the 0.975 posterior quantile for x0tx_{0t}, where x0tx_{0t} is the true value of the calibration target from the second stage of experimentation, then (x0tL,x0tU)(x_{0t}^{L},x_{0t}^{U}) is a 95%95\% credible interval. The coveraged probability (CPCP) is calculated as such

    CP=1Tt=1TψtCP=\frac{1}{T}\sum_{t=1}^{T}\psi_{t}

    where

    ψt=P[x0tL<x0t<x0tU]={0if x0t(x0tL,x0tU);1if x0t(x0tL,x0tU).\psi_{t}=P[x_{0t}^{L}<x_{0t}<x_{0t}^{U}]=\left\{\begin{array}[]{ll}0&\mbox{if $x_{0t}\not\in(x_{0t}^{L},x_{0t}^{U})$};\\ &\\ 1&\mbox{if $x_{0t}\in(x_{0t}^{L},x_{0t}^{U})$}.\end{array}\right.

    The average coverage probability (AvCPAvCP) is calculated by averaging across the number of replications in the simulation study, where

    AvCP=1100j=1100CPj.AvCP=\frac{1}{100}\sum_{j=1}^{100}CP_{j}.

    Another quantity of interest to compare the average interval widths (AvIWAvIW) for the methods, where the average interval widths (IWIW) across the simulated time series is calculated as follows:

    IW=1Tt=1T(x0tUx0tL)IW=\frac{1}{T}\sum_{t=1}^{T}(x_{0t}^{U}-x_{0t}^{L})

    with the average interval width across the simulation study given as

    AvIW=1100j=1100IWj,AvIW=\frac{1}{100}\sum_{j=1}^{100}IW_{j},

    where IWjIW_{j} is the average interval width for the jthj^{th} simulation replicate. The performance of the dynamic calibration approaches will be assess using the average coverage probability (AvCPAvCP), average interval width (AvIWAvIW) and average mean square (AvMSEAvMSE).
    We consider the performance of the methods under two conditions: interpolation and extrapolation. Interpolation case is of interest to understand how the method performed when the calibrated time series is within the range of the reference values, [20, 100]. Extrapolation case also conducted to examine the methods when x0tx_{0t} falls outside of the range of the calibration references, where x0t>100x_{0t}>100. While it not preferable to do extrapolation in the regression case, it is often done in practice in microwave radiometry.
    All simulations were carried out on the Compile server running R3.0.2R~3.0.2 (R Development Core Team, 2013) at Virginia Commonwealth University. The Compile server has a Linux OS with 16 CPU cores and 32 GB Ram. Each iteration in the study took approximately 15 minutes with a total of 25.63 hours.

    1.   Interpolation case

    In the following tables, the simulation results for the dynamic and static calibration methods are provided. The results of simulation studies provide insight into the properties of the calibration approaches. The results in Tables 1 and 1 indicate that all of the estimators do a good job at approximating the true values of x0tx_{0t} when the gain flucuation gtg_{t} is set to 0. Even in this case we see as the signal-to-noise ratio rr increases so does the AvMSEAvMSE values. All of the methods have an average coverage probability AvCPAvCP of 1 or close. The high coverage rate is of no surprise for a stable system. There does not appear to be an advantage by including more reference measurements (i.e 2- or 5-points) in the model when the system is stable in time. The clear difference is the AvIWAvIW values for the dynamic methods compared to the static methods. In Tables 1 and 1 when r=10r=10 and r=2r=2, the interval for the dynamic methods is wider than those of the four static methods but as rr increases the interval width of the dynamic methods remain nearly unchanged as the interval widths for the static methods are 4 to 5 times wider.
    The simulation results for the stepped gain fluctuations are provided in Tables 1 and 1. Clearly the presence of the stepped gtg_{t} has an effect on the fit of the models. The results in Tables 1 show that in nearly all cases, the two dynamic methods MD1M_{D1} and MD2M_{D2} have AvMSEAvMSE values smaller than the two static Bayesian approaches. The AvMSEAvMSE values for the dynamic methods are reasonably lower for r=200r=200. When r=1000r=1000, notice the dynamic models MD1M_{D1} and MD2M_{D2} have smaller average mean square errors smaller than the static method MF2M_{F2}. The average coverage probability AvCPAvCP is comparable for all of the methods and number of references. The dynamic methods consistently have shorter interval widths. The widths of the 95% credible intervals for MD1M_{D1} and MD2M_{D2} is not affected by the increases in rr.

    Table 1: Comparison of calibration approaches when interpolating to estimate x0tx_{0t} without gain fluctuations based on 100 data sets. AvMSE is the average mean squared error, AvCP is the average coverage probability, and AvIW is the average 95% interval width. The signal-to-noise ratio is denoted as rr.
    Constant gt=0g_{t}=0
    r=10r=10 r=100r=100 r=1000r=1000
    Ref. Model AvMSE AvCP AvIW AvMSE AvCP AvIW AvMSE AvCP AvIW
    2 MD1M_{D1} 0.0008 0.995 2.519 0.0035 0.983 2.523 0.0307 0.939 2.517
    MD2M_{D2} 0.0012 1.000 3.782 0.0038 1.000 3.782 0.0308 1.000 3.782
    MF1M_{F1} 0.0001 1.000 1.224 0.0012 1.000 3.868 0.0123 1.000 12.229
    MF2M_{F2} 0.0001 1.000 1.223 0.0016 1.000 3.863 0.0335 1.000 12.168
    MB1M_{B1} 0.0002 0.997 1.182 0.0022 1.000 3.866 0.0386 1.000 12.177
    MB2M_{B2} 0.0014 1.000 1.458 0.0139 1.000 4.606 0.1391 1.000 14.565
    5 MD1M_{D1} 0.0008 0.995 2.496 0.0035 0.983 2.509 0.0307 0.941 2.514
    MD2M_{D2} 0.0013 1.000 3.983 0.0039 1.000 3.983 0.0307 1.000 3.983
    MF1M_{F1} 0.0001 1.000 1.223 0.0012 1.000 3.865 0.0123 1.000 12.220
    MF2M_{F2} 0.0001 1.000 1.222 0.0022 1.000 3.860 0.0813 1.000 12.113
    MB1M_{B1} 0.0002 1.000 1.223 0.0023 1.000 3.861 0.0792 1.000 12.116
    MB2M_{B2} 0.0014 1.000 1.457 0.0139 1.000 4.604 0.1069 1.000 10.748
    Table 2: Comparison of calibration approaches when interpolating to estimate x0tx_{0t} without gain fluctuations based on 100 data sets. AvMSE is the average mean squared error, AvCP is the average coverage probability, and AvIW is the average 95% interval width. The signal-to-noise ratio is denoted as rr.
    Constant gt=0g_{t}=0
    r=2r=2 r=20r=20 r=200r=200
    Ref. Model AvMSE AvCP AvIW AvMSE AvCP AvIW AvMSE AvCP AvIW
    2 MD1M_{D1} 0.0012 0.992 2.519 0.0041 0.981 2.520 0.0323 0.939 2.528
    MD2M_{D2} 0.0015 1.000 3.782 0.0044 1.000 3.782 0.0325 1.000 3.782
    MF1M_{F1} 0.0001 1.000 1.230 0.0010 1.000 3.871 0.0114 1.000 12.231
    MF2M_{F2} 0.0001 1.000 1.229 0.0012 1.000 3.866 0.0314 1.000 12.170
    MB1M_{B1} 0.0001 1.000 1.230 0.0019 1.000 3.869 0.0371 1.000 12.179
    MB2M_{B2} 0.0190 1.000 1.155 0.0243 1.000 3.767 0.1381 1.000 14.567
    5 MD1M_{D1} 0.0011 0.992 2.508 0.0041 0.981 2.510 0.032 0.939 2.514
    MD2M_{D2} 0.0017 1.000 3.983 0.0045 1.000 3.983 0.032 1.000 3.983
    MF1M_{F1} 0.0001 1.000 1.228 0.0010 1.000 3.868 0.011 1.000 12.222
    MF2M_{F2} 0.0001 1.000 1.227 0.0019 1.000 3.863 0.081 1.000 12.114
    MB1M_{B1} 0.0001 1.000 1.227 0.0021 1.000 3.864 0.082 1.000 12.118
    MB2M_{B2} 0.0013 1.000 1.462 0.0137 1.000 4.607 0.138 1.000 14.560
    Table 3: Comparison of calibration approaches when interpolating to estimate x0tx_{0t} with stepped gain fluctuations based on 100 data sets. AvMSE is the average mean squared error, AvCP is the average coverage probability, and AvIW is the average 95% interval width. The signal-to-noise ratio is denoted as rr.
    Stepped gt=aig_{t}=a_{i}
    r=10r=10 r=100r=100 r=1000r=1000
    Ref. Model AvMSE AvCP AvIW AvMSE AvCP AvIW AvMSE AvCP AvIW
    2 MD1M_{D1} 0.0191 0.961 2.509 0.0198 0.953 2.506 0.0406 0.926 2.543
    MD2M_{D2} 0.0196 1.000 3.782 0.0201 1.000 3.782 0.0408 1.000 3.783
    MF1M_{F1} 0.0001 1.000 9.094 0.0004 1.000 9.813 0.0094 1.000 15.209
    MF2M_{F2} 0.0046 1.000 9.065 0.0073 1.000 9.779 0.0528 1.000 15.098
    MB1M_{B1} 0.0859 1.000 9.072 0.0838 1.000 9.786 0.1866 1.000 15.109
    MB2M_{B2} 0.1399 1.000 10.830 0.0823 1.000 11.687 0.1836 1.000 18.115
    5 MD1M_{D1} 0.0191 0.961 2.510 0.0197 0.954 2.511 0.0405 0.924 2.516
    MD2M_{D2} 0.0196 1.000 3.983 0.0201 1.000 3.983 0.0405 1.000 3.983
    MF1M_{F1} 0.0001 1.000 9.087 0.0004 1.000 9.806 0.0094 1.000 15.199
    MF2M_{F2} 0.0184 1.000 9.041 0.0267 1.000 9.749 0.1620 1.000 14.995
    MB1M_{B1} 0.0199 1.000 9.044 0.0267 1.000 9.752 0.1559 1.000 14.999
    MB2M_{B2} 0.0706 1.000 10.826 0.0618 1.000 8.625 0.1742 1.000 15.091
    Table 4: Comparison of calibration approaches when interpolating to estimate x0tx_{0t} with stepped gain fluctuations based on 100 data sets. AvMSE is the average mean squared error, AvCP is the average coverage probability, and AvIW is the average 95% interval width. The signal-to-noise ratio is denoted as rr.
    Stepped gt=aig_{t}=a_{i}
    r=2r=2 r=20r=20 r=200r=200
    Ref. Model AvMSE AvCP AvIW AvMSE AvCP AvIW AvMSE AvCP AvIW
    2 MD1M_{D1} 0.0209 0.957 2.520 0.0219 0.950 2.522 0.0436 0.921 2.526
    MD2M_{D2} 0.0214 1.000 3.782 0.0222 1.000 3.782 0.0438 1.000 3.782
    MF1M_{F1} 0.0001 1.000 9.103 0.0003 1.000 9.822 0.0086 1.000 15.216
    MF2M_{F2} 0.0047 1.000 9.075 0.0073 1.000 9.788 0.0511 1.000 15.105
    MB1M_{B1} 0.0084 1.000 9.081 0.0115 1.000 9.795 0.0601 1.000 15.116
    MB2M_{B2} 0.0709 1.000 10.842 0.0826 1.000 11.698 0.2054 1.000 18.122
    5 MD1M_{D1} 0.0209 0.957 2.509 0.0218 0.949 2.511 0.0436 0.920 2.516
    MD2M_{D2} 0.0214 1.000 3.983 0.0221 1.000 3.983 0.0435 1.000 3.983
    MF1M_{F1} 0.0001 1.000 9.096 0.0003 1.000 9.815 0.0086 1.000 15.205
    MF2M_{F2} 0.0185 1.000 9.050 0.0267 1.000 9.758 0.1616 1.000 15.002
    MB1M_{B1} 0.0199 1.000 9.053 0.0281 1.000 9.761 0.1641 1.000 15.006
    MB2M_{B2} 0.0708 1.000 10.836 0.0825 1.000 11.693 0.2052 1.000 18.114
    Table 5: Comparison of calibration approaches when interpolating to estimate x0tx_{0t} with sinusoidal gain fluctuations based on 100 data sets. AvMSE is the average mean squared error, AvCP is the average coverage probability, and AvIW is the average 95% interval width. The signal-to-noise ratio is denoted as rr.
    Sinusoidal gt=0.1sin(0.025t)g_{t}=0.1\mbox{sin}(0.025t)
    r=10r=10 r=100r=100 r=1000r=1000
    Ref. Model AvMSE AvCP AvIW AvMSE AvCP AvIW AvMSE AvCP AvIW
    2 MD1M_{D1} 4.4088 0.628 2.657 4.4794 0.629 2.648 4.7214 0.638 2.681
    MD2M_{D2} 4.4002 0.829 3.783 4.4708 0.825 3.783 4.7123 0.810 3.783
    MF1M_{F1} 0.0001 1.000 21.980 0.0012 1.000 22.307 0.0123 1.000 25.206
    MF2M_{F2} 0.1541 1.000 21.665 0.1670 1.000 21.978 0.2943 1.000 24.738
    MB1M_{B1} 0.1689 0.975 20.933 0.1868 1.000 21.994 0.3174 1.000 24.757
    MB2M_{B2} 0.4127 1.000 26.178 0.4258 1.000 26.567 0.5531 1.000 30.020
    5 MD1M_{D1} 4.4087 0.628 2.646 4.4793 0.630 2.648 4.7214 0.635 2.653
    MD2M_{D2} 4.3906 0.845 3.984 4.4609 0.839 3.984 4.7023 0.824 3.984
    MF1M_{F1} 0.0001 1.000 21.964 0.0012 1.000 22.291 0.0123 1.000 25.188
    MF2M_{F2} 0.5810 1.000 21.371 0.6218 1.000 21.671 1.0152 1.000 24.306
    MB1M_{B1} 0.5956 1.000 21.377 0.5909 1.000 21.678 0.9628 1.000 24.314
    MB2M_{B2} 0.4123 1.000 26.166 0.3087 1.000 18.973 0.4658 1.000 25.009
    Table 6: Comparison of calibration approaches when interpolating to estimate x0tx_{0t} with sinusoidal gain fluctuations based on 100 data sets. AvMSE is the average mean squared error, AvCP is the average coverage probability, and AvIW is the average 95% interval width. The signal-to-noise ratio is denoted as rr.
    Sinusoidal gt=0.1sin(0.025t)g_{t}=0.1\mbox{sin}(0.025t)
    r=2r=2 r=20r=20 r=200r=200
    Ref. Model AvMSE AvCP AvIW AvMSE AvCP AvIW AvMSE AvCP AvIW
    2 MD1M_{D1} 4.4504 0.625 2.658 4.5213 0.628 2.660 4.7643 0.6331 2.665
    MD2M_{D2} 4.4419 0.828 3.783 4.5127 0.823 3.783 4.7553 0.8083 3.783
    MF1M_{F1} 0.0001 1.000 21.968 0.0010 1.000 22.295 0.0114 1.000 25.196
    MF2M_{F2} 0.1538 1.000 21.653 0.1672 1.000 21.966 0.2896 1.000 24.729
    MB1M_{B1} 0.1732 1.000 21.669 0.1867 1.000 21.982 0.3127 1.000 24.748
    MB2M_{B2} 0.4122 1.000 26.164 0.4253 1.000 26.553 0.5518 1.000 30.008
    5 MD1M_{D1} 4.4504 0.625 2.647 4.5213 0.627 2.648 4.7643 0.633 2.654
    MD2M_{D2} 4.4322 0.844 3.984 4.5029 0.838 3.984 4.7451 0.823 3.984
    MF1M_{F1} 0.0001 1.000 21.952 0.0010 1.000 22.278 0.0114 1.000 25.178
    MF2M_{F2} 0.5799 1.000 21.359 0.6206 1.000 21.660 1.0133 1.000 24.297
    MB1M_{B1} 0.5851 1.000 21.366 0.6256 1.000 21.667 1.0183 1.000 24.304
    MB2M_{B2} 0.4119 1.000 26.152 0.4247 1.000 26.541 0.5513 1.000 29.995

    The results provided in Tables 1 and 1 summarize the performance of the methods when the gain fluctuation is sinusoidal noise. The results for rr values of 10, 100, and 1000 are given in Table 1 with r=2,20r=2,~20 and 200 given in Table 1. When gtg_{t} is sinusoidal, the AvMSEAvMSE values for the dynamic methods are consistently larger than any of the static methods. For all of the chosen rr values, the AvCPAvCP is considerably lower than the opposing methods. The dynamic methods still have average interval widths extremely shorter than any of the static methods. The AvIWAvIW is constant across the signal-to-noise ratios.
    The simulation study shows that methods MD1M_{D1} and MD2M_{D2} do a good job at estimating calibrated values that are interior to the range of reference measurements. Both methods display high coverage probabilities in the presence of drifting parameters. For the three possible gain fluctuations, the interval widths for the dynamic methods were consistently shorter than the static calibration approaches. When fitting data where there is a definite linear relationship the dynamic methods are invariant to the number of reference measurements. When using the proposed methods in this paper, not much will be gained by using more than 2 reference measurements. Overall, when interpolating to estimate x0tx_{0t}, the dynamic methods outperform the static Bayesian approaches across the different signal-to-noise ratios. In the following section the performance of the dynamic methods are assessed when the calibrated values fall outside of the range of reference measurements.

    2.   Extrapolation case

    At this point in the paper we examine the calibration approaches when the calibrated values are outside of the reference measurements. The range of the measurement references is from 20 to 100. The true x0tx_{0t} behaved as a random walk bounded between 100 and 110. We assessed the performance of the dynamic methods under three possible gain fluctuation patterns. First, the simulation study is conducted without the presence of additional gain fluctuation (i.e. gt=0g_{t}=0); second, the gain gtg_{t} is a stepped pattern influencing the time-varying slope β1t\beta_{1t} over time; lastly, a sinusoidal gtg_{t} is added to β1t\beta_{1t}. Just as the previous results, the methods are assessed by the average mean square error (AvMSEAvMSE), average coverage probability (AvCPAvCP), and the average interval width (AvIWAvIW) under different signal-noise-ratios.

    Table 7: Comparison of calibration approaches when extrapolating to estimate x0tx_{0t} without gain fluctuations based on 100 data sets. AvMSE is the average mean squared error, AvCP is the average coverage probability, and AvIW is the average 95% interval width. The signal-to-noise ratio is denoted as rr.
    Constant gt=0g_{t}=0
    r=10r=10 r=100r=100 r=1000r=1000
    Ref. Model AvMSE AvCP AvIW AvMSE AvCP AvIW AvMSE AvCP AvIW
    2 MD1M_{D1} 0.0018 1.000 5.255 0.0043 1.000 5.255 0.0309 1.000 5.255
    MD2M_{D2} 0.0016 1.000 3.910 0.0042 1.000 3.910 0.0311 1.000 3.910
    MF1M_{F1} 0.0001 1.000 1.224 0.0012 1.000 3.869 0.0123 1.000 12.234
    MF2M_{F2} 0.0001 1.000 1.223 0.0001 1.000 3.863 0.1019 1.000 12.168
    MB1M_{B1} 0.0001 1.000 1.225 0.0008 1.000 3.867 0.1115 1.000 12.181
    MB2M_{B2} 0.0014 1.000 1.458 0.0139 1.000 4.606 0.1391 1.000 14.565
    5 MD1M_{D1} 0.0019 1.000 5.233 0.0043 1.000 5.233 0.0309 1.000 5.233
    MD2M_{D2} 0.0029 1.000 4.106 0.0054 1.000 4.106 0.0323 1.000 4.106
    MF1M_{F1} 0.0001 1.000 1.223 0.0012 1.000 3.866 0.0123 1.000 12.224
    MF2M_{F2} 0.0001 1.000 1.222 0.0027 1.000 3.860 0.5502 1.000 12.113
    MB1M_{B1} 0.0001 1.000 1.223 0.0030 1.000 3.862 0.5566 1.000 12.120
    MB2M_{B2} 0.0014 1.000 1.457 0.0139 1.000 4.604 0.1389 1.000 14.558
    Table 8: Comparison of calibration approaches when extrapolating to estimate x0tx_{0t} without gain fluctuations based on 100 data sets. AvMSE is the average mean squared error, AvCP is the average coverage probability, and AvIW is the average 95% interval width. The signal-to-noise ratio is denoted as rr.
    Constant gt=0g_{t}=0
    r=2r=2 r=20r=20 r=200r=200
    Ref. Model AvMSE AvCP AvIW AvMSE AvCP AvIW AvMSE AvCP AvIW
    2 MD1M_{D1} 0.0031 1.000 5.253 0.0060 1.000 5.253 0.0340 1.000 5.253
    MD2M_{D2} 0.0034 1.000 3.910 0.0064 1.000 3.910 0.0347 1.000 3.910
    MF1M_{F1} 0.0001 1.000 1.230 0.0008 1.000 3.872 0.0107 1.000 12.236
    MF2M_{F2} 0.0001 1.000 1.229 0.0003 1.000 3.866 0.1068 1.000 12.170
    MB1M_{B1} 0.0001 1.000 1.230 0.0010 1.000 3.871 0.1164 1.000 12.183
    MB2M_{B2} 0.0013 1.000 1.465 0.0135 1.000 4.610 0.1376 1.000 14.567
    5 MD1M_{D1} 0.0032 1.000 5.231 0.0060 1.000 5.231 0.0340 1.000 5.232
    MD2M_{D2} 0.0053 1.000 4.106 0.0083 1.000 4.106 0.0365 1.000 4.106
    MF1M_{F1} 0.0001 1.000 1.228 0.0008 1.000 3.869 0.0107 1.000 12.226
    MF2M_{F2} 0.0001 1.000 1.227 0.0035 1.000 3.863 0.5616 1.000 12.114
    MB1M_{B1} 0.0001 1.000 1.228 0.0039 1.000 3.865 0.5680 1.000 12.121
    MB2M_{B2} 0.0013 1.000 1.462 0.0135 1.000 4.607 0.1375 1.000 14.560

    The results are provided in Tables 2 and2 for the statistical calibration methods without gain fluctuations. The performance of the proposed method is stable across the signal-to-noise ratios. A point of interest is the reported AvIWAvIW values for methods MD1M_{D1} and MD2M_{D2}. We see for r=10r=10 and r=2r=2 that the AvIWAvIW is 3 to 5 times wider than those for the static approaches. When r=100r=100 and r=20r=20 the interval width for all competing methods are relatively close. The dynamic approaches outperform the static methods in noisy conditions such as r=1000r=1000 and r=200r=200. The interval widths for the dynamic methods are considerably shorter than the those for the static methods. The simulation results reveal that when the data is characteristic of having a large signal-to-noise ratio, the dynamic methods, MD1M_{D1} and MD2M_{D2}, will outperform static Bayesian approaches and the inverse approach.

    Table 9: Comparison of calibration approaches when extrapolating to estimate x0tx_{0t} with stepped gain fluctuations based on 100 data sets. AvMSE is the average mean squared error, AvCP is the average coverage probability, and AvIW is the average 95% interval width. The signal-to-noise ratio is denoted as rr.
    Stepped gt=aig_{t}=a_{i}
    r=10r=10 r=100r=100 r=1000r=1000
    Ref. Model AvMSE AvCP AvIW AvMSE AvCP AvIW AvMSE AvCP AvIW
    2 MD1M_{D1} 0.0206 1.000 5.247 0.0210 1.000 5.247 0.0412 1.000 5.247
    MD2M_{D2} 0.0225 1.000 3.910 0.0230 1.000 3.910 0.0435 1.000 3.910
    MF1M_{F1} 0.0001 1.000 9.097 0.0004 1.000 9.817 0.0094 1.000 15.215
    MF2M_{F2} 0.0581 1.000 9.065 0.0656 1.000 9.779 0.3191 1.000 15.098
    MB1M_{B1} 0.0634 1.000 9.075 0.0718 1.000 9.789 0.3361 1.000 15.115
    MB2M_{B2} 0.0707 1.000 10.830 0.0826 1.000 11.687 0.2060 1.000 18.115
    5 MD1M_{D1} 0.0209 1.000 5.226 0.0213 1.000 5.226 0.0412 1.000 5.226
    MD2M_{D2} 0.0268 1.000 4.106 0.0273 1.00 4.106 0.0483 1.000 4.106
    MF1M_{F1} 0.0001 1.000 9.090 0.0004 1.000 9.809 0.0094 1.000 15.203
    MF2M_{F2} 0.2274 1.000 9.041 0.2812 1.000 9.749 1.4628 1.000 14.995
    MB1M_{B1} 0.2307 1.000 9.047 0.2851 1.000 9.755 1.4744 1.000 15.004
    MB2M_{B2} 0.0706 1.000 10.826 0.0825 1.000 11.682 0.2058 1.000 18.106
    Table 10: Comparison of calibration approaches when extrapolating to estimate x0tx_{0t} with stepped gain fluctuations based on 100 data sets. AvMSE is the average mean squared error, AvCP is the average coverage probability, and AvIW is the average 95% interval width. The signal-to-noise ratio is denoted as rr.
    Stepped gt=aig_{t}=a_{i}
    r=2r=2 r=20r=20 r=200r=200
    Ref. Model AvMSE AvCP AvIW AvMSE AvCP AvIW AvMSE AvCP AvIW
    2 MD1M_{D1} 0.0242 1.000 5.245 0.0250 1.000 5.245 0.0466 1.000 5.245
    MD2M_{D2} 0.0266 1.000 3.910 0.0275 1.000 3.910 0.0494 1.000 3.910
    MF1M_{F1} 0.0001 1.000 9.106 0.0002 1.000 9.826 0.0080 1.000 15.222
    MF2M_{F2} 0.0620 1.000 9.075 0.0698 1.000 9.788 0.3284 1.000 15.105
    MB1M_{B1} 0.0674 1.000 9.085 0.0760 1.000 9.799 0.3447 1.000 15.121
    MB2M_{B2} 0.0710 1.000 10.842 0.0825 1.000 11.698 0.2048 1.000 18.122
    5 MD1M_{D1} 0.0245 1.000 5.224 0.0254 1.000 5.224 0.0427 1.000 5.226
    MD2M_{D2} 0.0315 1.000 4.106 0.0324 1.000 4.106 0.0485 1.000 4.106
    MF1M_{F1} 0.0001 1.000 9.099 0.0002 1.000 9.818 0.0089 1.000 14.255
    MF2M_{F2} 0.2354 1.000 9.050 0.2902 1.000 9.758 1.1896 1.000 14.078
    MB1M_{B1} 0.2388 1.000 9.056 0.2941 1.000 9.764 1.1995 1.000 14.086
    MB2M_{B2} 0.0709 1.000 10.836 0.0824 1.000 11.693 0.1831 1.000 16.977
    Table 11: Comparison of calibration approaches when extrapolating to estimate x0tx_{0t} with sinusoidal gain fluctuations based on 100 data sets. AvMSE is the average mean squared error, AvCP is the average coverage probability, and AvIW is the average 95% interval width. The signal-to-noise ratio is denoted as rr.
    Sinusoidal gt=0.1sin(0.025t)g_{t}=0.1\mbox{sin}(0.025t)
    r=10r=10 r=100r=100 r=1000r=1000
    Ref. Model AvMSE AvCP AvIW AvMSE AvCP AvIW AvMSE AvCP AvIW
    2 MD1M_{D1} 4.4096 0.873 5.127 4.4800 0.872 5.127 4.7214 0.866 5.127
    MD2M_{D2} 4.4410 0.833 3.904 4.5114 0.825 3.904 4.7530 0.813 3.904
    MF1M_{F1} 0.0001 1.000 21.988 0.0012 1.000 22.315 0.0123 1.000 25.216
    MF2M_{F2} 1.8193 1.000 21.665 1.8636 1.000 21.978 2.7760 1.000 24.739
    MB1M_{B1} 1.8602 1.000 21.688 1.9056 1.000 22.002 2.8312 1.000 24.766
    MB2M_{B2} 0.4127 1.000 26.178 0.4258 1.000 26.567 0.5531 1.000 30.020
    5 MD1M_{D1} 4.4105 0.872 5.106 4.4808 0.872 5.106 4.7222 0.866 5.107
    MD2M_{D2} 4.4889 0.842 4.100 4.5593 0.835 4.101 4.8007 0.822 4.100
    MF1M_{F1} 0.0001 1.000 21.971 0.0012 1.000 22.297 0.0123 1.000 25.195
    MF2M_{F2} 6.9539 1.000 21.371 7.2327 1.000 21.671 11.0337 1.000 24.306
    MB1M_{B1} 6.9852 1.000 21.383 7.2650 1.000 21.684 11.0772 1.000 24.320
    MB2M_{B2} 0.4123 1.000 26.166 0.4254 1.000 26.555 0.5526 1.000 30.007
    Table 12: Comparison of calibration approaches when extrapolating to estimate x0tx_{0t} with sinusoidal gain fluctuations based on 100 data sets. AvMSE is the average mean squared error, AvCP is the average coverage probability, and AvIW is the average 95% interval width. The signal-to-noise ratio is denoted as rr.
    Sinusoidal gt=0.1sin(0.025t)g_{t}=0.1\mbox{sin}(0.025t)
    r=2r=2 r=20r=20 r=200r=200
    Ref. Model AvMSE AvCP AvIW AvMSE AvCP AvIW AvMSE AvCP AvIW
    2 MD1M_{D1} 4.4491 0.872 5.125 4.5199 0.871 5.125 4.7626 0.866 5.126
    MD2M_{D2} 4.4809 0.828 3.904 4.5518 0.821 3.904 4.7948 0.807 3.904
    MF1M_{F1} 0.0001 1.000 21.976 0.0008 1.000 22.303 0.0107 1.000 25.205
    MF2M_{F2} 1.8350 1.000 21.653 1.8796 1.000 21.966 2.7956 1.000 24.729
    MB1M_{B1} 1.8759 1.000 21.676 1.9216 1.000 21.990 2.8508 1.000 24.756
    MB2M_{B2} 0.4123 1.000 26.164 0.4250 1.000 26.553 0.5511 1.000 30.008
    5 MD1M_{D1} 4.4497 0.872 5.10 4.5205 0.871 5.105 4.7633 0.865 5.105
    MD2M_{D2} 4.5292 0.836 4.100 4.6000 0.832 4.100 4.842 0.814 4.100
    MF1M_{F1} 0.0001 1.000 21.958 0.0008 1.000 22.285 0.0107 1.000 25.185
    MF2M_{F2} 6.9764 1.000 21.359 7.2560 1.000 21.660 11.0629 1.000 24.297
    MB1M_{B1} 7.0077 1.000 21.372 7.2882 1.000 21.673 11.1064 1.000 24.311
    MB2M_{B2} 0.4119 1.000 26.152 0.4246 1.000 26.541 0.5507 1.000 29.995

    Next, we impose a stepped gain fluctuation gtg_{t} to the data generated and wanted to evaluate the behavior of the calibration methods. The results for the stepped case are given in Tables 2 and 2. We see by the AvMSEAvMSE values in both tables that the dynamic methods perform better than most static methods. If the calibrated values by chance drift outside of the reference range the dynamic methods will do a good job at capturing it with certainty while having a narrower credible interval than confidence intervals of the static methods. The dynamic approaches outperform all of the static method in terms of AvIWAvIW. These results of the simulation study do not change much across the number of references used. Once again, when the relationship is assumed to be linear there is no benefit to adding more references.
    Lastly, the study is conducted with a sinusoidal gain fluctuation while extrapolating to estimate x0tx_{0t}. The results for the sinusoidal case are given in Tables 2 and 2. The dynamic methods MD1M_{D1} and MD2M_{D2} exhibit the same behavior as before in Tables 1 and 1 with AvMSEAvMSE values ranging for 4.4 to 4.8. Even though the average mean square errors are larger than those of the static methods when using a 2-reference model, the two dynamic methods outperform the static methods MF2M_{F2} and MB1M_{B1} which are based on the inverse approach. The dynamic models have average coverage probabilities smaller than the static model across all of the signal-to-noise ratios. We can not fail to point out that once again the AvIWAvIW are 4 to 6 times shorter than the average widths for the static models.

    4. Application to Microwave Radiometer

    In this example, we apply the dynamic calibration approaches to the calibration of a microwave radiometer for an earth observing satellite. Engineers and scientist commonly use microwave radiometers to measure the electromagnetic radiation emitted by some source or a particular surface such as ice or land surface. Radiometers are very sensitive instruments that are capable of measuring extremely low levels of radiation. The transmission source of the radiant power is the target of the radiometers antenna. When the region of interest, such as terrain, is observed by a microwave radiometer, the radiation received by the antenna is partly due to self-emission by the area of interest and partly due to the reflected radiation originating from the surroundings (Ulaby et al. 1981) such as cosmic background radiation, ocean surface, or a heated surface used for the purpose of calibration.
    A basic diagram of a radiometer is shown in Figure 2 where the radiant power with equivalent brightness temperature (i.e. the term brightness temperature represents the intensity of the radiation emitted by the scene under observation) TAT_{A} enters the radiometer receiver and is converted to the output signal v(t)v(t).

    Refer to caption
    Figure 2: Schematic of Simple Radiometer

    The schematic features the common components of most microwave radiometers. As the radiometer captures a signal (i.e. Brightness Temperature TAT_{A}), it couples the signal into a transmission line which then carries the signal to and from the various elements of the circuit. In Figure 2, a signal TAT_{A} is introduced directly into the antenna, then it is mixed, amplified and filtered to produce the output signal v(t)v(t). This filtering and amplification of the signal is carried out through the following components of the radiometer: an amplifier (g0)(g_{0}); pre-detection filter (H)(H); a square law detector (ξ2)(\xi^{2}); and a post-detection filter (W)(W). The output of the radiometer is denoted as v(t)v(t). See Ulaby et al. (1981) for a detailed discussion.
    Racette and Lang (2005) state that at the core of every radiometer measurement is a calibrated receiver. Calibration is required due to the fact that the current electronic hardware is unable to maintain a stable input/output relationship. For space observing instruments, stable calibration without any drifts is a key to detect proper trends of climate (Imaoka et al. 2010). Due to problems such as amplifier gain instability and exterior temperature variations of critical components that may cause this relationship to drift over time (Bremer 1979). During the calibration process, the radiometer receiver measures the voltage output power v(t)v(t), and its corresponding input temperature of a known reference. Two or more known reference temperatures are needed for calibration of a radiometer. Ulaby et al. (1981); Racette and Lang (2005) state that the relationship between the output, v(t)v(t) and the input, TAT_{A} is approximately linear, and can be expressed as

    T^A=β0+β1v(t)\hat{T}_{A}=\beta_{0}+\beta_{1}v(t)

    where, T^A\hat{T}_{A} is the estimated value of the brightness temperature, v(t)v(t) is the observed output voltage. Using this relationship, the output value, v(t)v(t), is used to derive an estimate for the input, TAT_{A} (Racette and Lang, 2005).

    Refer to caption
    Figure 3: Known Reference Temperature Collection

    Traditional calibration methods use measurements taken from known calibration references, for example see Figure 3. Due to possible cost constraints it is common to use between two and five references. The reference temperatures are converted to their equivalent power measurement prior to the calibration algorithm. The radiometer outputs are observed when the radiometer measures the reference temperatures, giving an ordered calibration pair (Ti,vi)(T_{i},v_{i}). The viv_{i} values are observed from the process of the electronics within the radiometer (see Figure 2) (Ulaby et al. 1981; Racette and Lang 2005). Through the process of calibration, the unknown brightness temperature TjT_{j} is estimated by plugging its observed output vjv_{j} into either Equation (3) or Equation (4).
    It is of interest to develop a calibration approach that can detect gain abnormalities, and/or correct for slow drifts that affect the quality of the instrument measurements. To demonstrate the dynamic approach in terms of application appeal, the two dynamic methods were used to characterize a calibration target over time for a microwave radiometer. The data used for this example was collected during a calibration experiment that was conducted on the Millimeter-wave Imaging Radiometer (MIR) (Racette et al. 1995). The purpose of the experiment was to validate predictions of radiometer calibration.

    Refer to caption
    Figure 4: Time series of MIR output voltage measurement data VcoldV_{cold}, VhotV_{hot}, and VskyV_{sky}.

    The MIR was built with two internal blackbody references which will be used to observe a third stable temperature reference for an extended period of time. The third reference was a custom designed cryogenically cooled reference. Racette (2005) conducted the MIR experiment under two scenarios: the first experiment denoted as T295T295 examined the calibration predictions when the unknown target is interior (i.e. interpolation) to the reference measurements; the second set of measurements (denoted as T80T80) where taken when the unknown temperature to be estimated is outside (i.e. extrapolation) of the range of calibration references.
    For demonstration purposes we will only consider the T80T80 experiment, for details of the T295T295 experiment see Racette (2005). For the T80T80 run of the experiment, the reference temperatures are as follows:

    1. 1.

      Tcold293.69KT_{cold}\sim 293.69K

    2. 2.

      Thot325.59KT_{hot}~\sim 325.59K

    with the unknown target temperature that must be estimated denoted as TskyT_{sky}. Each temperature measure has a corresponding observed time series of output measurements; VcoldV_{cold}, VhotV_{hot}, and VskyV_{sky} (see Figure 4). Therefore in this example we only consider a 2-point calibration set-up as we use TcoldT_{cold} and ThotT_{hot} as the known reference standards and use VskyV_{sky} to derive estimates of TskyT_{sky} for the first 1000 time periods.
    The results of the dynamic approaches: MD1M_{D1} and MD2M_{D2}, will be compared to the “inverse" calibration method (Krutchkoff 1967) implemented by Racette (2005). The method considered by Racette (2005) will be denoted as M1uM_{1u}. As in practice, rarely does one know the value of the true temperature to be estimated so the aim of this example is to assess the contribution of the calibration approach to the variability in the measurement estimate. Racette (2005) analysis did not consider biases that may exist in calibration, continuing in the same spirit, the existence of biases will not be considered in the analysis. We will apply the M1uM_{1u}, MD1M_{D1}, and MD2M_{D2} approaches to the data to estimate the temperature TskyT_{sky}; the standard deviation of the estimated time series σ^Tsky\hat{\sigma}_{T_{sky}} is used as a measure of uncertainty including the contribution of the calibration algorithm.

    Refer to caption
    Figure 5: Time series of calibrated temperature for MIR T80T80 experiment. Black lines indicate the results using the “inverse" calibration approach M1uM_{1u}; the green lines are the results using the dynamic approach MD1M_{D1} with the 95% credible intervals in purple.
    Refer to caption
    Figure 6: Time series of calibrated temperature for MIR T80T80 experiment. Black lines indicate the results using the “inverse" calibration approach T1uT_{1u}; the red lines are the results using the dynamic approach MD2M_{D2} with the 95% credible intervals in purple.

    Figures 5 shows the time series of the temperature estimates for TskyT_{sky} using Krutchkoff’s (1967) “inverse" approach M1uM_{1u} and the dynamic approach MD1M_{D1}. The standard deviations for the M1uM_{1u} and MD1M_{D1} approaches are σ^Tsky(M1u)=1.482K\hat{\sigma}_{T_{sky}}(M_{1u})=1.482K and σ^Tsky(MD1)=0.998K\hat{\sigma}_{T_{sky}}(M_{D1})=0.998K, respectfully. We see the dynamic model MD1M_{D1} improves the estimation process over the static model M1uM_{1u} by observing the corresponding standard deviation values. The dynamic model decreased the measurement uncertainty by roughly 33%. In Figure 6 the time series of the temperature estimates for TskyT_{sky} using the “inverse" approach M1uM_{1u} and the dynamic approach MD2M_{D2} is given. The standard deviations for the M1uM_{1u} and MD2M_{D2} approaches are σ^Tsky(M1u)=1.482K\hat{\sigma}_{T_{sky}}(M_{1u})=1.482K and σ^Tsky(MD1)=0.974K\hat{\sigma}_{T_{sky}}(M_{D1})=0.974K. Again, the dynamic approach outperforms the static model M1uM_{1u}. In this case, dynamic model MD2M_{D2} decreased the measurement uncertainty by roughly 34%.

    5. Discussion

    Two new novel approaches to the statistical calibration problem have been presented in this paper. In was shown by the simulation results that the use of the dynamic approach has its benefits over the static methods. If the linear relationship in the first stage of calibration is known to be stable then the traditional methods should be used. The dynamic methods showed promise in the cases when the signal-to-noise ratio was high. There is also a computation expense to implementing the dynamic methods compared to the static methods, but in the sense of electronics these methods allow for near real time calibration and monitoring.
    It is worth noting that the dynamic method shows possible deficiencies when the gain fluctuations is sinusoidal, referring to results in Table 1. In Figure 7 it is evident the largest source of the error is in the beginning of estimation process, roughly from t=1t=1 to t=200t=200. The MSE values for the dynamic approaches; MD1M_{D1} and MD2M_{D2} were 4.41 and 4.40, respectively, which was vastly different than those reported for the static methods. This problem can be addressed by extending the burn-in period.

    Refer to caption
    Figure 7: Time series of MD1M_{D1} and MD2M_{D2} estimates in the interpolation case with gain fluctuations gt=0.1sin(0.025t)g_{t}=0.1\mbox{sin}(0.025t) (burn-in = 0).

    We increased the burn-in period to 200 which allowed the algorithm more time to learn and hence results in a lower MSE value. In Figure 8 we see that the estimates fit better to the true values of x0tx_{0t}. The MSE decreased from 4.41 to 0.64 for MD1M_{D1} and 0.63 for MD2M_{D2}. The increased burn-in period improves the coverage probability but the interval width isn’t noticeably affected. The coverage probability increased from 0.628 to 0.722 for MD1M_{D1} and from 0.829 to 0.964 for MD2M_{D2}.

    Refer to caption
    Figure 8: Time series of MD1M_{D1} and MD2M_{D2} estimates in the interpolation case with gain fluctuations gt=0.1sin(0.025t)g_{t}=0.1\mbox{sin}(0.025t) (burn-in = 200).
    Table 13: Comparison of calibration approaches MD1M_{D1} and MD2M_{D2} when interpolating to estimate x0tx_{0t} with sinusoidal gain fluctuation.
    2 References-Sinusoidal Gain- w//Burn In =200=200
    Mean Squared Error Coverage Probability Interval Width
    MD1M_{D1} 0.63553 0.72185 1.15722
    MD2M_{D2} 0.63333 0.96380 3.77191
    Refer to caption
    Figure 9: True time series of β0t\beta_{0t} and β1t\beta_{1t}
    Refer to caption
    Figure 10: Time series of true x0tx_{0t} and MD1M_{D1} estimate of x0tx_{0t}

    For completeness we consider the behavior of the method if β1t\beta_{1t} crosses zero. It is absurd to believe that this would happen in practice because one would test the significance of β1t\beta_{1t} (Myers 1990; Montgomery et al. 2012) for using any method where the possibility of dividing by zero could occur. We demonstrate this by generating data where β0t2\beta_{0t}\approx 2 for all time and β1t\beta_{1t} drifts from 1 to -1 over time where t=1,,1000t=1,\dots,1000 (see Figure 9). Figure 10 shows the dynamic method is close to the true values of x0tx_{0t} until β1t\beta_{1t} get close to 0. Within the region where the slope crosses the xaxisx-\mbox{axis} the posterior estimates become unstable. Here we define unstable as meaning that we are within a region where there is division by zero. This instability is only present when |β1t|<ϵ|\beta_{1t}|<\epsilon, for every ϵ>0\epsilon>0. As long as |β1t|>0|\beta_{1t}|>0 the dynamic method will perform well when estimating x0tx_{0t}.
    Some calibration problems are not linear or approximately linearly related in x0tx_{0t} and y0ty_{0t}. Future work is to investigate the dynamic calibration methods in the presence of nonlinearity. In such settings we may not have the ability to use only 2-points as references. Any approach will require more references in order to accurately capture the nonlinear behavior. Another area to be explored is using semiparametric regression which also allow for parameter variation across time and could be implemented in a near real time setting.

    References