This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

An Empirical Evaluation of the Approximation of Subjective Logic Operators Using Monte Carlo Simulations

Fabio Massimo Zennaro fabiomz@ifi.uio.no Magdalena Ivanovska magdalei@ifi.uio.no Audun Jøsang josang@mn.uio.no Department of Informatics, University of Oslo
PO Box 1080 Blindern
0316 Oslo, Norway
Abstract

In this paper we analyze the use of subjective logic as a framework for performing approximate transformations over probability distribution functions. As for any approximation, we evaluate subjective logic in terms of computational efficiency and bias. However, while the computational cost may be easily estimated, the bias of subjective logic operators have not yet been investigated. In order to evaluate this bias, we propose an experimental protocol that exploits Monte Carlo simulations and their properties to assess the distance between the result produced by subjective logic operators and the true result of the corresponding transformation over probability distributions. This protocol allows a modeler to get an estimate of the degree of approximation she must be ready to accept as a trade-off for the computational efficiency and the interpretability of the subjective logic framework. Concretely, we apply our method to the relevant case study of the subjective logic operator for binomial multiplication and fusion, and we study empirically their degree of approximation.

keywords:
subjective logic , Monte Carlo simulation , Beta distributions , binomial product , subjective logic fusion

1 Introduction

Subjective logic (SL) [4] defines a framework for expressing uncertain probabilistic statements in the form of subjective opinions. A subjective opinion allows a modeler to state probabilities over a set of alternative events along with a measure of the global uncertainty of such modeling. Subjective opinions thus integrate a form of first-order uncertainty, relative to the distribution of probability mass over events, and a form of second-order uncertainty, due to the incertitude in distributing the probability mass. Subjective opinions provide a simple, clean and interpretable way to encode and manipulate uncertainty; as such, they constitute a useful modelling tool in sensitive scenarios in which statistical models can not be inferred from data, but must be built relying on the domain knowledge or the intuition of experts. In this fashion, SL has been extensively adopted to model uncertainty in several fields such as trust modeling, biomedical data analysis or forensics analysis [4].

From a purely statistical point of view, subjective opinions can be seen as an alternative representation for standard probability distribution functions (pdfs), such as Beta pdfs or Dirichlet pdfs. Indeed, under certain assumptions, it is possible to define a unique mapping between subjective opinions and probability distribution functions [4]. This means that subjective opinions may be interpreted as a re-parametrization of standard distributions from the statistical literature.

SL also defines several operators over subjective opinions. These operators allow to carry out transformations over subjective opinions in a very efficient way. With respect to the underlying probability distributions, SL operators provide an extremely quick approximation of operations over probability distributions that would be otherwise very difficult or impossible to evaluate analytically.

Thus, beyond its original application, SL may also be seen as an effective statistical tool to compute approximate probability distributions generated by the transformations encoded into the SL operators. However, while the efficiency of SL operators may be easily evaluated, estimates about their bias are lacking. This shortcoming may limit the adoption of SL in favor of other better-studied approaches, such as Monte Carlo (MC) simulations. Modern probabilistic programming languages [2] provide a versatile language in which operations over probability distributions may be easily defined and evaluated using pre-coded inference algorithms. While being computationally more expensive, these techniques provide comforting guarantees on the convergence of the algorithms as a function of the number of sampling iterations. These guarantees, contrasting the lack of formal bounds of SL operators, may be a strong argument for many researchers to overlook SL and the related set of operators.

In this paper, we propose a protocol to address numerically the problem of characterizing the approximation of SL operators by offering an empirical analysis of their bias with respect to MC simulations. SL operators and MC simulations are taken as two distinct frameworks to approximate operations over pdfs, each one with its strenghts and limitations. Our analysis defines a quantitative comparison in which SL operators and MC simulations are contrasted in terms of the trade-off between computational efficiency and bias. More specifically, our approach allows to answer the question: What amount of approximation should we be ready to accept in exchange for the computational efficiency of subjective logic?

To show the usefulness of our protocol, we consider the specific case of binomial multiplication and fusion. Binomial multiplication is a simple SL operator that returns the approximation of the product of two Beta pdfs. Computing the product of independent Beta pdfs is a non-trivial problem [1] with relevant applications in fields such as reliability analysis and operations research [10]. Binomial multiplication in SL may then be seen as a simple and effective algorithm to compute an approximate solution to the problem of multiplying together two Beta pdfs. Fusion is a SL operator used for merging the opinions of different agents. This operator has been studied and applied in the context of second-order Bayesian networks [6]. For both operators, we compare the approximation obtained using SL to moment-matching approximation and kernel-density approximation produced via MC simulations. In this way, we are able to get an understanding of the amount of approximation that we should be ready to accept if we want to work in the framework of SL.

The rest of the paper is organized as follows. Section 2 reviews the basics of subjective logic and Section 3 presents the main aspects of computational statistics relevant to this work. Section 4 describes the computational complexity of SL approximations and MC approximations, while Section 5 discusses the bias of the same techniques. Section 6 proposes a grounded framework for evaluating the degree of approximation of SL operators in relation to MC simulations. Section 7 makes this framework concrete by applying it to the case study of the product of Beta pdfs, and it presents a set of empirical simulations to validate our approach; similarly, Section 8 applies our framework to another case study, the fusion of Beta pdfs, and it validates our methodology via empirical simulations. Finally, Section 9 summarizes the results and discusses possible directions for future work. For convenience and reference, Table 1 summarizes the notation that will be used throughout this paper.

Ω\Omega Collection of mutually exclusive events
MM Number of mutually exclusive events
X,Y,ZX,Y,Z... Random variables over Ω\Omega
x,y,zx,y,z... Sample of a random variable
ωX\omega_{X} Subjective logic opinion
𝐛,b\mathbf{b},b Belief (vector and scalar)
dd Disbelief (scalar)
𝐚,a\mathbf{a},a Prior probability (vector, scalar)
uu Uncertainty (scalar)
pXp_{X} Probability distribution function (pdf) of X
Mi[X]M_{i}\left[X\right] ii-th moment of the pdf of X
E[X]E\left[X\right], Var[X]Var\left[X\right] Expected value and variance of the pdf of X
DA[pX,pY]D_{A}\left[p_{X},p_{Y}\right] Distance AA between the pdf of X and the pdf of Y
p^X\hat{p}_{X} Empirical pdf for X estimated from samples
NN Number of samples
pXSLp_{X}^{SL} Pdf underlying a subjective logic opinion
p^XMC\hat{p}_{X}^{\mathrm{MC}} Empirical pdf for X estimated via Monte Carlo (MC) sampling
p^XKDE\hat{p}_{X}^{\mathrm{KDE}} Empirical pdf for X estimated via MC and kernel density estimation (KDE)
p^XMM\hat{p}_{X}^{\mathrm{MM}} Empirical pdf for X estimated via MC and moment matching (MM)
p^XGAUSS\hat{p}_{X}^{\mathrm{GAUSS}} Empirical pdf for X estimated via MC and MM with a Gaussian approximation
p^XBETA\hat{p}_{X}^{\mathrm{BETA}} Empirical pdf for X estimated via MC and ad MM with a Beta approximation
pXGAUSS{p}_{X}^{\mathrm{GAUSS}} Pdf for X estimated via analytic MM with a Gaussian approximation
pXBETA{p}_{X}^{\mathrm{BETA}} Pdf for X estimated via analytic MM with a Beta approximation
𝒫\mathcal{P} Space of probability distribution functions
𝒮\mathcal{S} Space of subjective opinions
P:𝒫×𝒫𝒫\circ_{P}:\mathcal{P}\times\mathcal{P}\rightarrow\mathcal{P} Binary operator on the space of probability distribution functions
SL:𝒮×𝒮𝒮\circ_{SL}:\mathcal{S}\times\mathcal{S}\rightarrow\mathcal{S} Binary operator on the space of subjective logic opinions
Table 1: Summary of notation.

2 Subjective Logic

In this section, we present the fundamentals of SL. We start with a formalization of subjective opinions and we show how they may be mapped to probability distributions.

Subjective opinions

Let Ω\Omega be a discrete collection of MM mutually exclusive and exhaustive events. A subjective opinion ω\omega is a triple:

(𝐛,u,𝐚),\left(\mathbf{b},u,\mathbf{a}\right), (1)

such that

i=1Mbi+u=1,\sum_{i=1}^{M}b_{i}+u=1, (2)

where 𝐛M\mathbf{b}\in\mathbb{R}^{M}, with bi0b_{i}\in\mathbb{R}_{\geq 0}, is the belief vector expressing the probability mass that the modeler places on each event xix_{i} in Ω\Omega, u0u\in\mathbb{R}_{\geq 0} is the uncertainty scalar quantifying the uncertainty of the modeler in its definition of 𝐛\mathbf{b}, and 𝐚M\mathbf{a}\in\mathbb{R}^{M} is the prior vector encoding a prior probability distribution over the events in Ω\Omega. This subjective opinion is called a multinomial opinion.
Notice that the constraint in Equation 2 limits the degrees of freedom of 𝐛\mathbf{b} and uu to MM and, consequently, defines a MM-dimensional simplex on which subjective opinions may be represented.

The limit-case multinomial opinion is the binomial opinion for M=2M=2. In this case Ω={x,x¯}\Omega=\{x,\overline{x}\} and the subjective opinion in Equation 1 may be re-written for simplicity as:

(b,d,u,a),\left(b,d,u,a\right), (3)

such that

b+d+u=1,b+d+u=1, (4)

where b0b\in\mathbb{R}_{\geq 0} is the belief scalar expressing the probability of xx, d0d\in\mathbb{R}_{\geq 0} is the disbelief scalar expressing the probability of x¯\overline{x}, u0u\in\mathbb{R}_{\geq 0} is the uncertainty scalar and a0a\in\mathbb{R}_{\geq 0} is a scalar expressing the prior probability of xx.
Having only two degrees of freedom, binomial opinions in the form (b,d,u,a)\left(b,d,u,a\right) belong to a two-dimensional simplex and may be visualized together with aa in a barycentric coordinate system111See http://folk.uio.no/josang/sl/BV.html for an illustration..

Mapping of subjective opinions

In order to ground SL, a mapping has been defined between multinomial opinions and Dirichlet pdfs and between binomial opinions and Beta pdfs.

Given a mapping constant W0W\in\mathbb{R}_{\geq 0}, it is possible to define a unique mapping from opinions to pdfs. Let ω=(𝐛,u,𝐚)\omega=\left(\mathbf{b},u,\mathbf{a}\right) be a multinomial opinion with u0u\neq 0; ω\omega can be mapped to a Dirichlet pdf pp with distribution 𝙳𝚒𝚛(𝜶)\mathtt{Dir}\left(\boldsymbol{\alpha}\right), where the vector of parameters 𝜶\boldsymbol{\alpha} is defined as:

𝜶=W(𝐛u+𝐚).\boldsymbol{\alpha}=W\left(\frac{\mathbf{b}}{u}+\mathbf{a}\right). (5)

For binomial opinions, specifically, given a mapping constant W0W\in\mathbb{R}_{\geq 0}, it is possible to define a unique mapping from opinions to pdfs. Let ω=(b,d,u,a)\omega=\left(b,d,u,a\right) be a binomial opinion with u0u\neq 0; ω\omega can be mapped to a Beta pdf pp with distribution 𝙱𝚎𝚝𝚊(α,β)\mathtt{Beta}\left(\alpha,\beta\right), where α\alpha and β\beta are parameters defined as:

{α=W(bu+a)β=W(du+(1a)).\begin{cases}\alpha=W\left(\frac{b}{u}+a\right)\\ \beta=W\left(\frac{d}{u}+(1-a)\right).\end{cases} (6)

Notice that, for reasons of consistency, WW is usually fixed to 22 [4]. We then have a mapping ss from opinion ω\omega to pdf pp:

s:ωp.s:\omega\mapsto p.

Vice versa, given a mapping constant W0W\in\mathbb{R}_{\geq 0} and a fixed prior distribution 𝐚\mathbf{a}, it is possible to define a unique mapping from pdfs to opinions. Let pp be a Dirichlet pdf with distribution 𝙳𝚒𝚛(𝜶)\mathtt{Dir}\left(\boldsymbol{\alpha}\right) with αi>1\alpha_{i}>1; pp can be mapped to a multinomial opinion ω=(𝐛,u,𝐚)\omega=\left(\mathbf{b},u,\mathbf{a}\right), where the parameters are computed as:

{𝐛=𝜶W𝐚W+i(𝜶iW𝐚i)=𝜶W𝐚i𝜶i𝐮=WW+i(𝜶iW𝐚i)=Wi𝜶i𝐚=𝐚\begin{cases}\mathbf{b}=\frac{\boldsymbol{\alpha}-W\mathbf{a}}{W+\sum_{i}\left(\boldsymbol{\alpha}_{i}-W\mathbf{a}_{i}\right)}=\frac{\boldsymbol{\alpha}-W\mathbf{a}}{\sum_{i}\boldsymbol{\alpha}_{i}}\\ \mathbf{u}=\frac{W}{W+\sum_{i}\left(\boldsymbol{\alpha}_{i}-W\mathbf{a}_{i}\right)}=\frac{W}{\sum_{i}\boldsymbol{\alpha}_{i}}\\ \mathbf{a}=\mathbf{a}\end{cases} (7)

Again, for a binomial opinion, given a mapping constant W0W\in\mathbb{R}_{\geq 0} and a fixed prior distribution aa, it is possible to define a unique mapping from pdfs to opinions. Let pp be a Beta pdf with distribution 𝙱𝚎𝚝𝚊(α,β)\mathtt{Beta}\left(\alpha,\beta\right) with α,β>1\alpha,\beta>1; pp can be mapped to a binomial opinion ω=(b,d,u,a)\omega=\left(b,d,u,a\right), where the parameters are computed as:

{b=αWaα+βd=βW(1a)α+βu=Wα+βa=a\begin{cases}b=\frac{\alpha-Wa}{\alpha+\beta}\\ d=\frac{\beta-W(1-a)}{\alpha+\beta}\\ u=\frac{W}{\alpha+\beta}\\ a=a\end{cases} (8)

For reasons of consistency, WW is usually fixed to 22 [4]. Given a prior distribution aa, this generates the mapping tt from pdf pp to opinion ω\omega:

t:pω.t:p\mapsto\omega.
Subjective opinion operators

SL defines several operators over subjective opinions, such as addition, product or fusion [4]. In general, these operators are computed over the parameters of subjective opinions. Let ωX=(𝐛X,uX,𝐚X)\omega_{X}=\left(\mathbf{b}_{X},u_{X},\mathbf{a}_{X}\right) and ωY=(𝐛Y,uY,𝐚Y)\omega_{Y}=\left(\mathbf{b}_{Y},u_{Y},\mathbf{a}_{Y}\right) be two subjective opinions and let SL:𝒮×𝒮𝒮\circ_{SL}:\mathcal{S}\times\mathcal{S}\rightarrow\mathcal{S} be a generic operator over the space of subjective opinions 𝒮\mathcal{S}. Then, ωZ=(𝐛Z,uZ,𝐚Z)\omega_{Z}=\left(\mathbf{b}_{Z},u_{Z},\mathbf{a}_{Z}\right) resulting from the application of the operator to ωX\omega_{X} and ωY\omega_{Y} is given as:

ωZ=ωXSLωY={𝐛Z=fb(ωX,ωY)uZ=fu(ωX,ωY)𝐚Z=fa(ωX,ωY),\omega_{Z}=\omega_{X}\circ_{SL}\omega_{Y}=\begin{cases}\mathbf{b}_{Z}=f_{b}\left(\omega_{X},\omega_{Y}\right)\\ u_{Z}=f_{u}\left(\omega_{X},\omega_{Y}\right)\\ \mathbf{a}_{Z}=f_{a}\left(\omega_{X},\omega_{Y}\right),\end{cases} (9)

where fb,fu,fa:𝒮×𝒮0f_{b},f_{u},f_{a}:\mathcal{S}\times\mathcal{S}\rightarrow\mathbb{R}_{\geq 0} are operator-specific functions returning the values of belief, uncertainty and prior for the opinion ωZ\omega_{Z}.

Subjective opinion operators for evaluating operations over pdfs

When properly defined, SL operators can be used to approximate operations over probability distribution functions. Suppose we are given two pdfs, pXp_{X} and pYp_{Y}, and we want to compute a generic operation over them, P:𝒫×𝒫𝒫\circ_{P}:\mathcal{P}\times\mathcal{P}\rightarrow\mathcal{P} over the space of probability distributions 𝒫\mathcal{P}. Computing this operation over probability distributions may be very complex. However, if we have an SL operator SL:𝒮×𝒮𝒮\circ_{SL}:\mathcal{S}\times\mathcal{S}\rightarrow\mathcal{S} that approximates P\circ_{P}, we may find a workaround computing pXPpYp_{X}\circ_{P}p_{Y} by projecting the two distribution onto the opinions ωX\omega_{X} and ωY\omega_{Y}, computing the resulting opinion ωZ=ωXSLωY\omega_{Z}=\omega_{X}\circ_{SL}\omega_{Y}, and then mapping the result back onto a probability distribution function pZSLp_{Z}^{SL}. In this way, the resulting pdf pZSLp_{Z}^{SL} provides an easy-to-compute approximation of the real pdf pZp_{Z} (see Figure 1).

(pX,pY){\left(p_{X},p_{Y}\right)}pZ{p_{Z}}(ωX,ωY){\left(\omega_{X},\omega_{Y}\right)}ωZ{\omega_{Z}}pZSL{p_{Z}^{SL}}P\scriptstyle{\circ_{P}}t()\scriptstyle{t()}SL\scriptstyle{\circ_{SL}}s()\scriptstyle{s()}
Figure 1: If the application of the operator P\circ_{P} to two pdfs pXp_{X} and pYp_{Y} can not be solved analytically, we can map pXp_{X} and pYp_{Y} to the opinions ωX\omega_{X} and ωY\omega_{Y} and apply the SL operator SL\circ_{SL} to compute the opinion ωZ\omega_{Z}. The pdf pZSLp_{Z}^{SL} associated with ωZ\omega_{Z} provides an approximation of pZp_{Z}.

3 Computational Statistics

In this section, we review some elements of computational statistics that are relevant to our work. We describe how sampling is used in MC simulations; we show how unbiased estimators can be built via MC integration; we discuss how unbiased estimators can be used to build moment-matching approximation; we show how pdfs may be reconstructed through kernel density estimation; and, finally, we bring these parts together to show how MC simulations may be used to compute the product of pdfs via moment-matching or kernel-density estimation.

Monte Carlo sampling

MC simulations are stochastic numerical algorithms designed to find approximate solutions through repeated random sampling. This paradigm has been applied in many areas of research to solve problems whose exact analytical solution is impossible or too difficult to derive. In statistics, MC simulations are widely used to evaluate probability distributions whose analytical form can not be explicitly expressed. Let XX be a random variable with a probability distribution pXp_{X} on the support Ω\Omega; let us also assume that the analytical form of pXp_{X} is unknown but that we can sample realizations xix_{i} of the random variable XX; then, MC simulations allow us to draw a large number of independent samples xix_{i} and use them to (i) compute useful empirical statistical descriptors S^X\hat{S}_{X} of the probability distribution pXp_{X}, or, eventually, (ii) reconstruct the approximate shape of the probability distribution pXp_{X}.

Monte Carlo integration

In order to compute useful empirical statistical descriptors SXS_{X} of the probability distribution pXp_{X}, MC simulations rely on integration and on the law of large numbers. Let SXS_{X} be a statistics of the probability distribution pXp_{X} that can be computed from a function f()f\left(\cdot\right) applied to the samples xix_{i}. The statistics SXS_{X} is then defined as:

SX=Ωf(x)pX(x)𝑑x.S_{X}=\int_{\Omega}f(x)p_{X}(x)dx. (10)

By the law of large numbers, an estimator of SXS_{X} can be computed using NN samples of xix_{i} as:

S^X=1Ni=1Nf(xi).\hat{S}_{X}=\frac{1}{N}\sum_{i=1}^{N}f\left(x_{i}\right). (11)

It is immediate to see that using Equation 11 and choosing an appropriate function f()f(\cdot) we can directly estimate useful statistics of the distribution pXp_{X}, such as moments and quantiles. Thus, through a MC simulation we can sample points from pXp_{X} and compute informative estimator statistics S^X\hat{S}_{X}.

Moment-matching approximation

A probability distribution pXp_{X} is completely characterized by the collection of all its moments; if we know the parametric form of the function pXp_{X} from which we are sampling from, but we ignore the exact value of its parameters, we can compute an estimate p^X\hat{p}_{X} by setting the moments to the estimated values M^i[X]\hat{M}_{i}\left[X\right]. In several scenarios of interest it may actually be possible to compute analytically the value of few lower moments Mi[X]M_{i}\left[X\right] of interest (such as, mean and variance); this approach is well-known and it has been used in the study of SL operators as well (see, for instance, [7]). In general, though, MC simulation and integration provide an empirical and robust way to compute estimators of the ii-th moments M^i[X]\hat{M}_{i}\left[X\right] of a probability distribution pXp_{X}, even when no exact analytical formula for computing the moments Mi[X]M_{i}\left[X\right] of interest is available.

Kernel density estimation

Beyond computing statistics, it is possible to use samples xix_{i} generated in a MC simulation to reconstruct the actual probability distribution pXp_{X}. A standard approach to reconstruct a continuous function pXp_{X} from a set of finite points xix_{i} is kernel density estimation (KDE). Any function may be expressed as a convolution with a kernel function κ()\kappa(\cdot):

pX=Ωκ(x)𝑑x.p_{X}=\int_{\Omega}\kappa(x)dx. (12)

Practically, it is possible to get an empirical approximation using only a finite set of points xix_{i}:

p^X(x)=1Nwi=1Nκ(xxiw),\hat{p}_{X}(x)=\frac{1}{Nw}\sum_{i=1}^{N}\kappa\left(\frac{x-x_{i}}{w}\right), (13)

where the kernel κ()\kappa(\cdot) is a symmetric function, like a triangular function or a Gaussian, and ww denotes the width of the kernel; empirical rules are available to select an optimal value for this parameter in relation to the number of samples available [11]. Thus, using the same MC simulation procedure to sample points from pXp_{X} it is possible also to estimate an approximate probability distribution p^X\hat{p}_{X}.

Monte Carlo simulation for evaluating operations over pdfs

Suppose we are given two probability distributions, pXp_{X} and pYp_{Y}, and suppose we want to compute the distribution pZp_{Z} determined by the application of operation P:𝒫×𝒫𝒫\circ_{P}:\mathcal{P}\times\mathcal{P}\rightarrow\mathcal{P}, that is, pZ=pXPpYp_{Z}=p_{X}\circ_{P}p_{Y}. If the pdf pZp_{Z} can not be computed analytically, MC simulations may be used to sample from pZp_{Z} and to estimate a pdf that approximates pZp_{Z}. As a first solution, we could rely on the samples {z1,z2zN}\{z_{1},z_{2}\dots z_{N}\} obtained by sampling from pXp_{X} and pYp_{Y} to estimate the moments M^i[Z]\hat{M}_{i}\left[Z\right] and then instantiate a moment-matching approximation p^ZMM\hat{p}_{Z}^{\mathrm{MM}} (see Figure 2). Alternatively, we could use the same samples {z1,z2zN}\{z_{1},z_{2}\dots z_{N}\} from pZp_{Z} to perform a kernel-density estimation and compute the KDE approximation p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}} (see Figure 3). Notice that, differently from the SL approximation pZSLp_{Z}^{SL}, we decorate the approximations computed via MC simulations p^ZMM\hat{p}_{Z}^{\mathrm{MM}} and p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}} with a hat to underline that they are empirical statistics.

(pX,pY){\left(p_{X},p_{Y}\right)}pZ{p_{Z}}{z1,z2zN}{\{z_{1},z_{2}\dots z_{N}\}}M^i[Z]{\hat{M}_{i}\left[Z\right]}p^ZMM{\hat{p}_{Z}^{\mathrm{MM}}}P\scriptstyle{\circ_{P}}MCS\scriptstyle{MCS}MCI\scriptstyle{MCI}
Figure 2: Moment-matching approximation via MC simulation. If the product of two pdfs pXp_{X} and pYp_{Y} can not be solved analytically, we can sample, integrate and estimate the pdf p^ZMM\hat{p}_{Z}^{\mathrm{MM}}, which provides an approximation of pZp_{Z}. MCS stands for MC sampling, MCI stands for MC integration.
(pX,pY){\left(p_{X},p_{Y}\right)}pZ{p_{Z}}{z1,z2zN}{\{z_{1},z_{2}\dots z_{N}\}}p^ZKDE{\hat{p}_{Z}^{\mathrm{KDE}}}P\scriptstyle{\circ_{P}}MCS\scriptstyle{MCS}KDE\scriptstyle{KDE}
Figure 3: Kernel-density approximation via MC simulation. If the product of two pdfs pXp_{X} and pYp_{Y} can not be solved analytically, we can sample and estimate the pdf p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}}, which provides an approximation of pZp_{Z}. MCS stands for MC sampling, KDE stands for kernel-density estimation.

4 Computational Complexity

In this section, we discuss and compare the computational complexity of SL operators and MC simulations. We will evaluate the computational complexity using the 𝒪()\mathcal{O}\left(\cdot\right) notation as the time complexity of running a given algorithm as a function of its input.

Subjective logic

SL operators are defined to be extremely efficient. Indeed, given two opinions ωX\omega_{X} and ωY\omega_{Y} and the generic operator SL:𝒮×𝒮𝒮\circ_{SL}:\mathcal{S}\times\mathcal{S}\rightarrow\mathcal{S}, the computation of ωZ=ωXSLωY\omega_{Z}=\omega_{X}\circ_{SL}\omega_{Y} usually requires only a limited number of function evaluations, as shown in Equation 9. The number of evaluations is 2M+12\cdot M+1, where MM is the number of events over which the opinions are defined. Thus, the overall complexity is 𝒪(M)\mathcal{O}\left(M\right): it depends only on the number of events considered, and it is independent of the actual form of the mapped distributions. This makes SL operators an attractive choice especially when working in lower dimensions.

Monte Carlo simulation

The MC approach is, by definition, computationally intensive. The computational complexity of a MC simulation scales as a function of the number NN of samples that must be produced. Each iteration requires random sampling and the execution of all the operations necessary to sample from pZp_{Z}. Overall, the computational complexity of the MC simulation is 𝒪(N)\mathcal{O}\left(N\right). If we are using MC simulations to estimate a moment-matching approximation p^ZMM\hat{p}_{Z}^{\mathrm{MM}}, MC integration allows us to compute statistics from the samples generated during the MC simulation with no additional overall computational complexity. However, if we want to estimate the actual pdf via KDE we have to take into account an increase in the overall computational complexity from the linear order to the quadratic order 𝒪(N2)\mathcal{O}\left(N^{2}\right). Computing the pdf p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}} is then a significantly computationally expensive procedure.

It is evident that, taking into account computational complexity only, SL operators dominate MC simulations, with or without KDE, especially considering that the number NN of samples in a MC simulation is required to grow large in order to return reliable results even in low dimensions.

5 Bias

In this section, we start analyzing the degree of approximation of SL operators and MC simulations. We will evaluate the degree of approximation in terms of bias of the estimator p^Z\hat{p}_{Z}, that is, as the expected value of the difference between the true distribution and the estimated approximation: E[pZp^Z]E\left[p_{Z}-\hat{p}_{Z}\right].

Subjective logic

The bias of SL operators is dependent on the definition of the specific operator, and a generic theoretical treatment is not possible. Moreover, an analytic study of the bias is not always available for all possible SL operators. In Section 7 we will consider the case study of the binomial operator for subjective logic and we will analyze more in detail its specific bias.

Monte Carlo simulation

MC simulations are known to provide asymptotically unbiased estimators. If we estimate a statistics SXS_{X} of the pdf pXp_{X} using a MC integration as in Equation 11, then S^X\hat{S}_{X} is an asymptotically unbiased estimator, that is, in the limit of infinite samples, it converges to the true quantity it approximates:

limNS^X(N)=SX,\lim_{N\rightarrow\infty}\hat{S}_{X}(N)=S_{X}, (14)

where we made explicit the dependence of S^X\hat{S}_{X} on the number of samples NN.

If we use MC integration to estimate the moments M^[Z]\hat{M}\left[Z\right] for a moment-matching approximation p^ZMM\hat{p}_{Z}^{\mathrm{MM}}, the MC simulation provides us with unbiased estimators of the moments; this means that, by increasing the number of samples generated in a MC simulation, we can get arbitrarily close to the true value of the estimated quantity. However, notice that while the estimated moments M^[Z]\hat{M}\left[Z\right] are asymptotically unbiased, the p^ZMM\hat{p}_{Z}^{\mathrm{MM}} is biased; this bias is due to the limited set of moments M^[Z]\hat{M}\left[Z\right] used to approximate pZp_{Z}.

If we use a MC simulation to estimate the true pdf directly via KDE, the empirical pdf p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}} is biased. In this case, it is known that the width parameter ww of KDE regulates the trade-off between bias and variance. In general, the bias can be shown to be proportional to the width ww of the kernel κ()\kappa(\cdot):

EKDE[pZp^ZKDE]w2,E_{KDE}\left[p_{Z}-\hat{p}_{Z}^{\mathrm{KDE}}\right]\propto w^{2}, (15)

under the constraint that ww can not be reduced to zero, for statistical and computational reasons [11]. When using a Gaussian kernel, the widely-adopted Silverman rule suggests the adoption of a kernel width of the following size:

w=1.06σ^1N5,w=1.06\hat{\sigma}\frac{1}{\sqrt[5]{N}}, (16)

where σ^\hat{\sigma} is the empirical standard deviation computed from the samples:

σ^=i=1N(xiμ^)2N1,\hat{\sigma}=\sqrt{\frac{\sum_{i=1}^{N}\left(x_{i}-\hat{\mu}\right)^{2}}{N-1}}, (17)

where μ^\hat{\mu} is the empirical mean. It follows, then, that the bias of the KDE approximation is proportional to:

EKDE[pZp^ZKDE]1.062σ^2N25.E_{KDE}\left[p_{Z}-\hat{p}_{Z}^{\mathrm{KDE}}\right]\propto 1.06^{2}\hat{\sigma}^{2}N^{\frac{2}{5}}. (18)

As said, this bias can never be reduced to zero. However, in specific computational setting, this bias may be bounded by finding an optimal trade-off between the number of samples NN and the empirical standard deviation σ^\hat{\sigma}. In particular, if the domain of pZp_{Z} is a discrete domain, as in the case of multinomial opinions and Dirichlet pdfs which underlie subjective opinions, then the empirical standard deviation σ^\hat{\sigma} may be bounded and it may be possible to estimate the magnitude of the bias as a function of the number of samples NN.

In summary, from the point of view of approximation, MC simulations represents a safer choice than SL operators, as they are grounded in solid theory and they allow us to quantify and to control the bias. The lack of any bound for SL operators may be seen as an obstacle in adopting them when working in critical domains where precise approximations are required. In the next section, we will introduce our protocol to solve this problem and estimate the degree of approximation of SL operators.

6 Computational Evaluation of the Degree of Approximation of Subjective Logic Operators

In this section we present a framework to evaluate the bias of an SL operator. We start by discussing how MC approximations may be related to SL approximation using a distance measure; then, we define what precise distance measure we will use and how it relates to bias.

Relating subjective logic approximation and Monte Carlo approximations via a distance measure

In the previous sections we illustrated two methodologies for finding an approximation of the pdf pZp_{Z}, one based on SL operators (pZSLp_{Z}^{SL}) and one relying on MC simulations (p^ZKDE,p^ZMM\hat{p}_{Z}^{\mathrm{KDE}},\hat{p}_{Z}^{\mathrm{MM}}). Figure 4 merges the graphs in Figure 1, 2 and 3 to illustrate the alternative computational paths that are offered to compute an approximation of pZp_{Z}; starting from the distributions (pX,pY)\left(p_{X},p_{Y}\right), the upper path represents the SL approach to finding an approximation of pZ=pXPpYp_{Z}=p_{X}\circ_{P}p_{Y}, while the lower paths represent MC approaches to finding an approximation of the same quantity pZp_{Z}.

(ωX,ωY){\left(\omega_{X},\omega_{Y}\right)}ωZ{\omega_{Z}}pZSL{p_{Z}^{SL}}(pX,pY){\left(p_{X},p_{Y}\right)}pZ{p_{Z}}{z1,z2zN}{\{z_{1},z_{2}\dots z_{N}\}}p^ZKDE{\hat{p}_{Z}^{\mathrm{KDE}}}M^i[Z]{\hat{M}_{i}\left[Z\right]}p^ZMM{\hat{p}_{Z}^{\mathrm{MM}}}SL\scriptstyle{\circ_{SL}}s()\scriptstyle{s()}P\scriptstyle{\circ_{P}}t()\scriptstyle{t()}MCS\scriptstyle{MCS}MCI\scriptstyle{MCI}KDE\scriptstyle{KDE}
Figure 4: Approximations of pZp_{Z}. MCS stands for MC sampling, MCI stands for MC integration, KDE stands for kernel-density estimation.

Now, approximate methods trade off precision in the results for simplicity in computation. In order to make a grounded decision on which approximation path in Figure 4 to use, it is necessary to quantify the trade-off between computational complexity and bias. As discussed in Section 4 and 5, in the case of KDE approximation via MC simulations, both complexity and bias are known. However, in the case of SL operators, we may easily derive their computational complexity, but we have no simple way of evaluating their bias. Exploiting the properties of MC integration and the idea of distance between pdfs, it is possible to assess the degree of approximation of SL operators in a computational fashion by relating them to MC simulations.

A simple way to evaluate how well a pdf pp approximates another pdf qq is to estimate the distance between them, D[p,q]D\left[p,q\right], where D[,]D\left[\cdot,\cdot\right] is a measure of distance or divergence between pdfs [8]. The degree of approximation of pZSLp_{Z}^{SL} could then be obtained by measuring the distance from the true pdf pZp_{Z}:

D[pZ,pZSL].D\left[p_{Z},p_{Z}^{SL}\right]. (19)

However, since the true pdf pZp_{Z} is taken to be unknown or hard to compute, it is challenging to get a direct estimate of these quantities. Since we can not rely directly on pZp_{Z}, we can instead exploit MC simulations and its properties.

From Equation 15 in Section 5, we know that the KDE estimation p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}} is biased and we know how to evaluate it. Moreover, from Equation 18 in Section 5, we see that this bias depends on the number of samples NN and the standard deviation σ^\hat{\sigma}. Now, if the domain of pZp_{Z} is a discrete domain, as in the case of multinomial opinions and Dirichlet pdfs, then the empirical standard deviation σ^\hat{\sigma} may be bounded and it may be possible to estimate the magnitude of the bias as a function of the number of samples NN. It may be possible to select a number of samples NN that shrinks the bias to a negligible quantity; in such case, we can then accept the KDE estimation p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}} as a close approximation of the true pdf pZp_{Z}:

D[pZ,p^ZKDE(N)]0.D\left[p_{Z},\hat{p}_{Z}^{\mathrm{KDE}}(N)\right]\approx 0. (20)

We underline that this approximation holds only under the assumption that, for an increasing number of samples NN, the bias of the estimate p^ZKDE(N)\hat{p}_{Z}^{\mathrm{KDE}}(N) tends, if not to zero, to a quantity whose order of magnitude is negligible with respect to further analysis; in other words, the validity of the approximation in Equation 20 is conditional on the pdf pZp_{Z} we are considering, the analysis we will be carrying out, and the number of samples we can produce (for an example of an evaluation of these conditions, see the application to the case study of the product of Beta pdfs in Section 7 and Section 7.1).

The approximation in Equation 20 is extremely useful because it means that while we can not evaluate absolute distances with respect to the true distribution pZp_{Z}, we can still evaluate the relative distance between the KDE approximation and the SL approximation, and use it as a proxy for the distance between the SL approximation pZSLp_{Z}^{SL} and the true distribution pZp_{Z}:

D[p^ZKDE(N),pZSL]D[pZ,pZSL].D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),p_{Z}^{SL}\right]\approx D\left[p_{Z},p_{Z}^{SL}\right]. (21)

Thus, given only a finite set of samples NN we can obtain an empirical statistic of the distance as:

D^[pZ,pZSL]=^D[p^ZKDE(N),pZSL],\hat{D}\left[p_{Z},p_{Z}^{SL}\right]\hat{=}D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),p_{Z}^{SL}\right], (22)

If the condition in Equation 20 holds, we expect the distance D[p^ZKDE(N),pZSL]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),p_{Z}^{SL}\right] to be orders of magnitudes greater than D[pZ,p^ZKDE(N)]D\left[p_{Z},\hat{p}_{Z}^{\mathrm{KDE}}(N)\right]; this would indeed confirm that the bias of p^ZKDE(N)\hat{p}_{Z}^{\mathrm{KDE}}(N) is negligible and that the computation of D^[pZ,pZSL]\hat{D}\left[p_{Z},p_{Z}^{SL}\right] (using a finite number of samples) provides a good estimate of the degree of approximation of the SL approximation.

Relating distance measure to bias

So far, we have discussed distance measures in abstract terms. The quantity D^[pZ,pZSL]\hat{D}\left[p_{Z},p_{Z}^{SL}\right] may indeed be computed using different pdf distance, such as ϕ\phi-divergences or integral probability metrics [13].

In this paper, we will rely on computing a simple integral distance, defined as:

DI[p,q]=+|p(x)q(x)|𝑑x.D_{I}\left[p,q\right]=\int_{-\infty}^{+\infty}\left|p(x)-q(x)\right|dx. (23)

This distance DI[p,q]D_{I}\left[p,q\right] is the same as the total variation distance except for the scaling constant:

DTV[p,q]=12+|p(x)q(x)|𝑑x.D_{TV}\left[p,q\right]=\frac{1}{2}\int_{-\infty}^{+\infty}\left|p(x)-q(x)\right|dx. (24)

The constant 12\frac{1}{2} rescales the distance on the interval [0,1][0,1]. However, in order to get an absolute evaluation of how the mass of the two distributions pp and qq overlaps, we drop the scaling constant.

The choice of an integral distance DI[,]D_{I}\left[\cdot,\cdot\right] is justified for three reasons. First, from a conceptual point of view, an integral distance allows us to get a complete picture of the difference between two pdfs. While measures based on the evaluation of a limited set of synthetic statistics such as moments would provide us with a rough evaluation of the difference between two distributions, an integral distance provides a more precise way to assess the distribution of the mass of probability, taking into account, for instance, the potential presence of multiple modes or how mass subtly distributes on the tails.

Second, from a computational point of view, the integral distance DI[,]D_{I}\left[\cdot,\cdot\right] allows us, once again, to exploit MC integration. Recall that we want to get an estimation of D^[pZ,pZSL]\hat{D}\left[p_{Z},p_{Z}^{SL}\right] through the approximation D[p^ZKDE(N),pZSL]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),p_{Z}^{SL}\right]. Now, if we reconstruct p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}} via KDE, we can estimate the integral distance via MC integration over the domain Ω\Omega of the events as:

Ω|p^ZKDE(z)pZSL(z)|𝑑z=^1Ni=1N|p^ZKDE(zi)pZSL(zi)|.\int_{\Omega}\left|\hat{p}_{Z}^{\mathrm{KDE}}(z)-p_{Z}^{SL}(z)\right|dz\hat{=}\frac{1}{N}\sum_{i=1}^{N}\left|\hat{p}_{Z}^{\mathrm{KDE}}(z_{i})-p_{Z}^{SL}(z_{i})\right|. (25)

Third, from a theoretical point of view, the integral in Equation 25 is related to the bias:

Ω|p^ZKDE(z)pZSL(z)|𝑑z\displaystyle\int_{\Omega}\left|\hat{p}_{Z}^{\mathrm{KDE}}(z)-p_{Z}^{SL}(z)\right|dz =^1Ni=1N|p^ZKDE(zi)pZSL(zi)|\displaystyle\hat{=}\frac{1}{N}\sum_{i=1}^{N}\left|\hat{p}_{Z}^{\mathrm{KDE}}(z_{i})-p_{Z}^{SL}(z_{i})\right| (26)
=^E[|p^ZKDE(Z)pZSL(Z)|]\displaystyle\hat{=}E\left[\left|\hat{p}_{Z}^{\mathrm{KDE}}(Z)-p_{Z}^{SL}(Z)\right|\right] (27)
E[p^ZKDE(Z)pZSL(Z)].\displaystyle\geq E\left[\hat{p}_{Z}^{\mathrm{KDE}}(Z)-p_{Z}^{SL}(Z)\right]. (28)

Thus, using the integral distance DI[p^ZKDE,pZSL]D_{I}\left[\hat{p}_{Z}^{\mathrm{KDE}},p_{Z}^{SL}\right] we can obtain an estimation of the distance D^[pZ,pZSL]\hat{D}\left[p_{Z},p_{Z}^{SL}\right] as well as an upper bound on the bias of pZSLp_{Z}^{SL}. Notice that the absolute value in the integral distance provides a more honest evaluation of the absolute difference between pdfs, avoiding an averaging effect in absence of the absolute value operator.

Figure 5 summarizes our overall framework to evaluate the degree of approximation of the SL approximation as the integral distance |pZSLp^ZKDE|\int\left|p_{Z}^{SL}-\hat{p}_{Z}^{\mathrm{KDE}}\right|, under the assumption that D[pZ,p^ZKDE(N)]0D\left[p_{Z},\hat{p}_{Z}^{\mathrm{KDE}}(N)\right]\approx 0. This approach is generic and it is not tied to the SL approximation. If the condition in Equation 20 can be guaranteed, the same approach may be used to get an estimation of the distance between the true pdf pZp_{Z} and other potential approximation. For instance, Figure 5 shows our methodology applied also to the problem of estimating the distance from the true pdf of the moment-matching approximation D[pZ,p^ZMM(N)]D\left[p_{Z},\hat{p}_{Z}^{\mathrm{MM}}(N)\right] by computing the distance |p^ZMMp^ZKDE|\int\left|\hat{p}_{Z}^{\mathrm{MM}}-\hat{p}_{Z}^{\mathrm{KDE}}\right|.

(ωX,ωY){\left(\omega_{X},\omega_{Y}\right)}ωZ{\omega_{Z}}pZSL{p_{Z}^{SL}}|pZSLp^ZKDE|{\int\left|p_{Z}^{SL}-\hat{p}_{Z}^{\mathrm{KDE}}\right|}D^I[pZSL,pZ]{\hat{D}_{I}\left[p_{Z}^{SL},p_{Z}\right]}(pX,pY){\left(p_{X},p_{Y}\right)}pZ{p_{Z}}{z1,z2zN}{\{z_{1},z_{2}\dots z_{N}\}}p^ZKDE{\hat{p}_{Z}^{\mathrm{KDE}}}M^i[Z]{\hat{M}_{i}\left[Z\right]}p^ZMM{\hat{p}_{Z}^{\mathrm{MM}}}|p^ZMMp^ZKDE|{\int\left|\hat{p}_{Z}^{\mathrm{MM}}-\hat{p}_{Z}^{\mathrm{KDE}}\right|}D^I[p^ZMM,pZ]{\hat{D}_{I}\left[\hat{p}_{Z}^{\mathrm{MM}},p_{Z}\right]}SL\scriptstyle{\circ_{SL}}s()\scriptstyle{s()}MCI\scriptstyle{MCI}P\scriptstyle{\circ_{P}}t()\scriptstyle{t()}MCS\scriptstyle{MCS}MCI\scriptstyle{MCI}KDE\scriptstyle{KDE}MCI\scriptstyle{MCI}
Figure 5: Evaluations of the distances between pZp_{Z} and its approximations based on the assumption that D[pZ,p^ZKDE(N)]0D\left[p_{Z},\hat{p}_{Z}^{\mathrm{KDE}}(N)\right]\approx 0. MCS stands for MC sampling, MCI stands for MC integration, KDE stands for kernel-density estimation.

7 Case Study: Product of Beta Distributions

In this section, we show how our framework may be applied to the problem of computing the product of Beta distributions. We first recall the definition of a Beta distribution and the definition of the product of Beta distributions; we then introduce the SL operator for binomial multiplication and we discuss how it can be used for approximating the distribution of the random variable given by the product of two independent random variables with Beta distributions; we work out the computational complexity of binomial multiplication and show the lack of generic estimate of its degree of approximation; to solve this problem, we apply our framework to get an evaluation of the degree of approximation of binomial multiplication; finally, we run an extensive set of empirical simulations to validate our theoretical results.

Beta pdf

Let XX be a random variable on the support [0,1][0,1]; we say that XX follows a Beta distribution X𝙱𝚎𝚝𝚊(α,β)X\sim\mathtt{Beta}\left(\alpha,\beta\right) with parameters α0\alpha\in\mathbb{R}_{\geq 0} and β0\beta\in\mathbb{R}_{\geq 0} when its probability density function pXp_{X} has the following form:

pX(x;α,β)=1B(α,β)xα1(1x)β1,p_{X}\left(x;\alpha,\beta\right)=\frac{1}{B\left(\alpha,\beta\right)}x^{\alpha-1}\left(1-x\right)^{\beta-1}, (29)

where B(α,β)B\left(\alpha,\beta\right) is the Beta function.

Product of Beta pdfs

Let X𝙱𝚎𝚝𝚊(αX,βX)X\sim\mathtt{Beta}\left(\alpha_{X},\beta_{X}\right) and Y𝙱𝚎𝚝𝚊(αY,βY)Y\sim\mathtt{Beta}\left(\alpha_{Y},\beta_{Y}\right) be two independent Beta random variables with associated pdfs pXp_{X} and pYp_{Y}. Let us define a third random variable ZZ as the product of the two Beta random variables Z=XYZ=X\cdot Y. The probability density function pZp_{Z} of ZZ does not follow a Beta distribution anymore, and its precise analytical form can not be easily expressed using elementary functions [9]. An analytical solution to the evaluation of the pdf of the product of two Beta distributions has been offered in [10]222This paper actually presents the more generic solution to the problem of multiplying two general Beta distribution, which subsume the multiplication of two simple Beta distributions as defined above.:

pZ(z;αX,αY,βX,βY)=B(βX,βY)zβX(1z)βX+βY1zαYFD(3)(βX;1αX,1αY,αX+βX1;βX+βY;0,z1z,z1z)B(αX,βX)B(αY,βY),\begin{array}[]{c}p_{Z}\left(z;\alpha_{X},\alpha_{Y},\beta_{X},\beta_{Y}\right)=\\ B\left(\beta_{X},\beta_{Y}\right)\cdot z^{-\beta_{X}}\cdot\left(1-z\right)^{\beta_{X}+\beta_{Y}-1}\cdot z^{\alpha_{Y}}\cdot\\ \cdot\frac{F_{D}^{(3)}\left(\beta_{X};1-\alpha_{X},1-\alpha_{Y},\alpha_{X}+\beta_{X}-1;\beta_{X}+\beta_{Y};0,\frac{z-1}{z},\frac{z-1}{z}\right)}{B\left(\alpha_{X},\beta_{X}\right)B\left(\alpha_{Y},\beta_{Y}\right)},\end{array} (30)

where FD(3)F_{D}^{(3)} is the Lauricella D hyper-geometric series. While this formula provides an elegant solution to the problem of finding the pdf of the product of two Beta pdfs, its straightforward evaluation is challenging as the Lauricella function requires the computation of factorial products and series.

Other analytical approaches to evaluate the product of two or more Beta distributions have been proposed, including methods relying on high-order functions, such as the Meijer G-function or Fox’s H function, or modeling the pdf of the product using an infinite mixture of simpler distributions [14, 1]. These approaches also present computational challenges, despite more efficient solutions have been investigated [14, 1].

Finally, common approaches rely on MC simulations to sample points from the probability distribution of ZZ and to compute statistics of the pdf by matching the moments or the quantiles of ZZ [9], as we reviewed in Section 3. Notice that when considering the product of two independent random variables Z=XYZ=X\cdot Y, it is straightforward to compute the mean E[Z]E\left[Z\right] and the variance Var[Z]Var\left[Z\right] of pZp_{Z} analytically as:

E[Z]=E[X]E[Y]Var[Z]=E[X]2Var[Y]+E[Y]2Var[X]+Var[X]Var[Y].\begin{array}[]{c}E\left[Z\right]=E\left[X\right]\cdot E\left[Y\right]\\ Var\left[Z\right]=E\left[X\right]^{2}\cdot Var\left[Y\right]+E\left[Y\right]^{2}\cdot Var\left[X\right]+Var\left[X\right]\cdot Var\left[Y\right].\end{array} (31)

Thus, if we were to perform a moment-matching approximation considering only the first two moments, we could instantiate such an approximation without running any MC simulation with a constant computational complexity of 𝒪(1)\mathcal{O}\left(1\right).

Subjective logic binomial multiplication

An alternative solution to compute the product of Beta distributions is based on the use of the SL operator for binomial multiplication. Given two binomial opinions ωX\omega_{X} and ωY\omega_{Y} defined on different domains, the binomial opinion ωZ\omega_{Z} resulting from the multiplication ωXωY\omega_{X}\cdot\omega_{Y} is computed as [4]:

ωZ={bZ=bXbY+(1aX)aYbXuY+aX(1aY)uXbY1aXaYdZ=dX+dYdXdYuZ=uXuY+(1aY)bXuY+(1aX)uXbY1aXaYaZ=aXaY.\omega_{Z}=\begin{cases}b_{Z}=b_{X}b_{Y}+\frac{\left(1-a_{X}\right)a_{Y}b_{X}u_{Y}+a_{X}\left(1-a_{Y}\right)u_{X}b_{Y}}{1-a_{X}a_{Y}}\\ d_{Z}=d_{X}+d_{Y}-d_{X}d_{Y}\\ u_{Z}=u_{X}u_{Y}+\frac{\left(1-a_{Y}\right)b_{X}u_{Y}+(1-a_{X})u_{X}b_{Y}}{1-a_{X}a_{Y}}\\ a_{Z}=a_{X}a_{Y}.\end{cases} (32)

Practically, a binomial product operator allows us to evaluate the combination of two opinions over two different facts. In the domain of probability distributions, the multiplication of opinions ωZ=ωXωY\omega_{Z}=\omega_{X}\cdot\omega_{Y} translates into the multiplication of the mapped pdfs pZ=pXpYp_{Z}=p_{X}\cdot p_{Y}.

Approximating the product of Beta pdfs

Now, assume we are interested in computing the product Z=XYZ=X\cdot Y, where XX and YY are two independent Beta random variables. Since an analytic solution is hard to compute, we may decide to rely either on the SL approximation or on a MC approximation.

Concerning moment-matching approximations we may consider a Gaussian pdf and a Beta pdf. Using a Gaussian pdf is a choice motivated by the simplicity and the ubiquity of this distribution; however, this is clearly a naive choice, as a Gaussian pdf has an unbounded support, is symmetrical and it assumes that all the statistical moments greater than the second are zero. Using a Beta distribution is a more prudent choice: even if it is known that the product of two Betas is not, in general, a Beta distribution, a Beta pdf still fits the right support and it may have other moments different from zero. In order to evaluate the parameters of our Gaussian and Beta approximation, we may rely on MC simulations or on an analytic evaluation. If we opt for MC simulations, we can use MC integration to estimate the mean μ^\hat{\mu} and the variance σ^2\hat{\sigma}^{2} of pZp_{Z}; the Gaussian approximation p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}} is then instantiated as 𝙽(μ^,σ^2)\mathtt{N}\left(\hat{\mu},\hat{\sigma}^{2}\right), while the Beta approximation p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}} is defined as 𝙱𝚎𝚝𝚊(μ^(σ^2+μ^2μ^)σ^2,(μ^1)(σ^2+μ^2μ^)σ^2)\mathtt{Beta}\left(\frac{-\hat{\mu}\left(\hat{\sigma}^{2}+\hat{\mu}^{2}-\hat{\mu}\right)}{\hat{\sigma}^{2}},\frac{\left(\hat{\mu}-1\right)\left(\hat{\sigma}^{2}+\hat{\mu}^{2}-\hat{\mu}\right)}{\hat{\sigma}^{2}}\right), thus guaranteeing that pZp_{Z}, p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}} and p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}} have the same mean and variance. If we rely on an analytic approach, we can easily compute the mean μ\mu and the variance σ2\sigma^{2} of pZp_{Z} using Equation 31; as before, the Gaussian approximation pZGAUSS{p}_{Z}^{\mathrm{GAUSS}} is then instantiated as 𝙽(μ,σ2)\mathtt{N}\left({\mu},{\sigma}^{2}\right), while the Beta approximation pZBETA{p}_{Z}^{\mathrm{BETA}} is defined as 𝙱𝚎𝚝𝚊(μ(σ2+μ2μ)σ2,(μ1)(σ2+μ2μ)σ2).\mathtt{Beta}\left(\frac{-{\mu}\left({\sigma}^{2}+{\mu}^{2}-{\mu}\right)}{{\sigma}^{2}},\frac{\left({\mu}-1\right)\left({\sigma}^{2}+{\mu}^{2}-{\mu}\right)}{{\sigma}^{2}}\right).

Figure 6 provides a concrete instantiation of the diagram in Figure 4, in which the generic operators P\circ_{P} and SL\circ_{SL} have been substituted with multiplication and the generic moment-matching approximation p^ZMM\hat{p}_{Z}^{\mathrm{MM}} has been replaced by the empirical Gaussian approximation p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}}, the empirical Beta approximation p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}}, the analytic Gaussian approximation pZGAUSS{p}_{Z}^{\mathrm{GAUSS}}, and the analytic Beta approximation pZBETA{p}_{Z}^{\mathrm{BETA}}.

(ωX,ωY){\left(\omega_{X},\omega_{Y}\right)}ωZ{\omega_{Z}}pZSL{p_{Z}^{SL}}(pX,pY){\left(p_{X},p_{Y}\right)}pZ{p_{Z}}{z1,z2zN}{\{z_{1},z_{2}\dots z_{N}\}}p^ZKDE{\hat{p}_{Z}^{\mathrm{KDE}}}(μ^,σ^){\left(\hat{\mu},\hat{\sigma}\right)}p^ZGAUSS{\hat{p}_{Z}^{\mathrm{GAUSS}}}p^ZBETA{\hat{p}_{Z}^{\mathrm{BETA}}}(μ,σ){\left({\mu},{\sigma}\right)}pZGAUSS{{p}_{Z}^{\mathrm{GAUSS}}}pZBETA{{p}_{Z}^{\mathrm{BETA}}}\scriptstyle{\cdot}s()\scriptstyle{s()}\scriptstyle{\cdot}AMC\scriptstyle{AMC}t()\scriptstyle{t()}MCS\scriptstyle{MCS}MCI\scriptstyle{MCI}KDE\scriptstyle{KDE}
Figure 6: Approximations of the product of two Beta pdfs, pXp_{X} and pYp_{Y}. MCS stands for MC sampling, MCI stands for MC integration, KDE stands for kernel-density estimation, AMC stands for analytic moment computation.

As discussed earlier, choosing which path to take, whether to follow the SL approximation path in upper part of the graph or opt for one of the MC approximations in the lower part, requires evaluating the trade-off between computational complexity and degree of approximation of the different approaches. As these parameters are known in the case of MC simulations, we will review here the computational complexity and the approximation of the binomial multiplication.

Computational complexity of the binomial product operator

Binomial multiplication is extremely efficient. Given two binomial opinions ωX\omega_{X} and ωY\omega_{Y} it is possible to compute their product ωZ\omega_{Z} through a fixed and finite number of arithmetic operations. Independently from the actual form of the mapped distributions, the product is always computed in the same amount of time. As such, the computational complexity of these SL operators is constant 𝒪(1)\mathcal{O}\left(1\right).

Approximation of the binomial operator

The original paper that introduced the SL operator for binomial multiplication [5] proposed a first qualitative analysis of the degree of approximation of this operator. In particular, it considered the specific instance of the multiplication of two Beta pdfs of the form X,Y𝙱𝚎𝚝𝚊(1,1)X,Y\sim\mathtt{Beta}\left(1,1\right); the pdf of XX and YY reduces to a uniform distribution over [0,1][0,1], which is taken to be a worst-case scenario with maximal variance and entropy. The analytical solution Z=XYZ=X\cdot Y to this particular case was then computed and graphically compared to the pdf associated with product ωZ=ωXωY\omega_{Z}=\omega_{X}\cdot\omega_{Y}. This study provided a clear visual appraisal of the difference between the exact pdf and the SL-approximated pdf, but no quantitative estimation were provided for more general cases.

Relating the binomial multiplication and Monte Carlo approximations

In order to compute a numerical estimation of the degree of approximation of the SL operator for binomial multiplication we want to rely on the framework described in Section 6.

The basic condition expressed in Equation 20 requires the bias of p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}} to be bounded and negligible. Recall that this bias, using a Gaussian kernel with width computed using the Silverman rule, is EKDE[pZp^ZKDE]1.062σ^2N25E_{KDE}\left[p_{Z}-\hat{p}_{Z}^{\mathrm{KDE}}\right]\propto 1.06^{2}\hat{\sigma}^{2}N^{\frac{2}{5}}, where

σ^=i=1N(xiμ^)2N1.\hat{\sigma}=\sqrt{\frac{\sum_{i=1}^{N}\left(x_{i}-\hat{\mu}\right)^{2}}{N-1}}. (33)

Now, notice that on our bounded support [0,1]\left[0,1\right] we can expect the difference (xiμ^)\left(x_{i}-\hat{\mu}\right) to, be at most, in the order of 10110^{-1}. This implies that, in the worst case, the order of magnitude of σ^\hat{\sigma} may be estimated as:

σ^<N(101)2N1\hat{\sigma}<\sqrt{\frac{{N}\left(10^{-1}\right)^{2}}{N-1}} (34)
σ^<101.\hat{\sigma}<10^{-1}. (35)

Consequently, relying on Silverman rule in Equation 16, the order of magnitude of largest kernel width ww may be bounded as:

w<1.061011N5w<1.06\cdot 10^{-1}\frac{1}{\sqrt[5]{N}} (36)
w<101N15.w<10^{-1}N^{-\frac{1}{5}}. (37)

As such, from Equation 18, the bias will be proportional to this upper bound:

EKDE[pZp^ZKDE](101N15)2.E_{KDE}\left[p_{Z}-\hat{p}_{Z}^{\mathrm{KDE}}\right]\propto\left(10^{-1}N^{-\frac{1}{5}}\right)^{2}. (38)

Thus, for instance, if we were to run our MC simulation sampling N=105N=10^{5} samples, then we can expect the bias of p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}} to be in the order of 10410^{-4}. This analysis on the bias allows us to consider the bias negligible if we are comparing it with quantities, such as D[p^ZKDE(N),pZSL]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),p_{Z}^{SL}\right], order of magnitude greater than 10410^{-4}. If this condition is met, then we can estimate the degree of approximation of binomial multiplication adopting the framework illustrated in Figure 5 and instantiated for this specific SL operator as in Figure 7.

(ωX,ωY){\left(\omega_{X},\omega_{Y}\right)}ωZ{\omega_{Z}}pZSL{p_{Z}^{SL}}|pZSLp^ZKDE|{\int\left|p_{Z}^{SL}-\hat{p}_{Z}^{\mathrm{KDE}}\right|}D^I[pZSL,pZ]{\hat{D}_{I}\left[p_{Z}^{SL},p_{Z}\right]}(pX,pY){\left(p_{X},p_{Y}\right)}pZ{p_{Z}}{z1,z2zN}{\{z_{1},z_{2}\dots z_{N}\}}p^ZKDE{\hat{p}_{Z}^{\mathrm{KDE}}}(μ^,σ^){\left(\hat{\mu},\hat{\sigma}\right)}p^ZGAUSS{\hat{p}_{Z}^{\mathrm{GAUSS}}}|p^ZGAUSSp^ZKDE|{\int\left|\hat{p}_{Z}^{\mathrm{GAUSS}}-\hat{p}_{Z}^{\mathrm{KDE}}\right|}D^I[p^ZGAUSS,pZ]{\hat{D}_{I}\left[\hat{p}_{Z}^{\mathrm{GAUSS}},p_{Z}\right]}p^ZBETA{\hat{p}_{Z}^{\mathrm{BETA}}}|p^ZBETAp^ZKDE|{\int\left|\hat{p}_{Z}^{\mathrm{BETA}}-\hat{p}_{Z}^{\mathrm{KDE}}\right|}D^I[p^ZBETA,pZ]{\hat{D}_{I}\left[\hat{p}_{Z}^{\mathrm{BETA}},p_{Z}\right]}(μ,σ){\left({\mu},{\sigma}\right)}pZGAUSS{{p}_{Z}^{\mathrm{GAUSS}}}|pZGAUSSp^ZKDE|{\int\left|{p}_{Z}^{\mathrm{GAUSS}}-\hat{p}_{Z}^{\mathrm{KDE}}\right|}D^I[pZGAUSS,pZ]{\hat{D}_{I}\left[{p}_{Z}^{\mathrm{GAUSS}},p_{Z}\right]}pZBETA{{p}_{Z}^{\mathrm{BETA}}}|pZBETAp^ZKDE|{\int\left|{p}_{Z}^{\mathrm{BETA}}-\hat{p}_{Z}^{\mathrm{KDE}}\right|}D^I[pZBETA,pZ]{\hat{D}_{I}\left[{p}_{Z}^{\mathrm{BETA}},p_{Z}\right]}\scriptstyle{\cdot}s()\scriptstyle{s()}MCI\scriptstyle{MCI}\scriptstyle{\cdot}AMC\scriptstyle{AMC}t()\scriptstyle{t()}MCS\scriptstyle{MCS}MCI\scriptstyle{MCI}KDE\scriptstyle{KDE}MCI\scriptstyle{MCI}MCI\scriptstyle{MCI}MCI\scriptstyle{MCI}MCI\scriptstyle{MCI}

Figure 7: Evaluations of the distances between pZp_{Z} and its approximations based on the assumption that D[pZ,p^ZKDE(N)]0D\left[p_{Z},\hat{p}_{Z}^{\mathrm{KDE}}(N)\right]\approx 0. MCS stands for MC sampling, MCI stands for MC integration, KDE stands for kernel-density estimation, AMC stands for analytic moment computation.

7.1 Empirical Evaluation

In this section we describe our experimental simulations for the evaluation of the degree of approximation of the binomial product. We will first offer a qualitative analysis of the SL approximation pZSLp_{Z}^{SL} and the approximation generated via MC simulation p^ZMC\hat{p}_{Z}^{\mathrm{MC}}; then, we will provide a quantitative statistical assessment of the distance D[pZ,pZSL]D\left[p_{Z},p_{Z}^{SL}\right]; next, we will analyze a specific study case concerning the worst-case scenario of the product of two degenerate Beta random variables; finally, we will assess quantitatively the degree of approximation in the product of multiple opinions.

In our simulations starting in the domain of subjective logic, opinions ω=(b,d,u,a)\omega=\left(b,d,u,a\right) are sampled randomly. The parameters bb, dd and uu must be sampled from a simplex defined by the constraint b+d+u=1b+d+u=1; therefore we sample them from a Dirichlet distribution 𝙳𝚒𝚛(𝜶)\mathtt{Dir}\left(\boldsymbol{\alpha}\right) with 𝜶=[1,1,1]\boldsymbol{\alpha}=\left[1,1,1\right], which guarantees a uniform sampling over the simplex. The parameter aa, instead, is sampled from a uniform distribution 𝚄𝚗𝚒𝚏(0,1)\mathtt{Unif}\left(0,1\right). In the simulations starting in the domain of probability distributions, Beta distributions 𝙱𝚎𝚝𝚊(α,β)\mathtt{Beta}\left(\alpha,\beta\right) are sampled randomly; both parameters α\alpha and β\beta are drawn from a uniform pdf on a bounded domain, 𝚄𝚗𝚒𝚏(0,10)\mathtt{Unif}\left(0,10\right).

All the simulations are carried out using the WebPPL probabilistic programming language [3] and the scripts are available online333https://github.com/FMZennaro/SLMC/BinomialProduct.

7.1.1 Qualitative simulations

In the qualitative simulations we aim at getting a first intuitive feeling about the approximation of pZSLp_{Z}^{SL}.

Protocol

In order to compare the SL approximation pZSLp_{Z}^{SL} and the MC approximation p^ZMC\hat{p}_{Z}^{\mathrm{MC}} we adopt the following protocol: (i) we sample two random opinions ωX\omega_{X} and ωY\omega_{Y}; (ii) we compute their product ωZ=ωXωY\omega_{Z}=\omega_{X}\cdot\omega_{Y}; (iii) we project ωZ\omega_{Z} onto the distribution pZSLp_{Z}^{SL}; (iv) we project the opinions ωX\omega_{X} and ωY\omega_{Y} onto the Beta distributions pXp_{X} and pYp_{Y}; (v) we re-create p^ZMC\hat{p}_{Z}^{\mathrm{MC}} using MC simulation to draw NN samples {z1,z2,,zN}\{z_{1},z_{2},\dots,z_{N}\} from pZp_{Z}; finally, (vi) we plot pZSLp_{Z}^{SL} against p^ZMC\hat{p}_{Z}^{\mathrm{MC}} numerically, without any smoothing or interpolation. On the side, (vii) we use the samples {z1,z2,,zN}\{z_{1},z_{2},\dots,z_{N}\} to estimate moments via MC integration and then instantiate the moment-matching approximations p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}} and p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}}; (viii) we use Equation 31 to compute the analytic moment-matching approximations pZGAUSS{p}_{Z}^{\mathrm{GAUSS}} and pZBETA{p}_{Z}^{\mathrm{BETA}}; (ix) we plot the moment-matching approximations against p^ZMC\hat{p}_{Z}^{\mathrm{MC}} numerically.

Results

Figures 8 and 9 illustrates the difference between the SL binomial multiplication pZSLp_{Z}^{SL} and the approximation of the true pdf pZp_{Z} plotted via MC. In some instances, pZSLp_{Z}^{SL} seems to provide a very good approximation of pZp_{Z}, as shown in Figure 8. In other instances, as shown in Figure 9, this approximation is more coarse, especially when it comes to values of the support near the extremes.

Refer to caption
Figure 8: Qualitative simulation. The first opinion ωX\omega_{X} has parameters b=0.61b=0.61, d=0.30d=0.30, u=0.09u=0.09 and a=0.79a=0.79, the second opinion ωY\omega_{Y} has parameters b=0.28b=0.28, d=0.66d=0.66, u=0.06u=0.06 and a=0.46a=0.46. The number of samples is N=105N=10^{5}.
Refer to caption
Figure 9: Qualitative simulation. The first opinion ωX\omega_{X} has parameters b=0.16b=0.16, d=0.55d=0.55, u=0.29u=0.29 and a=0.001a=0.001, the second opinion ωY\omega_{Y} has parameters b=0.43b=0.43, d=0.16d=0.16, u=0.41u=0.41 and a=0.39a=0.39. The number of samples is N=105N=10^{5}.

The discrepancy shown in Figure 9 may be theoretically imputed to a poor approximation of the MC simulation due to a limited number of samples. In order to confute this hypothesis, another identical simulation with a number of samples one order of magnitude larger was run. Figure 10 shows that this simulation returned the same qualitative result. This suggests that the gap between p^ZMC\hat{p}_{Z}^{\mathrm{MC}} and pZSLp_{Z}^{SL} may not be imputed to a poor MC approximation.

Refer to caption
Figure 10: Qualitative simulation. Same settings as in Figure 9, except for the number of samples N=106N=10^{6}.

Figure 11 offers a visual comparison of the approximations offered by p^ZMC\hat{p}_{Z}^{\mathrm{MC}} and pZSLp_{Z}^{SL} contrasted now with the Gaussian p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}}, pZGAUSS{p}_{Z}^{\mathrm{GAUSS}} and the Beta p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}}, pZBETA{p}_{Z}^{\mathrm{BETA}} approximations. The analytic and empirical (via MC) approximations behave in a very similar fashion. Overall, the Gaussian approximations are the farthest from pZp_{Z}, while the Beta approximations follow very closely pZSLp_{Z}^{SL} and p^ZMC\hat{p}_{Z}^{\mathrm{MC}}; in particular, the analytic Beta approximation pZBETA{p}_{Z}^{\mathrm{BETA}} almost overlap pZSLp_{Z}^{SL}, because of a similar approach in evaluating the moments of pZp_{Z}.

Refer to caption
Figure 11: Qualitative simulation. The first opinion ωX\omega_{X} has parameters b=0.35b=0.35, d=0.23d=0.23, u=0.42u=0.42 and a=0.83a=0.83, the second opinion ωY\omega_{Y} has parameters b=0.14b=0.14, d=0.26d=0.26, u=0.60u=0.60 and a=0.77a=0.77. The number of samples is N=105N=10^{5}.
Discussion

This analysis suggests that the SL approximation pZSLp_{Z}^{SL} may provide a good and useful estimation of the true pdf pZp_{Z}; indeed, pZSLp_{Z}^{SL} follows very closely the shape p^ZMC\hat{p}_{Z}^{\mathrm{MC}} which, in turn, is close to pZp_{Z}. Given that pZSLp_{Z}^{SL} consists of a smooth Beta distribution, it is not surprising that the approximation suffers the worst near the boundaries where the MC estimate p^ZMC\hat{p}_{Z}^{\mathrm{MC}} diverges from pZSLp_{Z}^{SL}, as shown in Figure 9. The results also discourage the naive possibility of using a Gaussian approximation, since p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}} mismodels the true pdf pZp_{Z} by centering the mean but spreading probability mass too widely beyond the domain [0,1][0,1]. Instead, a Beta approximation p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}} or pZBETA{p}_{Z}^{\mathrm{BETA}} provides an approximation qualitatively very close to pZSLp_{Z}^{SL} and p^ZMC\hat{p}_{Z}^{\mathrm{MC}}; also, notice that the Beta approximation pZBETA{p}_{Z}^{\mathrm{BETA}} can be computed as cheaply as the SL approximation, with complexity 𝒪(1)\mathcal{O}\left(1\right), by evaluating its mean and variance analytically.

7.1.2 Quantitative simulations

Quantifying the gap between pZp_{Z} and pZSLp_{Z}^{SL} that we observed in the qualitative study above is the aim of the quantitative simulations.

Protocol

The first part of our quantitative protocol is the same as the qualitative protocol: (i-a) we sample two random opinions ωX\omega_{X} and ωY\omega_{Y}; (ii-a) we compute their product ωZ=ωXωY\omega_{Z}=\omega_{X}\cdot\omega_{Y}; (iii-a) we project ωZ\omega_{Z} onto the distribution pZSLp_{Z}^{SL}; (iv-a) we project the opinions ωX\omega_{X} and ωY\omega_{Y} onto the Beta distributions pXp_{X} and pYp_{Y}; (v-a) we re-create p^ZMC\hat{p}_{Z}^{\mathrm{MC}} using MC simulation to draw NN samples {z1,z2,,zN}\{z_{1},z_{2},\dots,z_{N}\} from pZp_{Z}. Then, instead of plotting our results, (vi-a) we use a KDE to explicitly estimate p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}}; and, (vii-a) we compute via MC integration the area determined by the integral 01|pZSL(z)p^ZKDE(z)|𝑑z\int_{0}^{1}\left|p_{Z}^{SL}(z)-\hat{p}_{Z}^{\mathrm{KDE}}(z)\right|dz. On the side, we compute moment-matching approximations as before: (viii) we use the samples {z1,z2,,zN}\{z_{1},z_{2},\dots,z_{N}\} to estimate moments via MC integration and then instantiate p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}} and p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}}; (ix) we use Equation 31 to compute the analytic approximations pZGAUSS{p}_{Z}^{\mathrm{GAUSS}} and pZBETA{p}_{Z}^{\mathrm{BETA}}; (x) we compute via MC integration the area determined by the absolute difference between p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}} and each moment-matching approximation.

For completeness, we also run a simulation starting in the domain of pdfs: (i-b) we sample two random Beta pdfs pXp_{X} and pYp_{Y}; (ii-b) we re-create p^ZMC\hat{p}_{Z}^{\mathrm{MC}} using MC simulation to draw NN samples {z1,z2,,zN}\{z_{1},z_{2},\dots,z_{N}\} from pZp_{Z}; (iii-b) we use a KDE to explicitly estimate p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}}; (iv-b) we map the Beta distributions pXp_{X} and pYp_{Y} onto the opinions ωX\omega_{X} and ωY\omega_{Y}; (v-b) we compute their product ωZ=ωXωY\omega_{Z}=\omega_{X}\cdot\omega_{Y}; (vi-b) we project ωZ\omega_{Z} onto the distribution pZSLp_{Z}^{SL}; and, (vii-b) we compute via MC integration the area determined by the integral 01|pZSL(z)p^ZKDE(z)|𝑑z\int_{0}^{1}\left|p_{Z}^{SL}(z)-\hat{p}_{Z}^{\mathrm{KDE}}(z)\right|dz. As before, we also compute distances between p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}} and empirical (via MC) and analytical moment-matching approximations as explained in the steps (viii)-(x) above.

In order to get significant statistical result, we repeat each simulation 100100 times and we compute the mean and the standard deviation of the distance D^I[pZ,pZSL]\hat{D}_{I}\left[p_{Z},p_{Z}^{SL}\right].

Notice that, since the pdf p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}} that we are trying to estimate is defined on a bounded interval, using a Gaussian kernel for KDE is a sub-optimal choice. The Gaussian kernel distributes the mass of probability over the entire real line, and thus we would inevitably spill part of the probability mass beyond the domain [0,1][0,1]. To solve this problem we adopt the logit trick [12]: instead of applying a Gaussian KDE to estimate p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}} directly from the samples {z1,z2,,zN}\{z_{1},z_{2},\dots,z_{N}\}, we use a logit transform logit(x)=logx1x\textnormal{logit}(x)=\log\frac{x}{1-x} to project the sample {z1,z2,,zN}\{z_{1},z_{2},\dots,z_{N}\} onto the entire real line; we then apply a Gaussian KDE to the projected samples and rescale back the learned pdf to p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}}.

Refer to Figure 7 for the diagram of the experimental protocol for the quantitative simulations.

Results

Figure 12 shows the variation in the distances D^[pZ,]\hat{D}\left[p_{Z},\cdot\right] estimated as D[p^ZKDE(N),]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),\cdot\right] as a function of the number samples NN generated in the MC simulation. All the statistics are computed from 100100 repetitions and using 10310^{3} uniformly sampled points on the support [0,1][0,1] to perform MC integration.

Refer to caption
Figure 12: Quantitative simulation. Mean and standard deviation of the distance D[p^ZKDE(N),]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),\cdot\right] when considering different approximations: pZSLp_{Z}^{SL} in the simulation starting from opinions and following the steps (i-a)-(vii-a) (SL), pZSLp_{Z}^{SL} in the simulation starting from Beta pdfs and following the steps (i-b)-(vii-b) (SL’), p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}} (Gauss), p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}} (Beta). Note that the xx-axis is in a logarithmic (in base 10) scale.

The stable trend of all the distances D[p^ZKDE(N),]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),\cdot\right] suggests that the MC simulations sampled enough points, for all the values of NN that we considered, to return a good approximation.

More importantly, recall that our whole analysis holds only if Equation 20 is satisfied. Using N=105N=10^{5}, we know from Equation 38 that the bias in evaluating p^ZKDE(N)\hat{p}_{Z}^{\mathrm{KDE}}(N) is in the order of 10410^{-4}. Thus compared to the scale of the mean and variance error in our results, which are in the scale of 10110^{-1}, we can confirm that the bias is negligible. We can then state that D[p^ZKDE(N),]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),\cdot\right] does indeed provide a good estimate of D^[pZ,]\hat{D}\left[p_{Z},\cdot\right].

Consistently with the previous experiments, the Gaussian approximations provides the worst approximation. Indeed, with distances D[p^ZKDE(N),p^ZGAUSS]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),\hat{p}_{Z}^{\mathrm{GAUSS}}\right] and D[p^ZKDE(N),pZGAUSS]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),{p}_{Z}^{\mathrm{GAUSS}}\right] averaging around 0.300.350.30-0.35, we can expect one sixth of the probability mass of a Gaussian approximation not to overlap with the true distribution pZp_{Z}.

The Beta approximations p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}} and pZBETA{p}_{Z}^{\mathrm{BETA}} clearly offer a better solution. Even if the product of two Beta distributions pZp_{Z} is not a Beta distribution, it is clear from these results that the shape of pZp_{Z} is in general very close to a Beta pdf. Indeed, the expected value of D[p^ZKDE(N),p^ZBETA]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),\hat{p}_{Z}^{\mathrm{BETA}}\right] and D[p^ZKDE(N),pZBETA]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),{p}_{Z}^{\mathrm{BETA}}\right] point out that 95%95\% of the mass of a Beta approximation and pZp_{Z} overlap with very limited variance.

The SL approximation pZSLp_{Z}^{SL} also offers a good solution. The result of the simulation in which we started from opinions and the one in which we started from Beta pdfs are extremely close. This offers a confirmation of the robustness of the transformations between the domain of opinions and the domain of pdfs. Overall, the expected value of D[p^ZKDE(N),pZSL]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),p_{Z}^{SL}\right] suggests that the typical overlap between the mass of pZSLp_{Z}^{SL} and pZp_{Z} settles around 90%90\%, slightly worse than the Beta approximation. The high variance points to a strong case-by-case variability: in certain scenario pZSLp_{Z}^{SL} may provide a model as good or better than p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}} or pZBETA{p}_{Z}^{\mathrm{BETA}}, but on other instances its quality may degrade further.

Discussion

The results of our quantitative analysis agree with the qualitative study. A Gaussian approximation p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}} or pZGAUSS{p}_{Z}^{\mathrm{GAUSS}} was shown to be a poor choice for modelling the product of two Beta distributions (and, for this reason, we will drop this approximation from the next simulations). Instead, the SL approximation pZSLp_{Z}^{SL} and the Beta approximations p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}} or pZBETA{p}_{Z}^{\mathrm{BETA}} are both good approximations, assuming that we can accept a difference between the true pdf and the approximation up to 5%10%5\%-10\% of the probability density. With limited computational resources, pZBETA{p}_{Z}^{\mathrm{BETA}} seems to be, on average, the best bet.

7.1.3 Limit-case Study

In this limit-case study, we consider the worst-case scenario considered in [5]. This study provides a way to enrich the previous study and reconnect this paper to it.

Protocol

We quantitatively analyze the case in which both opinions ωX\omega_{X} and ωY\omega_{Y} are degenerate Beta pdfs of the form 𝙱𝚎𝚝𝚊(1,1)\mathtt{Beta}\left(1,1\right) with a=12a=\frac{1}{2}. To provide a quantitative analysis we follow the same protocol used in Section 7.1.2: first, we derive the MC approximation p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}} (using the KDE algorithm), the SL approximation pZSLp_{Z}^{SL}, the empirical (via MC) Beta approximation p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}} and the analytic Beta approximation pZBETA{p}_{Z}^{\mathrm{BETA}}; then, we compute the distance between the aforementioned distributions and the true pdf pZp_{Z}, whose exact form, log(z)-\log(z), is given in [5].

Results

Table 2 shows the evaluation of the distance D^[pZ,]\hat{D}\left[p_{Z},\cdot\right] with respect to the true pdf pZ=log(z)p_{Z}=-\log(z), when performing MC simulations with N=106N=10^{6} points and using 10310^{3} uniformly sampled points on the support [0,1][0,1] to perform MC integration.

The results show that the difference between the approximations is about one order of magnitude from each other. The MC approximation is, as expected, very close to the true pdf pZp_{Z}, with a distance averaging around 91039\cdot 10^{-3}. This is higher than the theoretically computed value, likely due to the fact that we are evaluating a limit case; however, this difference is still small enough to allow us a comparison with the other approximations. The Beta approximations have a slightly higher distance around 31023\cdot 10^{-2}. Finally, the SL approximation has the highest distance at around 21012\cdot 10^{-1}, meaning that the probability mass of pZSLp_{Z}^{SL} and pZp_{Z} overlap for about 90%90\%. All the results also show a high variance, which is caused by the difficulty in numerically approximating values near 0, where the true pdf pZ=log(z)p_{Z}=-\log(z) diverges.

D^[pZ,]\hat{D}\left[p_{Z},\cdot\right]
KDE 0.00871 ±\pm 0.01242
SL 0.20793 ±\pm 0.93301
Beta (MC) 0.03070 ±\pm 0.10018
Beta (Analytic) 0.03635 ±\pm 0.14617
Table 2: Limit case study. Evaluation of the distance D^[pZ,]\hat{D}\left[p_{Z},\cdot\right] when using a MC approximation p^ZMC\hat{p}_{Z}^{\mathrm{MC}}, a SL approximation pZSLp_{Z}^{SL} and an empirical Beta approximation p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}}, and an analytic Beta approximation pZBETA{p}_{Z}^{\mathrm{BETA}}.
Discussion

The results are consistent with our previous results obtained in the quantitative analysis in Section 7.1.2 and they confirm that the scenario considered in [5] with two degenerate Beta pdfs of the form 𝙱𝚎𝚝𝚊(1,1)\mathtt{Beta}\left(1,1\right) and a=12a=\frac{1}{2} is indeed a hard case for SL approximation. The MC approximation performs better in modeling the true form of the pdf log(z)-\log(z) and, compared to it, the SL approximation is two orders of magnitude less precise in terms of integral distance. This simulation thus clearly highlights the cost in terms of accuracy that the computational simplicity of SL implies.

7.1.4 Multiple Products

In this last experimental section we consider the product of multiple opinions and we examine how approximation spread.

Protocol

We quantitatively evaluate the product of multiple opinions ωX1\omega_{X_{1}}, ωX2\omega_{X_{2}}ωXL\omega_{X_{L}} by randomly sampling LL opinions and then defining the product ωZ=(((ωX1ωX2)ωX3)ωXL)\omega_{Z}=\left(\left(\left(\omega_{X_{1}}\cdot\omega_{X_{2}}\right)\cdot\omega_{X_{3}}\right)\dots\cdot\omega_{X_{L}}\right) and pZ=pX1pX2pXLp_{Z}=p_{X_{1}}\cdot p_{X_{2}}\dots\cdot p_{X_{L}}. The following analysis adopts the same protocol used in the quantitative simulations in Section 7.1.2 in order to compute the SL approximation pZSLp_{Z}^{SL} and the analytic Beta approximation pZBETA{p}_{Z}^{\mathrm{BETA}}, and then evaluate the integral distances D^I[pZ,]=^D[p^ZKDE(N),]\hat{D}_{I}\left[p_{Z},\cdot\right]\hat{=}D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),\cdot\right], where now the final pdf over ZZ is given by the product of multiple opinions. Notice that, while the convergence properties of the MC simulation remains the same, we may expect the precision of the SL approximation and the Beta approximation to degrade over multiple products as successive approximations cumulate.

Results

Figure 13 shows the variation of the distance D[p^ZKDE(N),pZSL]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),p_{Z}^{SL}\right] and D[p^ZKDE(N),pZBETA]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),{p}_{Z}^{\mathrm{BETA}}\right] as a function of the number LL of opinions that are multiplied together to determine ZZ. A slight increase in the degree of approximation may be observed as the number of factors increases from 22 to 55.

Refer to caption
Figure 13: Multiple products. Mean and standard deviation of the distance D[p^ZKDE(N),pZSL]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),p_{Z}^{SL}\right] and D[p^ZKDE(N),pZBETA]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),{p}_{Z}^{\mathrm{BETA}}\right] when the distribution ZZ is given by the product of multiple factors.
Discussion

The hypothesis that the degree of approximation of the SL operator and analytical Beta approximation degrades over multiple products because of the accumulation of approximation appears to be correct. As more factors are taken into consideration, both pZSLp_{Z}^{SL} and pZBETA{p}_{Z}^{\mathrm{BETA}} slowly diverges from the true pdf pZp_{Z}. A modeler should be aware of these dynamics in case she were to use SL to approximate the product of multiple Beta random variables.

8 Case Study: Fusion of Beta Distributions

In this section, we further showcase the versatility of our framework by applying it to yet another subjective logic operator. We first provide the definition of fusion of Beta distributions; we then introduce the SL operator for fusion and we discuss its approximation and complexity; finally, we apply our framework to the problem of getting an approximation of fusion and we run empirical simulations to validate our theoretical results.

Fusion of Beta pdfs

Let X𝙱𝚎𝚝𝚊(αX,βX)X\sim\mathtt{Beta}\left(\alpha_{X},\beta_{X}\right) and Y𝙱𝚎𝚝𝚊(αY,βY)Y\sim\mathtt{Beta}\left(\alpha_{Y},\beta_{Y}\right) be two independent Beta random variables with associated pdfs pXp_{X} and pYp_{Y}. Let us define a third random variable ZZ as the fusion of the two Beta random variables Z=XYZ=X\circledcirc Y [6]:

XY=xyxy+(1x)(1y),X\circledcirc Y=\frac{x\cdot y}{x\cdot y+(1-x)\cdot(1-y)},

where xx and yy are realizations of the random variables XX and YY. As in the case of the product of Beta random variables, determining the shape of pZp_{Z} is a non-trivial problem. MC simulations offer a robust method to sample points from the probability distribution of ZZ and to estimate pZp_{Z} by moment-matching or kernel density estimation. Notice that, differently from the previous case study, we do not have an exact analytical solution for computing the first two moments of pZp_{Z} [6].

Subjective logic fusion

A SL operator may be instantiated to compute an approximate fusion over two binomial opinions. Given two binomial opinions ωX\omega_{X} and ωY\omega_{Y} defined on the same domain Ω\Omega and with the same prior aa, we define the binomial opinion ωZ\omega_{Z} resulting from the fusion ωXωY\omega_{X}\circledcirc\omega_{Y} as:

ωZ={bZ=mZsZWasZdZ=(1mZ)sZW(1a)sZuZ=WsZaZ=a,\omega_{Z}=\begin{cases}b_{Z}=\frac{m_{Z}s_{Z}-Wa}{s_{Z}}\\ d_{Z}=\frac{(1-m_{Z})s_{Z}-W(1-a)}{s_{Z}}\\ u_{Z}=\frac{W}{s_{Z}}\\ a_{Z}=a,\end{cases} (39)

where

mZ=bXbY+bYauX+bXauY+a2uXuY2(bXbY+bYauX+bXauY+a2uXuY)+1bYauYbXauXm_{Z}=\frac{b_{X}b_{Y}+b_{Y}au_{X}+b_{X}au_{Y}+a^{2}u_{X}u_{Y}}{2\left(b_{X}b_{Y}+b_{Y}au_{X}+b_{X}au_{Y}+a^{2}u_{X}u_{Y}\right)+1-b_{Y}-au_{Y}-b_{X}-au_{X}}\\ (40)
sZ=max{WamZ,W(1a)(1mZ),(WuX+1)(WuY+1)(bXbX2auXa2uX2)(bYbY2auYa2uY2)mZ(1mZ)[(WuY+1)(bYbY2auYa2uY2)+(WuX+1)(bXbX2auXa2uX2)]1}.\begin{split}s_{Z}=&\max\left\{\frac{Wa}{m_{Z}},\frac{W(1-a)}{(1-m_{Z})},\right.\\ &\left.\frac{\left(\frac{W}{u_{X}}+1\right)\left(\frac{W}{u_{Y}}+1\right)\left(b_{X}-b_{X}^{2}-au_{X}-a^{2}u_{X}^{2}\right)\left(b_{Y}-b_{Y}^{2}-au_{Y}-a^{2}u_{Y}^{2}\right)}{m_{Z}(1-m_{Z})\left[\left(\frac{W}{u_{Y}}+1\right)\left(b_{Y}-b_{Y}^{2}-au_{Y}-a^{2}u_{Y}^{2}\right)+\left(\frac{W}{u_{X}}+1\right)\left(b_{X}-b_{X}^{2}-au_{X}-a^{2}u_{X}^{2}\right)\right]}-1\right\}.\end{split}

This formula expresses in the subjective logic formalism the moment-matching approximation of fusion defined in [6]. Practically, a fusion operator allows us to evaluate the aggregation of two different opinions on the same fact.

Approximating the product of Beta pdfs

Now, if we are interested in computing the fusion Z=XYZ=X\circledcirc Y, where XX and YY are two independent Beta random variables, we may rely either on the SL approximation defined in Equation 39 or on approximations computed via MC simulations; as before we will consider the following approximations for the true pdf pZp_{Z}: a SL pdf pZSLp_{Z}^{SL}, a KDE estimation p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}}, a Beta and a Gaussian moment-matching approximation p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}}, p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}} computed by evaluating mean and variance of pZp_{Z} via MC integration. Differently from before, we do not consider an exact analytical moment-matching approximation (pZBETA{p}_{Z}^{\mathrm{BETA}} or pZGAUSS{p}_{Z}^{\mathrm{GAUSS}}) because no exact analytical formula exists. Notice, also, that since we used the moment-matching approximation provided in [6] to define the SL operator in Equation 39, our results on the degree of approximation of the SL operator immediately extend to Operator 1 defined in [6].

Figure 14 provides the concrete instantiation of the diagram in Figure 4 for the SL operator of fusion and illustrates the alternative between the path of MC simulations and SL approximation.

(ωX,ωY){\left(\omega_{X},\omega_{Y}\right)}ωZ{\omega_{Z}}pZSL{p_{Z}^{SL}}(pX,pY){\left(p_{X},p_{Y}\right)}pZ{p_{Z}}{z1,z2zN}{\{z_{1},z_{2}\dots z_{N}\}}p^ZKDE{\hat{p}_{Z}^{\mathrm{KDE}}}(μ^,σ^){\left(\hat{\mu},\hat{\sigma}\right)}p^ZGAUSS{\hat{p}_{Z}^{\mathrm{GAUSS}}}p^ZBETA{\hat{p}_{Z}^{\mathrm{BETA}}}\scriptstyle{\circledcirc}s()\scriptstyle{s()}\scriptstyle{\circledcirc}t()\scriptstyle{t()}MCS\scriptstyle{MCS}MCI\scriptstyle{MCI}KDE\scriptstyle{KDE}
Figure 14: Approximations of the fusion of two Beta pdfs, pXp_{X} and pYp_{Y}. MCS stands for MC sampling, MCI stands for MC integration, KDE stands for kernel-density estimation.
Computational complexity of the fusion operator

Despite the more involved expressions in Equation 39 and 40, the computational complexity for evaluating ωZ\omega_{Z} is still constant with respect to the given opinions ωX\omega_{X} and ωY\omega_{Y}; as such, the asymptotic complexity of the SL operator for fusion is 𝒪(1)\mathcal{O}\left(1\right).

Approximation of the binomial operator

The degree of approximation of the SL operator for fusion is equivalent to the approximation of the moment-matching solution on which it is defined; [6] offers an evaluation of this approximation within the wider context of belief propagation in second-order Bayesian networks. In the following paragraphs, we will instead aim at estimating, in a more directed way, the approximation of the SL operator via the computation of the distance D^I[pZ,pZSL]\hat{D}_{I}\left[p_{Z},p_{Z}^{SL}\right].

Relating the binomial multiplication and Monte Carlo approximations

Following the approach described in the previous section, we will compute a numerical estimation of the degree of approximation of the SL operator for fusion relying on the framework described in Section 6.

Once again, we need to check that the basic condition expressed in Equation 20 requiring the bias of p^ZKDE\hat{p}_{Z}^{\mathrm{KDE}} to be negligible is satisfied. Given that we will compute KDE using a Gaussian kernel with width defined by the Silverman rule (Equation 16), and given that the support of ZZ is bounded on [0,1]\left[0,1\right], we can again expect the bias of our KDE estimator to be in the order of (101N15)2\left(10^{-1}N^{-\frac{1}{5}}\right)^{2} (Equation 38). Thus, the KDE bias may be considered negligible in the estimation of distances, if the distances we are considering are orders of magnitude greater than this bias. If this condition is met, then we can estimate the degree of approximation of fusion adopting the framework defined in Figure 5 and instantiated in Figure 15.

(ωX,ωY){\left(\omega_{X},\omega_{Y}\right)}ωZ{\omega_{Z}}pZSL{p_{Z}^{SL}}|pZSLp^ZKDE|{\int\left|p_{Z}^{SL}-\hat{p}_{Z}^{\mathrm{KDE}}\right|}D^I[pZSL,pZ]{\hat{D}_{I}\left[p_{Z}^{SL},p_{Z}\right]}(pX,pY){\left(p_{X},p_{Y}\right)}pZ{p_{Z}}{z1,z2zN}{\{z_{1},z_{2}\dots z_{N}\}}p^ZKDE{\hat{p}_{Z}^{\mathrm{KDE}}}(μ^,σ^){\left(\hat{\mu},\hat{\sigma}\right)}p^ZGAUSS{\hat{p}_{Z}^{\mathrm{GAUSS}}}|p^ZGAUSSp^ZKDE|{\int\left|\hat{p}_{Z}^{\mathrm{GAUSS}}-\hat{p}_{Z}^{\mathrm{KDE}}\right|}D^I[p^ZGAUSS,pZ]{\hat{D}_{I}\left[\hat{p}_{Z}^{\mathrm{GAUSS}},p_{Z}\right]}p^ZBETA{\hat{p}_{Z}^{\mathrm{BETA}}}|p^ZBETAp^ZKDE|{\int\left|\hat{p}_{Z}^{\mathrm{BETA}}-\hat{p}_{Z}^{\mathrm{KDE}}\right|}D^I[p^ZBETA,pZ]{\hat{D}_{I}\left[\hat{p}_{Z}^{\mathrm{BETA}},p_{Z}\right]}\scriptstyle{\circledcirc}s()\scriptstyle{s()}MCI\scriptstyle{MCI}\scriptstyle{\circledcirc}t()\scriptstyle{t()}MCS\scriptstyle{MCS}MCI\scriptstyle{MCI}KDE\scriptstyle{KDE}MCI\scriptstyle{MCI}MCI\scriptstyle{MCI}

Figure 15: Evaluations of the distances between pZp_{Z} and its approximations based on the assumption that D[pZ,p^ZKDE(N)]0D\left[p_{Z},\hat{p}_{Z}^{\mathrm{KDE}}(N)\right]\approx 0. MCS stands for MC sampling, MCI stands for MC integration, KDE stands for kernel-density estimation.

8.1 Empirical Evaluation

In this section we describe our experimental simulations for the evaluation of the degree of approximation of fusion. We will first provide a qualitative assessment of the SL approximation pZSLp_{Z}^{SL} and the approximations generated via MC simulations; then, we will provide a quantitative statistical evaluation of the distance D[pZ,]D\left[p_{Z},\cdot\right] for the SL approximation and for Beta and Gaussian moment-matching approximations.

In our simulations we randomly generate opinion and pdfs using the same protocol defined in Section 7.1. These simulations are also run using the WebPPL probabilistic programming language [3] and the scripts are available online444https://github.com/FMZennaro/SLMC/Fusion.

8.1.1 Qualitative simulations

In the qualitative simulations we try to offer a first visual assessment of the quality of the approximation of pZSLp_{Z}^{SL}.

Protocol

We follow the same protocol defined for the qualitative simulations in Section 7.1, just replacing the operations for binomial multiplication with the operations for fusion.

Results

Figures 16 and 17 offer a visual comparison of the approximations offered by p^ZMC\hat{p}_{Z}^{\mathrm{MC}} and pZSLp_{Z}^{SL} along with the Gaussian p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}} and the Beta p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}} moment-matching approximations computed via MC integration. While the pdf underlying the fusion of two random variables is not Beta distributed, both the SL approximation pZSLp_{Z}^{SL} and the moment-matching approximation p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}} seem to offer a good coarse approximation of the true distribution pZp_{Z}; both approximations match very well the true distribution within the support [0,1][0,1], but they may show problems modelling the behaviour of pZp_{Z} near the boundaries of the domain. As before, the Gaussian approximation p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}} offers the worst match, as its shape is not ideal to model a pdf on a bounded domain.

Refer to caption
Figure 16: Qualitative simulation. The first opinion ωX\omega_{X} has parameters b=0.16b=0.16, d=0.58d=0.58, u=0.26u=0.26 and a=0.57a=0.57, the second opinion ωY\omega_{Y} has parameters b=0.18b=0.18, d=0.64d=0.64, u=0.18u=0.18 and a=0.57a=0.57. The number of samples is N=106N=10^{6}.
Refer to caption
Figure 17: Qualitative simulation. The first opinion ωX\omega_{X} has parameters b=0.14b=0.14, d=0.63d=0.63, u=0.23u=0.23 and a=0.54a=0.54, the second opinion ωY\omega_{Y} has parameters b=0.83b=0.83, d=0.05d=0.05, u=0.12u=0.12 and a=0.54a=0.54. The number of samples is N=106N=10^{6}.
Discussion

This qualitative assessment suggests that the SL approximation pZSLp_{Z}^{SL} may provide a good enough approximation of the true pdf pZp_{Z} at a very low computational cost. The Beta moment-matching approximation evaluated computing mean and variance via MC simulations seems to behave in a very similar fashion; however, in this case where no exact analytical moment-matching approximation is possible, the cost of evaluating p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}} corresponds to the cost of running the whole MC simulation. Last, the Gaussian moment-matching approximation p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}} constitutes a sub-optimal choice, as its fit is worse than the alternative approximations and its computational cost is equivalent to the Beta moment-matching approximation.

8.1.2 Quantitative simulations

We now move to quantifying the gap between pZp_{Z} and pZSLp_{Z}^{SL} observed in the above simulations.

Protocol

We follow the same protocol defined for the quantitative simulations in Section 7.1, just replacing the operations for binomial multiplication with the operations for fusion. As before, Figure 7 offers a diagram of our experimental protocol.

Results

Figure 18 presents the estimation of the distances D^[pZ,]\hat{D}\left[p_{Z},\cdot\right] as a function of the number samples NN generated in the MC simulation. These statistics are computed performing 100100 repetitions and using 10310^{3} uniformly sampled points on the support [0,1][0,1] for MC integration.

Refer to caption
Figure 18: Quantitative simulation. Mean and standard deviation of the distance D[p^ZKDE(N),]D\left[\hat{p}_{Z}^{\mathrm{KDE}}(N),\cdot\right] when considering different approximations: pZSLp_{Z}^{SL} (SL), p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}} (Gauss) and p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}} (Beta). Note that the xx-axis is in a logarithmic (in base 10) scale.

In general, as in the previous simulation, we can make two preliminary observation: (i) all the distances show a stable trend, thus suggesting that our MC results are approaching their asymptotic limit; (ii) the order of magnitude of the distances is significantly greater than the KDE bias (Equation 38), thus meaning that its bias is negligible with respect to the distances.

Confirming the previous qualitative investigation, the distance between the true pdf pZp_{Z} and the Gaussian approximation p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}} reaches values as high as 0.50.5, meaning that up to one fourth of the probability mass of p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}} does not to overlap with the true distribution pZp_{Z}.

The Beta approximation p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}} models the true pdf pZp_{Z} better, with a distance around 0.100.150.10-0.15, suggesting a good approximation in which 95%95\% of the mass of p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}} overlaps with pZp_{Z}.

Similarly, the SL approximation pZSLp_{Z}^{SL} provides an equally good solution. The values of distance for pZSLp_{Z}^{SL} are well within the range of the standard deviation of p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}}, suggesting that the degree of approximation of SL and Beta moment-matching are very close.

Discussion

The results of this analysis agree with the previous qualitative simulations. Moreover, even if these results are quantitatively different, they are qualitatively in line with our study of the binomial product operator. The quantitative difference between the approximation of the binomial product and the fusion may be likely ascribed to the fact that it is easier to model the product of independent random variables instead of an arbitrary operation like fusion. From a qualitative point of view, though, the Gaussian approximation p^ZGAUSS\hat{p}_{Z}^{\mathrm{GAUSS}} ranks again last among the modeling options, while the SL approximation pZSLp_{Z}^{SL} and the Beta approximations p^ZBETA\hat{p}_{Z}^{\mathrm{BETA}} offer better solutions, at a computational cost that is constant (in the case of SL) or linear in the number of MC samples (in the case of Beta approximation).

9 Conclusion and Future Work

In this paper we studied the use of subjective logic as a framework for approximating operations over probability distributions. As in the case of any approximation, we considered SL operators from the perspective of the trade-off between the computational simplicity they guarantee and the precision they sacrifice. We proposed a protocol based on MC simulations to evaluate quantitatively this trade-off, estimating the distance between the SL approximation and a KDE estimation, under the assumption of a negligible bias between the KDE reconstruction and the true probability distribution.

We applied our protocol to the case study of the product and the fusion of two independent Beta distributions. The first case is relevant to fields like reliability analysis, while the second one is used in the field of subjective logic. In general, SL operators guarantee the preservation of the first moment, but do not strictly preserve higher moments or quantiles. To quantify the degree of approximation of the SL operators, we compared them with other standard approximations, such as moment-matching with a Gaussian pdf, moment-matching with a Beta pdf, and KDE via MC. Our simulations showed that, at the cost of accepting a difference between the SL approximation and the true pdf, SL offers a computationally efficient approximation. Both in the case of binomial products and in the case of fusion, the degree of approximation can be quantified in a mismatch between the SL approximation and the true pdf of up to 10%10\% of the probability mass. In general, KDE approximation and Beta approximation provided better estimation; KDE, however, has a computational cost that is quadratic in the number of samples generated via MC; moment-matching has a computational cost that can be constant and comparable to SL when moments of interest can be computed analytically, or, otherwise, linear in the number of samples generated via MC.

In summary, it is possible to enjoy the computational efficiency and the interpretability of SL if the modeling scenario allows room for approximations up to the amount estimated using our protocol. The recommendation is that, were SL operators to be used to model critical systems (as in the case of reliability analysis or when higher-order moments are critical), this divergence between the true pdf and the SL approximation that we highlighted should be factored in the analysis.

Further work will be developed for better characterizing the difference between true pdfs and SL approximations; in particular, understanding how the mass is differently allocated with respect to the overall shape of the pdf, whether, for instance, these differences are more accentuated near the mode (assuming one exists) or around the tail. According to the way in which probability mass is misplaced in SL approximations different forms of correction may be then considered.

Acknowledgment

The authors would like to express their gratitude to the anonymous reviewers who commented on the first version of this article and helped improving it.

References

  • [1] CA Coelho and RP Alberto. On the distribution of the product of independent beta random variables–applications. Technical report, Technical report, CMA 12 Google Scholar, 2012.
  • [2] Zoubin Ghahramani. Probabilistic machine learning and artificial intelligence. Nature, 521(7553):452, 2015.
  • [3] Noah D Goodman and Andreas Stuhlmüller. The design and implementation of probabilistic programming languages, 2014.
  • [4] A. Jøsang. Subjective Logic: A Formalism for Reasoning Under Uncertainty. Artificial Intelligence: Foundations, Theory, and Algorithms. Springer International Publishing, 2016.
  • [5] Audun Jøsang and David McAnally. Multiplication and comultiplication of beliefs. International Journal of Approximate Reasoning, 38(1):19–51, 2005.
  • [6] Lance Kaplan and Magdalena Ivanovska. Efficient belief propagation in second-order Bayesian networks for singly-connected graphs. International Journal of Approximate Reasoning, 93:132–152, 2018.
  • [7] Lance M Kaplan, Murat Şensoy, Yuqing Tang, Supriyo Chakraborty, Chatschik Bisdikian, and Geeth de Mel. Reasoning under uncertainty: Variations of subjective logic deduction. In Information Fusion (FUSION), 2013 16th International Conference on, pages 1910–1917. IEEE, 2013.
  • [8] David J.C. MacKay. Information theory, inference, and learning algorithms. Cambridge University Press, 2003.
  • [9] TG Pham and N Turkkan. Reliability of a standby system with beta-distributed component lives. IEEE Transactions on Reliability, 43(1):71–75, 1994.
  • [10] Thu Pham-Gia and Noyan Turkkan. The product and quotient of general beta distributions. Statistical Papers, 43(4):537–550, 2002.
  • [11] Jose C. Principe. Information theoretic learning: Rényi’s entropy and kernel perspectives. Springer, 2010.
  • [12] Cosma Shalizi. Advanced data analysis from an elementary point of view, 2013.
  • [13] Bharath K Sriperumbudur, Kenji Fukumizu, Arthur Gretton, Bernhard Schölkopf, and Gert RG Lanckriet. On integral probability metrics,\\backslashphi-divergences and binary classification. arXiv preprint arXiv:0901.2698, 2009.
  • [14] Jen Tang and AK Gupta. On the distribution of the product of independent beta random variables. Statistics & Probability Letters, 2(3):165–168, 1984.