This vignette shows how to transform the deterministic Markov model presented in `vignette("c-homogeneous", "heemod")`

in a probabilistic model.

We will start by re-specifying the deterministic model of HIV therapy described previously (a monotherapy strategy `mono`

and combined therapy strategy `comb`

).

But instead of defining transition probabilities and state values directly in `define_transition()`

or `define_state()`

(as in the previous vignette), parameters will be defined first in a `define_parameters()`

step. This is because only parameters defined this way can be resampled in a probabilistic analysis.

```
<- define_parameters(
param rr = .509,
p_AA_mono = .721,
p_AB_mono = .202,
p_AC_mono = .067,
p_AD_mono = .010,
p_BC_mono = .407,
p_BD_mono = .012,
p_CD_mono = .250,
p_AB_comb = p_AB_mono * rr,
p_AC_comb = p_AC_mono * rr,
p_AD_comb = p_AD_mono * rr,
p_BC_comb = p_BC_mono * rr,
p_BD_comb = p_BD_mono * rr,
p_CD_comb = p_CD_mono * rr,
p_AA_comb = 1 - (p_AB_comb + p_AC_comb + p_AD_comb),
cost_zido = 2278,
cost_lami = 2086,
cost_A = 2756,
cost_B = 3052,
cost_C = 9007
)
```

We need to define `p_AA_mono`

and `p_AA_comb`

in `define_parameters()`

because we will need to resample that value. Only values defined with `define_parameters()`

can be resampled. So we cannot use the complement alias `C`

to specify `p_AA_comb`

in `define_transition()`

, as we did before.

```
<- define_transition(
mat_trans_mono
p_AA_mono, p_AB_mono, p_AC_mono, p_AD_mono,0, C, p_BC_mono, p_BD_mono,
0, 0, C, p_CD_mono,
0, 0, 0, 1
)<- define_transition(
mat_trans_comb
p_AA_comb, p_AB_comb, p_AC_comb, p_AD_comb,0, C, p_BC_comb, p_BD_comb,
0, 0, C, p_CD_comb,
0, 0, 0, 1
)
```

State definition remains the same in this example.

```
<- define_state(
state_A cost_health = 2756,
cost_drugs = dispatch_strategy(
mono = cost_zido,
comb = cost_zido + cost_lami
),cost_total = discount(cost_health + cost_drugs, .06),
life_year = 1
)<- define_state(
state_B cost_health = 3052,
cost_drugs = dispatch_strategy(
mono = cost_zido,
comb = cost_zido + cost_lami
),cost_total = discount(cost_health + cost_drugs, .06),
life_year = 1
)<- define_state(
state_C cost_health = 9007,
cost_drugs = dispatch_strategy(
mono = cost_zido,
comb = cost_zido + cost_lami
),cost_total = discount(cost_health + cost_drugs, .06),
life_year = 1
)<- define_state(
state_D cost_health = 0,
cost_drugs = 0,
cost_total = discount(cost_health + cost_drugs, .06),
life_year = 0
)
```

Strategies must be first defined and run as in a standard deterministic analysis.

```
<- define_strategy(
strat_mono transition = mat_trans_mono,
state_A,
state_B,
state_C,
state_D
)
<- define_strategy(
strat_comb transition = mat_trans_comb,
state_A,
state_B,
state_C,
state_D
)
<- run_model(
res_mod mono = strat_mono,
comb = strat_comb,
parameters = param,
cycles = 50,
cost = cost_total,
effect = life_year
)
```

Now we can define the resampling distributions. The following parameters will be resampled:

- Relative risk.
- Costs (such that cost are always positive).
- Transition probability from AIDS to death.
- The transition probabilities from state A.

Since the log of a relative risk follows a lognormal distribution, relative risk follows a lognormal distribution whose mean is `rr`

and standard deviation on the log scale can be deduced from the relative risk confidence interval.

*r**r* ∼ *l**o**g**n**o**r**m**a**l*(*μ*=.509,*σ*=.173)

Programmed as:

`~ lognormal(mean = .509, sdlog = .173) rr `

Usually costs are resampled on a gamma distribution, which has the property of being always positive. Shape and scale parameters of the gamma distribution can be calculated from the mean and standard deviation desired in the distribution. Here we assume that *mean = variance*.

$$cost_A \sim \Gamma(\mu = 2756, \sigma = \sqrt{2756})$$

This can be programmed as:

`~ gamma(mean = 2756, sd = sqrt(2756)) cost_A `

Proportions follow a binomial distribution that can be estimated by giving the mean proportion and the size of the sample used to estimate that proportion with `p_CD ~ binomial(prob = .25, size = 40)`

.

Finally multinomial distributions are declared with the number of individuals in each group in the sample used to estimate the proportions. These proportions follow a Dirichlet distribution:

`p_AA + p_AB + p_AC + p_AD ~ multinomial(721, 202, 67, 10)`

```
<- define_psa(
rsp ~ lognormal(mean = .509, sdlog = .173),
rr
~ gamma(mean = 2756, sd = sqrt(2756)),
cost_A ~ gamma(mean = 3052, sd = sqrt(3052)),
cost_B ~ gamma(mean = 9007, sd = sqrt(9007)),
cost_C
~ binomial(prob = .25, size = 40),
p_CD_mono
+ p_AB_mono + p_AC_mono + p_AD_mono ~ multinomial(721, 202, 67, 10)
p_AA_mono )
```

Now that the distributions of parameters are set we can simply run the probabilistic model as follow:

```
<- run_psa(
pm model = res_mod,
psa = rsp,
N = 100
)
```

The average results are computed. In theory these values are more accurate than simple estimates because of non-linearities. An optional `threshold`

can be passed to `summary()`

to compute net monetary benefit.

```
summary(
pm, threshold = c(1000, 5000, 6000, 1e4))
```

The results of the analysis can be plotted on the cost-effectiveness plane. We can see there seem to be little uncertainty on the costs compared to the uncertainty on the effects, resulting in an uncertainty cloud that looks like a line.

`plot(pm, type = "ce")`

And as cost-effectiveness acceptability curves or EVPI:

```
plot(pm, type = "ac", max_wtp = 10000, log_scale = FALSE)
plot(pm, type = "evpi", max_wtp = 10000, log_scale = FALSE)
```

A covariance analysis can be performed on strategy results:

`plot(pm, type = "cov")`

Or on the difference between strategies:

`plot(pm, type = "cov", diff = TRUE, threshold = 5000)`

As usual plots can be modified with the standard `ggplot2`

syntax.

```
library(ggplot2)
plot(pm, type = "ce") +
xlab("Life-years gained") +
ylab("Additional cost") +
scale_color_brewer(
name = "Strategy",
palette = "Set1"
+
) theme_minimal()
```

Resampling can be significantly sped up by using parallel computing. This can be done in the following way:

- Define a cluster with the
`use_cluster()`

functions (i.e.`use_cluster(4)`

to use 4 cores). - Run the analysis as usual.
- To stop using parallel computing use the
`close_cluster()`

function.

Results may vary depending on the machine, but we found speed gains to be quite limited beyond 4 cores.

To compute EVPPI the results can also be exported with `export_savi()`

in a format compatible with the SAVI software (Sheffield Accelerated Value of Information).

The results can be post-processed by the `BCEA`

package with the `run_bcea()`

function.

```
<- run_bcea(pm, plot = TRUE, Kmax = 10000)
bcea summary(bcea)
```