library(Colossus)
#> Note: From versions 1.3.1 to 1.4.1 the expected inputs changed. Regressions are now run with CoxRun and PoisRun and formula inputs. Please see the 'Unified Equation Representation' vignette for more details.
library(data.table)
library(survival)
library(dplyr)
#>
#> Attaching package: 'dplyr'
#> The following objects are masked from 'package:data.table':
#>
#> between, first, last
#> The following objects are masked from 'package:stats':
#>
#> filter, lag
#> The following objects are masked from 'package:base':
#>
#> intersect, setdiff, setequal, union
Logistic Modeling for Binomial Odds and Binary Data
One alternative regression method in Colossus for predicting the probability of an event in independent trials is logistic regression. The theory is presented in the following section.
General Theory
We start by describing the structure of the data. The logistic regression models in Colossus assume the data can be represented in terms of trials, events, and risk factors. Similar to a Poisson model using person-years and number of events in each row, a logistic model requires some measure of the number of independent trials and events for each group. A special case is an analysis in which each row represents one trial, and either 0 or 1 events are observed. The likelihood () for a logistic model is a function of the trials (), events (), and the probability of a success () for each row, . In Colossus, the log-likelihood () is optimized.
Similar to other survival models in Colossus, a model is written as a function of the various risk factors (). This function is related to the probability through a linking function. Three options are currently available: odds, identity, and complementary log linking, based on the options present in 32-bit Epicure.
Similar to other regression models in Colossus, the first derivative vector and second derivative matrix are calculated to optimize the model parameters.
Examples of Use
The use of a logistic model in Colossus closely follows the other regression options. We start by loading and preparing the data. In this example, we will be using the Veterans’ Administration Lung Cancer study data from the survival package. Every row has 1 trial and a binary event status column. Once again, our model uses a trial column, an event column, and the “risk” equation that will be converted to a probability. In this case, the number of trials is 1, so the column can be omitted, but to illustrate, we will artificially create a trial column named “trial”.
#
veteran %>% setDT()
df <- copy(veteran)
# Make the same adjustments as Epicure example 6.5
karno <- df$karno
karno[93] <- 20
df$karno <- karno
df$trt <- df$trt - 1
df$trt <- as.integer(df$trt == 0)
cell_string <- df$celltype
cell <- case_when(
cell_string == "squamous" ~ 1,
cell_string == "smallcell" ~ 2,
cell_string == "adeno" ~ 3,
cell_string == "large" ~ 0
)
df$cell <- cell
df$karno50 <- df$karno - 50
df$trial <- 1
We will start with testing the effects of cell type. By default, Colossus will always remove the first level of a factored column. We can force Colossus to include every level by defining the lowest level to an unused value, -1 in this case. This gives the same final model as if we removed the baseline and added an intercept column (CONST). In this case, the p-values are for the null hypothesis that the cell type has no effect. If we removed the baseline and added an intercept, the p-values would be for the null hypothesis that the cell type has no effect above the baseline.
df$cell <- factor(df$cell, levels = c(-1, 0, 1, 2, 3))
model <- Logit(trial, status) ~ loglinear(cell)
model <- Logit(status) ~ loglinear(cell)
control <- list(verbose = 0, ncores = 2)
e <- LogisticRun(model, df, control = control)
print(e)
#> |-------------------------------------------------------------------|
#> Final Results
#> Covariate Subterm Term Number Central Estimate Standard Error 2-tail p-value
#> <char> <char> <int> <num> <num> <num>
#> 1: cell_0 loglin 0 3.258061 1.0190324 1.387638e-03
#> 2: cell_1 loglin 0 2.047690 0.5312790 1.160781e-04
#> 3: cell_2 loglin 0 2.708038 0.5962816 5.584424e-06
#> 4: cell_3 loglin 0 3.258061 1.0190324 1.387638e-03
#>
#> Logisitic Model Used
#> -2*Log-Likelihood: 64.43, Deviation: 64.43, AIC: 72.43, BIC: 84.11
#> Iterations run: 11
#> maximum step size: 1.07e-04, maximum first derivative: 3.45e-05
#> Analysis converged
#> Run finished in 0.02 seconds
#> |-------------------------------------------------------------------|
In this analysis, we were using the default of the odds ratio. The linking function can be changed by using the “link” option. The following examples show the impact of using each linking function. Note that the probability must be between 0-1, so the starting values are set to ensure that the initial probability estimates are valid. Because we are using categorical risk factors, every model converges to the same score and event probabilities; however, the standard error and p-values are affected. If we were using continuous variables, the linking functions would give a way to define further complicated risk models.
a_n <- c(0.1, 0.1, 0.1, 0.1)
e <- LogisticRun(model, df, control = control, a_n = a_n, link = "odds")
print(e)
#> |-------------------------------------------------------------------|
#> Final Results
#> Covariate Subterm Term Number Central Estimate Standard Error 2-tail p-value
#> <char> <char> <int> <num> <num> <num>
#> 1: cell_0 loglin 0 3.258065 1.0190343 1.387648e-03
#> 2: cell_1 loglin 0 2.047690 0.5312791 1.160781e-04
#> 3: cell_2 loglin 0 2.708039 0.5962820 5.584436e-06
#> 4: cell_3 loglin 0 3.258065 1.0190343 1.387648e-03
#>
#> Logisitic Model Used
#> -2*Log-Likelihood: 64.43, Deviation: 64.43, AIC: 72.43, BIC: 84.11
#> Iterations run: 11
#> maximum step size: 9.57e-05, maximum first derivative: 3.07e-05
#> Analysis converged
#> Run finished in 0.01 seconds
#> |-------------------------------------------------------------------|
a_n <- c(-1, -1, -1, -1)
e <- LogisticRun(model, df, control = control, a_n = a_n, link = "ident")
print(e)
#> |-------------------------------------------------------------------|
#> Final Results
#> Covariate Subterm Term Number Central Estimate Standard Error 2-tail p-value
#> <char> <char> <int> <num> <num> <num>
#> 1: cell_0 loglin 0 -0.03772402 0.03773426 0.31744183
#> 2: cell_1 loglin 0 -0.12136126 0.06071778 0.04563257
#> 3: cell_2 loglin 0 -0.06453783 0.03726759 0.08331963
#> 4: cell_3 loglin 0 -0.03772402 0.03773426 0.31744183
#>
#> Logisitic Model Used
#> -2*Log-Likelihood: 64.43, Deviation: 64.43, AIC: 72.43, BIC: 84.11
#> Iterations run: 11
#> maximum step size: 4.86e-05, maximum first derivative: 1.15e-02
#> Analysis did not converge, check convergence criteria or run further
#> Run finished in 0.01 seconds
#> |-------------------------------------------------------------------|
a_n <- c(0.1, 0.1, 0.1, 0.1)
e <- LogisticRun(model, df, control = control, a_n = a_n, link = "loglink")
print(e)
#> |-------------------------------------------------------------------|
#> Final Results
#> Covariate Subterm Term Number Central Estimate Standard Error 2-tail p-value
#> <char> <char> <int> <num> <num> <num>
#> 1: cell_0 loglin 0 -3.276984 1.0000387 1.049695e-03
#> 2: cell_1 loglin 0 -2.108983 0.5003059 2.493510e-05
#> 3: cell_2 loglin 0 -2.740478 0.5774463 2.076244e-06
#> 4: cell_3 loglin 0 -3.276984 1.0000387 1.049695e-03
#>
#> Logisitic Model Used
#> -2*Log-Likelihood: 64.43, Deviation: 64.43, AIC: 72.43, BIC: 84.11
#> Iterations run: 12
#> maximum step size: 1.26e-04, maximum first derivative: 4.51e-05
#> Analysis converged
#> Run finished in 0.01 seconds
#> |-------------------------------------------------------------------|