SynergyLMM (v1)

SynergyLMM Workflow Overview

SynergyLMM workflow overview.
Overview of SynergyLMM workflow. a, The first step is uploading the longitudinal tumor burden-related measurements for the different treatment groups. The input consists of a tabular data in long format with at least 4 columns containing information abouth the samples IDs, the time points for each measurement, the treatment group, and the tumor measurement. b, The input data will then be processed to estimate the linear mixed effect model that fits the tumor growth dynamics for each subject, and estimate the growth rates for each treatment group. c, SynergyLMM offer various functions to evaluate the model diagnostics and model performance, as well as for the identification of potential outliers and influential individuals for the model fit and the treatment groups. d, Once a proper model with satisfactory diagnostics has been obtained, the statistical assessment of combination effects is performed, with a time dependent estimation of the synergy score and combination index, along with their confidence intervals and statistical significance. The method allows for testing synergy using three reference models: Bliss independence, highest single agent (HSA), and response additivity (RA). e, SynergyLMM implements versatile functions to calculate the post hoc power of the experiment and the a priori power by modifying the values of key experimental variables, such as sample size, follow-up time, or frequency of the measurements.

How to Use SynergyLMM

Welcome to the SynergyLMM tutorial!
This quick-start tutorial shows how to use the web-app. An example dataset will be used, which can be downloaded by clicking in the link below:

Download Example Dataset

Data Upload and Model Fit

The user can upload the data in the 'Model Estimation' tab. The input data should be a data table in long format with at least four columns: one for sample IDs, one for time points, one for treatment groups, and one for tumor volume measurements. (See the example dataset for reference).

This is how the data should look like:

Input Table

Once the data is uploaded, the user will need to specify which columns contain each information. For this example, columns Mouse, Day, Group, and TumVol contain the information about the sample IDs, time points, treatments, and tumor measurements, respectively (Fig. 1a).

The next step is to define the time points for the experiment. By default, the minimum and maximum values in the 'Time' column will be used as the start and end time points, but these can be modified. The name of the different treatments in the experiment also need to be specified. In this example, 'Control', 'DrugA', 'DrugB', and 'Combination' correspond to the control, single drugs, and the drug combination groups (Fig. 1b).

Next, clicking on 'Run Analysis' will fit the model and produce a table with the model estimates, (Fig. 1c), including the tumor growth rates of the different groups, and the standard deviations for the random effects and for the residuals. If 'Show Plot' is selected, the plots with the tumor growth curves are displayed along with the regression line for the fixed effect coefficients (i.e., the tumor growth rates for the control and treatment groups, Fig. 1d).

The orange-shaded lines in Fig. 1d indicate two individuals, mouse 2 and mouse 10, whose measurements are (intentionally) notably different from the others, and which will greatly affect the model and the results. The impact of these individuals and how to deal with them will be discussed later.

Figure 1. a, Upload dataset. b, Set time and name of treatments. c, Table with model estimates. d, Plots of tumor growth data. The arrow heads indicate subjects for which only four observations were made. The orange-shaded lines indicate individuals whose measurements are notably different from the others.

Minimum Observations per Subject
In Fig. 1b, there is an option called 'Minimum Observations per Subject', which allows for excluding samples with fewer than the specified number of observations. For example, in Figure 1d, the arrow heads indicate subjects for which only four observations were made. If the the minimum observations per subject is set to five, these samples will be omitted from the analysis (Fig 2).

Figure 2. a, Choosing subjects with at least five observations. b, Plots of tumor growth data from which the individuals with less than five observations have been removed.

Synergy Analysis

Once the model is fitted, the Synergy Analysis can be performed. Fig. 3a shows the different options for synergy analysis. The user can choose between Bliss, highest single agent (HSA), or response additivity (RA) reference models. In the advance options, selecting 'Use Robust Estimators' will apply sandwich-based cluster-robust variance estimators. This is recommended and can help to deal with possible model misspecifications.

The analysis for the response additivity reference model is based on simulations, and the number of simulations to run can also be specified by the user.

The output of the analysis includes a plot with the combination index (CI) and synergy score (SS) for each time point (Fig. 3b), as well as a table with all the results. The plot in Fig. 3b shows the estimated synergy metric (CI or SS) value, together with the confidence interval (vertical lines), and the color of the dots indicates the statistical significance.

The interpretation of the CI and SS values follows the typical definition of CI and SS:

  • CI: Provides information about the observed drug combination effect versus the expected additive effect according to the reference synergy model. A drug combination effect larger than the expected (CI > 1) indicates synergism, a drug combination effect equal to the expected (CI =1) indicates additivity, and a lower drug combination effect than the expected (CI < 1) indicates antagonism. The value of the CI represents the proportion of tumor cell survival in the drug combination group compared to the expected tumor cell survival according to the reference model.
  • SS: The synergy score is defined as the excess response due to drug interaction compared to the reference model. Following this definition, a SS>0, SS=0, and SS<0, represent synergistic, additive and antagonistic effects, respectively.
Figure 3. a, Side panel for running the synergy analysis. b, Plot of synergy results. Each dot represents the point estimate for the Combination Index (CI) and Synergy Score (SS), while the vertical gray lines represent the 95% confidence interval. The color of the dots reflects the minus log10 p-value for the synergy test. The horizontal dashed lines indicate the thresholds for assessing synergy, corresponding to 1 for the CI and 0 for the SS.

Synergy Results

As observed in Fig. 3b and the table with the results, the only time point for which there is a statistically significant result is time 3. At this time point, the CI is lower than 1, and the SS is higher than 0, indicating synergy in the drug combination group. However, at the other time points, the results show that the confidence interval includes 1 for the CI and 0 for the SS, and the p-value is higher than 0.05, indicating that the additive effect of the drug combination cannot be rejected.

Model Diagnostics

An important step in the analysis is to verify that the model is appropriate for the input data. In the Model Diagnostics tab, the user can check the main assumptions: the normality of the random effects and the residuals.

The model diagnostics output includes four plots to help evaluate the adequacy of the model. The two plots at the top of Fig. 4a show the Q-Q plot of the random effects and normalized residuals. Most points in these plots should fall along the diagonal line. The two plots at the bottom of Fig. 4a show scatter plots of the standardized residuals versus fitted values and standardized residuals per time and per treatment, which give information about variability of the residuals and possible outlier observations. The residuals in these plots should appear randomly distributed around the horizontal line, indicating homoscedasticity. Fig. 4b shows the output of Shapiro-Wilk normality tests for the random effects and the residuals. A non-significant p-value obtained by the Shapiro-Wilk normality indicates that there is no evidence to reject normalilty.

Figure 4. a, Q-Q plots of the random effects (top left), normalized residuals (top right), scatter plots of normalized residuals versus fitted values (bottom left), and normalized residuals per time and treatment group (bottom right). b, Results of Shapiro-Wilk's normality test for the random effects and normalized residuals.

The results from the plots and the normality tests seem to indicate some departure from the normality, especially evident for the residuals. This could be related to those individuals which has a large variation in the measurements.

Observations with absolute normalilzed residuals greater than the 1−0.05/2 quantile of the standard normal distribution are identified as potential outlier observations. A table with these observations is also provided by SynergyLMM. It can be observed that many observations from mouse 2 and mouse 10 have large residual values.

Potential Outlier Observations

As mentioned before, mouse 2 and mouse 10 are intentionally introduced outliers in the model, and there are many observations from these subjectes that are indeed identified as potential outliers. More details about outlier identification are provided in the Influential Diagnostics section. But first, there are several approaches that the user can follow to try to improve the model diagnostics.

If the diagnostic plots and tests show evident violations of the model assumptions, there are several solutions that may help improving the model:

    - Define the model using unequal variances for the errors. This can be done in the Advanced Options of the Model Estimation tab.
    - Transform the time units to improve the model fit and ensure that its assumptions are satisfied. For example, a square root or logarithmic transformation of the time units could help improving the model.
    - Carefully address potential outliers. Individuals or measurements highlited as potential outliers may warrant further investigation to reveal the reasons behind unusual growth behaviours, and potentially exclude these before re-analysis, after careful reporting and justification.

In this case, the model will be fitted especifying a unequal variance for the errors. Concretely, a different variance for each treatment is defined (Fig. 5a). It can be seen that this has improved the model diagnostics, and now both the random effects and normalized residuals are approximately normally distributed. There are still some potential outlier observations, which is normal, since there are two intentional outlier subjects.

Figure 5. a, Q-Q plots of the random effects (top left), normalized residuals (top right), scatter plots of normalized residuals versus fitted values (bottom left), and normalized residuals per time and treatment group (bottom right) of the corrected model. b, Results of Shapiro-Wilk's normality test for the random effects and normalized residuals.

Influential Diagnostics

The Influential Diagnostics tab is useful to identify subjects that have a great impact in the model and that may warrant more careful analysis. To this end, SynergyLMM uses Cook's distances and subject-specific log-likelihood displacements.

Cook's distances indicate the influence of each subject to the beta coefficient estimates of the model (i.e., tumor growth rate for the control and treatment groups). The greater the value, the higher the influence of the subject on the coefficient for its treatment group.

Log-likelihood displacements provide information about the difference in log-likelihood between a model fitted with all the subjects and a model fitted without one specific subject. The greater the value of the log-likelihood displacement for a subject, the greater the influence of that subject on the overall model's likelihood.

Fig. 6 shows results of the influential diagnostics of our example model. It can be seen how our two outlier subjects are identified: Mouse 10 is identified as having a great influence in the value of its treatment group estimated growth rate, while Mouse 2 is identified as having an important influence in the overall fit of the model.

Figure 6. a, Plot of the Cook's distances values versus subjects. The highlighted points indicate those individuals whose log-likelihood displacement is higher than the 90th percentile of the values. b, Plot of the log-likelihood-displacement values versus subjects. The highlighted points indicate those individuals whose log-likelihood displacement is higher than the 90th percentile of the values.

Model Performance

Another important tool for the diagnostics of the model is the plot of the observed vs. predicted values, available in the Model Performance tab. The output from this tab includes a table with model performance metrics, such as R-square and various error measures (Fig. 7a), along with the plot of observed vs. predicted values (Fig. 7b).

Each plot in Fig. 7b corresponds to a subject. The blue dots represent the actual measurements, the continuous line shows the group (treatment) marginal predicted values, and the dashed line indicates each individual's predicted values

By examining how closely the lines fit the data points, the user can assess how well the model fits the data. In this example, the model generally fits the data well, except for certain subjects, such as mouse 2 and mouse 10.

Figure 7. a, Table with some metrics about the model performance. b, Plots of observed vs. predicted values for each subject. The dots indicate the actual measurements, while the continuous and dashed lines show the marginal and subject-specific predicted values.

Power Analysis

Post Hoc Power Analysis

The Post Hoc Power Analysis allows users to check the statistical power of the dataset and fitted model for synergy analysis using the Bliss and HSA reference models.

The post hoc power calculation is based on simulations, and depending on the dataset, it can take some time to run. The power is returned as the proportion of simulations resulting in a significant synergy hypothesis test.

Fig. 8a shows the side panel for running the post hoc power analysis. The user can define the number of simulations to run, the p-value threshold for considering significance, and the time point for which to perform the analysis.

Figure 8. a, Side panel for running the post hoc power analysis. b, Results of the post hoc power analysis.

A Priori Power Analysis

The user can also perform three different types of a priori power analyses by modifying various experimental variables. To do this, 'SynergyLMM' creates exemplary datasets using the estimates from the fitted model in the Model Estimation tab. Then, several parameters can be modified and evaluated by the user, such as: the number of subjects per group, the days of measurements, the coefficients (tumor growth rates) for each treatment group, the standard deviation of the random effects (between-subject variance), and the standard deviation of the residuals (within-subject variance).

SynergyLMM then provides the a priori power calculation for each exemplary dataset defined by a specific set of experimental parameters.

Sample Size

The first option that can be tested is how the power would vary as the sample size per group changes. To do this, an exemplary dataset is used to fit models with the same values for all parameters (coefficients for the different treatment groups, standard deviations of the random effects and residuals, number of time points, etc.) as in the fitted model, but varying the sample size in each group. The user can select the range of sample sizes to evaluate and the reference model to use (Fig. 9a).

When clicking 'Run Sample Size Power Analysis', a plot with the exemplary data and another showing the power variation with the sample size are displayed (Fig. 9b). It can be confirmed that the values shown in the exemplary data plot are the same as the model estimates shown in Fig. 1c.

Figure 9. a, Side panel for performing power analysis by varying the sample size per group. b, Plot displaying the regression lines and values of the exemplary data based on the estimates from the fitted model. The curve illustrates how the statistical power varies with sample size per group, while the dashed line indicates the statistical power threshold of 0.8.

The results from the a priori power analysis show that to achieve a statistical power of 0.8, assuming the other parameters remain the same, the sample size per group should be at least 13.

Time Power Analysis

Another parameter that can affect statistical power is the duration of follow-up or the frequency of measurements. Given the results of a pilot experiment, the user may wonder how the results would change if tumor growth were monitored for a longer time period, or how often tumor growth measurements should be taken. To address this, the user can evaluate statistical power by varying the maximum follow-up time or the frequency of measurements, while keeping the other parameters fixed, as determined from the fitted model.

Maximum Time of Follow-up

First, it can be examined how power changes if the follow-up time is increased or decreased. To do this, select the analysis based on the maximum time of follow-up ('max'), as shown in Fig. 10a. The sample size per group, the time interval between measurements, and the reference model to use can be also specified. For this example, the analysis evaluated how power varied from 3 to 42 days of follow-up, with 5 subjects per group, and measurements taken every 3 days.

Figure 10. a, Side panel for conducting power analysis by varying the maximum time of follow-up. Note that the type of analysis is set to 'max'. b, Plot displaying the regression lines and values of the exemplary data, along with a curve illustrating how statistical power varies with maximum follow-up time. The dashed line indicates the statistical power threshold of 0.8. c, Side panel for performing power analysis by varying the frequency of measurements. Note that the type of analysis is set to 'freq.' d, Plot displaying the regression lines and values of the exemplary data, along with a curve showing how statistical power varies with the frequency of measurements. The dashed line indicates the statistical power threshold of 0.8.

In Fig. 10b, it can be seen how the power varies with the maximum time of follow-up. As expected, when the follow-up period is short, the statistical power is low, and it increases as the follow-up time increases, until it reaches a point where the power does not increase significantly.

Frequency of Measurements

Another question that researchers may have is how frequently the measurements should be taken. Of course, this would depend, among other factors, on practical considerations, but it can be simulated to estimate the ideal frequency of measurements.

Again, for this analysis, all the parameters are fixed, and the only variation is how frequently the measurements are taken, which is defined by the number of evenly spaced measurements performed during the follow-up period. For this example, the analysis evaluated how the power varied from 3 to 26 evenly spaced measurements, with 5 subjects per group and a follow-up period of 42 days (Fig. 10c).

As shown in Fig. 10d, the power increases as the number of measurements increases, although the difference is not dramatic.

Data Variability Power Analysis

Finally, another factor that can affect statistical power is the variability in the data. The variability in the model is represented by the standard deviation of the random effects (between-subject variability) and the standard deviation of the residuals (within-subject variability), which can be obtained from the table shown in Fig. 1c.

Additionally, the statistical power is also influenced by the magnitude of the differences: the greater the effect of the drug combination, the higher the statistical power to detect synergistic effects.

The user can test how statistical power varies by modifying the variability and the drug combination effects in the 'Variability Power Analysis' side panel (Fig. 11a).

In this case, the parameters that vary are the standard deviations of the random effects and the residuals, as well as the effect of the drug combination, represented by the coefficient of the drug combination group (i.e., tumor growth rate in the drug combination group). By default, these parameters are evaluated over a range from 10% to 200% of the values obtained from the model fitted in the Model Estimation tab, but they can be modified by checking the 'Show Advanced Options' box.

Figure 11. a, Side panel for performing power analysis by varying the variability and the effect of the drug combination. b, Plots showing the regression lines and values of the exemplary data, along with the statistical power as a function of the standard deviation (SD) of the random effects and residuals, as well as the coefficient for the drug combination group. The dashed line indicates the statistical power threshold of 0.8. The red dot and vertical dashed line indicate the power for the given SD of random effects and residuals, and growth rate given by the original model.

From the results shown in Fig. 11b, it can be observed how the power increases as the variability decreases and as the coefficient for the drug combination group decreases (i.e., the growth rate for the drug combination is smaller, indicating a higher drug combination effect).

Note: All a priori analyses are conducted with certain parameters held constant while varying others. This approach provides users with insights into optimal experimental conditions for achieving adequate statistical power. However, this method may not accurately reflect real-world scenarios, where changes to one variable can influence others. For instance, increasing or decreasing the sample size or the frequency of measurements can impact the variability in the data.
Input Table

Download Model Estimates Table

Download Plot



Download Plot


Download Synergy Results

Download Plot


              


              

Download Outliers Table


              
                
                Download Cook's Distances
              
              


              
                
                Download log-likelihood displacements
              
            

Download Plot


Download Plot

Download Plot

Download Plot