8.7版最新更新
2021年春季，nQuery发布了最新的8.7版本软件，继续帮助生物统计学家和临床研究人员实现更快的、更低成本的、更高成功率的临床试验。
In the Spring 2021 nQuery Advanced 8.7 release,5 new sample size tables have been added to the Pro tier of nQuery covering various new adaptive designs.
In this release the following areas are targeted for development:
 MultiArm MultiStage Designs (MAMs) GSD
 Two Stage Phase II Designs
 Generalized MCPMod (Multiple Comparisons Procedure  Modelling)
 Group Sequential Designs
MultiArm MultiStage Designs (MAMs) GSD
Adaptive designs, such as the group sequential design, have become widely used in clinical drug development. However, much of this development has been confined to the twoarm setting. A common situation in clinical trials is that in which an optimal treatment or dose has not been identified prior to the phase III trial. In this situation, it may be desirable to begin a Phase II trial (or longer Phase III trial) with several treatment arms and conduct interim analyses that will allow the dropping of less promising arms and potentially seamlessly move those arms into Phase III. This type of design is commonly named a multiarm multistage design.
The Multi Arm Multi Stage (MAMs) group sequential design extends the common twoarm group sequential design to a design in which multiple treatment arms can be compared to a common control arm and allows the trial to be stopped for efficacy or futility based on the best performing arm. In addition, it can be shown that less effective arms may be dropped at interim analyses without affecting the Type I error.
MAMs designs provide the ability to assess more treatments in less time than could be done with a series of twoarm trials and can offer significantly smaller sample size requirements when compared to that required for the equivalent number of twoarm trials.
In this release, we have extended our MAMs Group Sequential designs to include an option for the proportions endpoint. This provides a familiar framework with which to conduct a MAMs design and explore the additional flexibility offered.
The table added is as follows:
 Multiple Arms Multiple Stage (MAMs) Group Sequential Design for Proportions
Two Stage Phase II Designs
Phase II designs are often used to determine whether a new procedure or treatment is likely to meet a basic level of efficacy to warrant further development or evaluation. Phase IIa designs are focussed on the proofofconcept part of Phase II trials. They aim to show the potential efficacy and safety of a proposed treatment. Twostage designs are common to allow for flexibility to stop trials early for futility as Phase II is the most common failure point in drug evaluation. We have implemented two new twostage Phase II designs in this release.
The first follows the method outlined in Bryant and Day (1995). As mentioned, often evaluation of toxicity is listed as a primary objective in Phase IIa trials. The Bryant and Day design is a two stage phase IIa design with a coprimary endpoint which allows one to evaluate both the efficacy and the toxicity of the drug or treatment within a single design.
In addition to this design, we have also implemented an adaptive response design based on the paper by Lin & Shih (2004). The main benefit of the method put forward by Lin and Shih over methods such as Simon's twostage design, is the way in which the proportion of successes under which the alternative hypothesis is powered can be specified. This method allows the specification of two response proportions, the first a low or pessimistic value, and the second an optimistic or high value. This allows one to combat the uncertainty in the specification of the proportion of desirable results and so mitigates the risk of having a proportion which is too high, and so results in the study being underpowered, or a proportion which is too low, and so results in patient resources being unnecessarily wasted or the study taking an unnecessarily long time to complete.
The tables added are as follows:
 Two Stage Phase II Design for Response and Toxicity (Bryant and Day)
 Adaptive Response Two Stage Phase II Design (Lin & Shih's Design)
MCPMod (Multiple Comparisons Procedure  Modelling)
MCPMod (Multiple Comparisons Procedure  Modelling) is an increasingly popular statistical methodology for dosefinding Phase IIb trials. Since its development at Novartis, MCPMod promises to devise proofofconcept and doseranging trials with greater evidence and data that can prove critical for Phase III clinical trial design.
Combining the robustness of multiple comparisons procedures with the flexibility of modelling, MCPmod combines these methods to provide superior statistical evidence from Phase II trials with regards to dose selection with the FDA & EMA approving MCPMod as fitforpurpose (FFP).
In this release, we have extended our MCPMod design to include the case where there are unequal standard deviations for each arm in the study.
The table added is as follows:
 Multiple Comparisons Procedure  Modelling for Continuous Outcome (Unequal Variance) (MCPMod)
Group Sequential Tests for Noninferiority Two Means
Group sequential designs are a widely used type of adaptive trial in confirmatory Phase III clinical trials. Group sequential designs differ from a standard fixed term trial by allowing a trial to end early at prespecified interim analyses for efficacy or futility. Group sequential designs achieve this by using a flexible error spending method which allows a set amount of the total Type I or Type II error at each interim analysis. The group sequential design allows the trialist the flexibility to end those trials early which otherwise would have led to another large cohort of subjects to be analysed unnecessarily.
Noninferiority testing is used to test if a new treatment is noninferior to a standard treatment by a prespecified amount. This is a common objective in areas such as medical devices and generic drug development. This release sees one new design being implemented in this area.
The table added is as follows:
 Group Sequential Test for Noninferiority of Mean Difference
List of new Adaptive tables
Phase II Multistage Designs
 Two Stage Phase II Design for Response and Toxicity (Bryant and Day)
Proportion > Two  Test
 Multiple Arms Multiple Stage (MaMs) Group Sequential Design for Proportions
Group Sequential Designs
 Group Sequential Test for Noninferiority of Mean Difference
Means > Two  Test
 Multiple Comparisons Procedure  Modelling for Continuous Outcome (Unequal Variance)
Phase II Multistage Designs
 Oncology Two Stage Phase IIA Design
如何更新
要访问自适应模块，您必须订阅 nQuery Pro Tier。 如果你这样做了，那么 nQuery 应该会自动提示你更新。
您可以通过单击Help>Check for updates来手动更新 nQuery Advanced。
8.6版最新更新
The Summer 2020 (v8.6) release sees 5 new sample size tables added to nQuery Advanced PRO tier, covering various new adaptive designs.
In this release, the following areas have been targeted for development:
 MultiArm MultiStage Designs (MAMS) GSD for Means
 Generalized MCPMod (Multiple Comparisons Procedure  Modelling)
 Phase II Group Sequential Tests for Proportions (Fleming’s Design)
MultiArm MultiStage Designs (MAMS) GSD for Means
Adaptive designs, such as the group sequential design, have become widely used in clinical drug development. However, much of this development has been confined to the twoarm setting. A common situation in clinical trials is that in which an optimal treatment or dose has not been identified prior to the phase III trial. In this situation it may be desirable to begin a Phase II trial (or longer Phase III trial) with several treatment arms and conduct interim analyses that will allow the dropping of less promising arms and potentially seamlessly move those arms into Phase III. This type of design is commonly named a multiarm multistage design.
The Multi Arm Multi Stage (MAMS) group sequential design extends the common twoarm group sequential design to a design in which multiple treatment arms can be compared to a common control arm and allows the trial to be stopped for efficacy or futility based on the best performing arm. In addition, it can be shown that less effective arms may be dropped at interim analyses without affecting the Type I error.
MAMS designs provide the ability to assess more treatments in less time than could be done with a series of twoarm trials and can offer significantly smaller sample size requirements when compared to that required for the equivalent number of twoarm trials.
In this release, a MAMS Group Sequential design has been implemented for a continuous endpoint under normality. This provides a familiar framework with which to conduct a MAMS design and explore the additional flexibility offered.
The table added is as follows:
 Multiple Arms Multiple Stage (MAMS) Group Sequential Design for Means
Generalized MCPMod (Multiple Comparisons Procedure  Modelling
MCPMod (Multiple Comparisons Procedure  Modelling) is an increasingly popular statistical methodology for dosefinding Phase IIb trials. Since its development at Novartis, MCPMod promises to devise proofofconcept and doseranging trials with greater evidence and data that can prove critical for Phase III clinical trial design.
Combining the robustness of multiple comparisons procedures with the flexibility of modelling, MCPmod combines these methods to provide superior statistical evidence from Phase II trials with regards to dose selection with the FDA & EMA approving MCPMod as fitforpurpose (FFP).
The MCPMod methodology has been well developed for the case of normally distributed, homoscedastic data from a parallel group study design. In practice however many other types of endpoints are encountered and often in more complex settings like longitudinal studies for example. Generalized MCPMod extends the original MCPMod methodology to the context of general parametric models and for general study designs.
In this release, 3 tables will be added for generalized MCPMod. This extends the MCPMod methodology for a continuous endpoint (released in nQuery 8.5) to one for binary and count endpoints. Similar to the continuous MCPMod table, these new tables focus on the proofofconcept stage of the study.
The tables added are as follows:
 Multiple Comparisons Procedure  Modelling for Poisson Rates (MCPMod)
 Multiple Comparisons Procedure  Modelling for Negative Binomial Rates (MCPMod)
 Multiple Comparisons Procedure  Modelling for Binary Outcome (MCPMod)
Phase II Group Sequential Tests for Proportions (Fleming’s Design)
Typically, a phase II trial is designed to have a single stage in which a certain number of subjects are treated and the number of successes/responses are observed. However for ethical and economic reasons it may be of interest to stop the trial early, in particular for futility due to high failure rates at Phase II.
Based on the method described in Fleming (1982), the Phase II group sequential test is a onesided multiple testing procedure that allows for early termination of the study if interim results are extreme. This is done by testing the accrued data at a number of interim stages and looking at the point estimate for the proportion of successes or responses observed at interim points.
Under this design, at each of the interim analyses, the trial may be stopped for futility if the interim cumulative number of successes is below an acceptance point and the trial may be stopped for efficacy if the interim number of successes is above a rejection point. In addition to the ethical benefit, this type of design allows for greater flexibility and cost savings while retaining operating characteristics similar to a fixed term trial.
In this release, a table will be added for a Phase II Group Sequential Tests for One Proportion. This is commonly referred to as Fleming’s Design.
The table added is as follows:
 Group Sequential Test of One Proportion (Fleming’s Design)
8.4版最新更新
From fixed to flexible trial designs, the Summer 2019 nQuery release (Ver 8.4) sees the continued strengthening of nQuery to help Biostatisticians and clinical researchers save costs and reduce risk.
41 new sample size tables have been added in total.
With increasing costs and failure rates in drug development an increasingly important issue, adaptive trials offer one mechanism to alleviate these problems and make clinical trials better reflect the statistical and practical needs of trial sponsors. These issues have also spurred an increasing openness towards innovative clinical trial designs in regulatory agencies around the world.
Summer 2019 nQuery Release Notes  ver 8.4
What's new in the nQuery Adaptive module?
The Summer 2019 release extends the number of tables on offer for adaptive designs. 7 new tables will be added to this area.
In this release the following areas are targeted for development:
 Conditional Power and Predictive Power
 Unblinded Sample Size Reestimation
 Blinded Sample Size Reestimation
A background to these areas along with a list of the sample size tables which are added in the Summer release is given in the adjacent sections.
Conditional Power and Predictive Power
In group sequential designs and other adaptive designs, access to the interim data gives the ability to answer the important question of how likely a trial is to succeed based on the information accrued so far. The two most commonly cited statistics to evaluate this are conditional power and predictive power.
Conditional power is the probability that the trial will reject the null hypothesis at a subsequent look given the current test statistic and the assumed parameter values, which are usually assumed to equal their interim estimates. Predictive power (a.k.a. Bayesian Predictive Power) is the conditional power averaged over the posterior distribution of the effect size. Both give an indication of how promising a study is based on the interim data and can be used as adhoc measures for futility testing or for defining “promising” results for unblinded sample size reestimation.
Building on the initial nQuery Adapt release, 1 table will be added for conditional power and predictive power as follows:
 Conditional Power for 2x2 Crossover Design
Crossover trials use a repeated measures design, where each subject receives more than one treatment, with different treatments given in different time periods. The main benefit of a crossover trial is the removal of the between subject effect that impacts parallel trials. This can yield a more efficient use of resources as fewer subjects may be required in the crossover design than comparable designs.
Unblinded Sample Size Reestimation
In group sequential designs and other similar designs, access to the interim data provides the opportunity to improve a study to better reflect the updated understanding of the study. One way a group sequential design can use the interim effect size estimate is not only to decide whether or not to stop a trial early but to increase the sample size if the interim effect size is considered “promising”. This optionality gives the trialist the chance to initially power for a more optimistic effect size, thus reducing upfront costs, while still being confident of being able to find for a smaller but clinically relevant effect size by increasing sample size if needed.
The most common way to define whether an interim effect size is promising is conditional power. Conditional power is the probability that the trial will reject the null hypothesis at a subsequent look given the current test statistic and the assumed parameter values, which are usually assumed to equal their interim estimates. For “promising” trials where the conditional power falls above a lower bound, a typical value would be 50%, the initial target power of the sample size can be increased to make the conditional power equal the target study power.
Building on the initial nQuery Adapt release, the following table will be added for unblinded sample size reestimation:
 Interim Monitoring and Unblinded Sample Size Reestimation for Survival
This table allows nQuery Adapt users to extend their initial group sequential design for survival in two groups using the LogRank test (with or without unequal followup) by giving tools which allow users to conduct interim monitoring and conduct a flexible sample size reestimate at a specified interim look.
This table will be accessible by designing a study using either of the two group sequential designs for survival tables and using the “Interim Monitoring & Sample Size Reestimation” option from the group sequential “Looks” table. This table will provide for two common approaches to unblinded sample size reestimation: ChenDeMetsLan and CuiHungWang. There is also an option to ignore the sample size reestimation and conduct interim monitoring for standard group sequential design.
The ChenDeMetsLan method allows a sample size increase while using the standard group sequential unweighted Wald statistics without appreciable error inflation, assuming an interim result has sufficiently "promising" conditional power. The primary advantages of the ChenDeMetsLan method are being able to use the standard group sequential test statistics and that each subject will be weighted equally to the equivalent group sequential design after a sample size increase. However, this design is restricted to the final interim analysis and Type I error control is expected but not guaranteed depending on the sample size reestimation rules.
The CuiHungWang method uses a weighted test statistic, using preset weights based on the initial sample size and the incremental interim test statistics, which strictly controls the type I error. However, this statistic will differ from that of a standard group sequential design after a sample size increase and since subjects are weighted on the initial sample size, those subjects in the postsample size increase cohort will be weighted less than those before.
There will be full control over the rules for the sample size reestimation including sample size reestimation look (for CuiHungWang), maximum sample size, whether to increase to the maximum sample size or the sample size to achieve the target conditional power and bounds for what a “promising” conditional power is, among others.
Blinded Sample Size Reestimation
Sample size determination always requires a level of uncertainty over the assumptions made to find the appropriate sample size. Many of these assumed values are for nuisance parameters which are not directly related to the effect size. Thus it would useful to have a better estimate for these values than relying on external sources or the cost of a separate pilot study but without the additional regulatory and logistical costs of using unblinded interim data. Blinded sample size reestimation allows the estimation of improved estimates for these nuisance parameters without unblinding the study.
In the Summer 2019 release, five tables will be added for blinded sample size reestimation using the internal pilot method. The internal pilot method assigns an initial cohort of subjects as the “pilot study” and then calculates an updated value for a nuisance parameter of interest. This updated nuisance parameter value is then used to increase the study sample size if required, with the final analysis conducted with standard fixed term analyses with the internal pilot data included.
The new additions to the Adapt Module expand the scope of the nQuery Adapt blinded sample size reestimation tables to the cases where unequal sample sizes and continuity corrections are needed. The new tables will be as follows:
 Blinded Sample Size Reestimation for Two Sample ttest for Inequality (common variance, unequal n's)
 Blinded Sample Size Reestimation for Two Sample ttest for Noninferiority (unequal n's)
 Blinded Sample Size Reestimation for Two Sample ttest for Equivalence (unequal n's)
 Blinded Internal Pilot Sample Size Reestimation for Two Sample χ2 Test for Noninferiority (Continuity Corrected)
 Blinded Internal Pilot Sample Size Reestimation for Two Sample χ2 Test for Inequality (Continuity Corrected)
These tables will provide full flexibility over the size of the internal pilot study, whether sample size decreases are allowable in addition to increase and tools to derive the bestblinded estimate from the internal pilot.
Blinded sample size reestimation for the two sample ttest updates the sample size based on a blinded estimate of the common withingroup standard deviation. Three methods are available to estimate the withingroup standard deviation from the internal pilot data: pilot standard deviation, biasadjusted pilot standard deviation, upper confidence limit for pilot standard deviation.
Blinded sample size reestimation for the two sample chisquared test updates the sample size based on a blinded estimate of the total proportion of successes and combining this with the initial proportion difference estimate. The user can enter either the proportion of successes or number of successes for the equivalent analysis.
List of New Adapt Tables
Blinded Sample Size Reestimation
 MTT24U Blinded Sample Size Reestimation for Two Sample ttest for Inequality (common variance, unequal n's)
 MTE28U Blinded Sample Size Reestimation for Two Sample ttest for Noninferiority (unequal n's)
 MTE29U Blinded Sample Size Reestimation for Two Sample ttest for Equivalence (unequal n's)
 PTE12 Blinded Internal Pilot Sample Size Reestimation for Two Sample χ2 Test for Noninferiority (Continuity Corrected)
 PTT27 Blinded Internal Pilot Sample Size Reestimation for Two Sample χ2 Test for Inequality (Continuity Corrected)
Conditional Power and Predictive Power
 MTT38 Conditional Power for 2x2 Crossover Design
Unblinded Sample Size Reestimation
 STT17 Unblinded Sample Size Reestimation and Interim Monitoring for Two Survival
How to Update
To access the adaptive module you must have a nQuery Advanced Pro subscription. If you do, then nQuery should automatically prompt you to update.
You can manually update nQuery Advanced by clicking Help>Check for updates.
nQuery Advanced 8.2最新更新版本说明
Release Notes
nQuery Advanced 8.2  April 2018 Update
The nQuery April 2018 release will add a wide range of sample size tables ranging from extensions of preexisting tables for a better and clearer user experience to the those based on the latest academic research and user feedback.
In the April 2018 release, we will be adding 52 new sample size tables to nQuery Advanced and 20 new tables to nQuery Bayes. This release summary will provide an overview of what areas have been targeted in this release along with the full list of tables being added.
nQuery Advanced Tables
In the April 2018 release, three main overarching areas were targeted for significant improvement. These were:
 Epidemiological Methods
 Noninferiority and Equivalence Tests
 Correlation and Diagnostic Testing (ROC) Methods
There is also a number of tables which do not fall into these categories related to areas such as inequality testing for lognormal data, testing of variance and standard deviations and nonparametric tests. These are described at the end of this document.
We will provide background on each of these below and a list of the sample size tables which will be added in the April release. References for each method are provided at the end of this article
1. Epidemiology
Epidemiology is the branch of medicine which primarily studies the incidence, distribution, and possible control of diseases and other factors relating to health. Epidemiological studies are cornerstone of research into areas such as public health, health policy or preventative medicine (e.g. vaccines).
Due to often having to study the effect of medicines and interventions at a more complex societywide level, processes and methods for epidemiology often adjust for problems that are less prominent in wellcontrolled clinical trials. These issues include being unable to individually randomise, relying on observational data or attempting to extract causal relationships from highly complex data.
Due to these and other issues, the study designs and statistical methods used by epidemiologists will often have to include adjustments for exogenous effects and have a greater reliance on processes such as pairmatching. While statistical methods for clinical trials provide a useful starting point for getting adequate sample size estimates, there is a growing desire for methods which have found traction in the epidemiological field.
In the April release, 12 new tables will be added with the main areas of focus in the Epidemiology upgrade being the following:
 CaseControl Studies (Observational and Prospective)
 Vaccine Efficacy Studies
 Cluster Randomized Trials (CRT)
 Mendelian Randomization Studies
These areas and the tables in each category are explored below.
CaseControl Studies
CaseControl studies are those where the analysis assumes that the effect of a treatment or intervention or prognostic factor can be modelled by comparing the effect on a paired cases and controls. In the epidemiological context, this is most commonly associated with retrospective studies attempting to find a relationship between a risk factor and a disease of interest (e.g. effect of smoking on lung cancer rates) using preexisting sources such as health databases.
In this context, the nQuery April 2018 release adds an additional four tables which should add additional flexibility when planning a casecontrol study using nQuery. These are:
 Test for Binary Covariate in Logistic Regression
 CaseControl Test with M Controls per Case (McNemar Extension)
 Conditional Logistic Regression with Binary Risk Factor
 Conditional Logistic Regression with Continuous Risk Factor
These tables complement our preexisting nQuery tables for chisquared tests, exact tests, correlated proportions and logistic regression and add options for conditional logistic regression to allow for greater flexibility when exploring sample estimates for casecontrol studies.
Vaccine Efficacy
Vaccine efficacy studies face significant challenges compared to other clinical trials. These include having a much larger scale due often being nation or region wide campaigns, dealing with rare diseases or conditions and the challenges of doing work in the field rather than in a fully controlled setting. For reasons such as these, vaccine efficacy designs and statistical methods have developed their own approaches and terminology to help the relevant researchers and public or private bodies of interest.
For vaccine efficacy, the nQuery April 2018 release adds two additional tables tailored for finding the sample size for the precision of an estimate of the vaccine efficacy. These are:
 Confidence Interval for Vaccine Efficacy in a Cohort Study
 Confidence Interval for Vaccine Efficacy in a CaseControl Study
In conjunction with our wide range of preexisting tables for binomial proportions and survival rates, these tables will give researchers in vaccine research more tailored options for their study.
Cluster Randomized Trials
Cluster randomized trials are studies where the unit of randomization is a cluster or group rather than the individual subject. This is a common design when there are natural blocks or clusters such as schools or hospitals. By assigning the same treatment to all subjects within a cluster, the administrative and financial cost of field trials can be reduced significantly. For this reason and others, this design type is very commonly seen in public health policy studies.
For cluster randomized trials, the nQuery April 2018 release includes four additional tables which expand upon our preexisting options for cluster randomized trials. These are:
 Cluster Randomized Trial for Inequality Test of Two Means (unequal sample sizes)
 Cluster Randomized Trial for Noninferiority Test of Two Means
 Cluster Randomized Trial for Equivalence Test of Two Means
 Cluster Randomized Trial for Superiority by a Margin of Two Means
These options expand upon our preexisting tables for cluster randomized trials comparing means, proportions, incidence rates and survival curves and for alternative cluster randomized trials such as the matchedpair design.
Mendelian Randomization
Mendelian Randomization is form of randomization which takes advantage of the growing availability and understanding of genetic information to make causal claims about potential treatments without using the common fully randomized approach. By using well characterised relationships between genes and phenotypes with a known secondary effect on an outcome of interest, mendelian randomization offers the opportunity to use genetic information as a instrumental variable to find the causal relationship between a risk factor of interest and a disease outcome.
For studies which use Mendelian Randomization, the nQuery April 2018 release provides two new tables. These are:
 Mendelian Randomization for Continuous Outcome
 Mendelian Randomization for Binary Outcome
These provide the first tables in nQuery which account for this novel design and innovative approaches such as this will be of active interest in the area
2. Noninferiority and Equivalence Testing
Noninferiority and equivalence testing are used to statistically evaluate how similar a proposed treatment is to a preexisting standard treatment. This is a very common objective in areas such as generics and medical devices. This is particularly important if using a placebo group would be required otherwise.
As noninferiority and equivalence testing will typically involve evaluation against a welldefined treatment (e.g. RLD), there is a lower incidence of the large parallel studies typically seen in Phase III clinical trials. Onesample, paired samples or crossover designs are common as these will generally require a lower cost and sample size.
In the April release, we will be adding an additional 20 (22 in the CRT Means reference above are included) tables for noninferiority and equivalence testing. These are focused on expanding the options available for the noninferiority and equivalence testing of continuous data, binary data and incidence rates. The focus areas are as follows:
 Continuous Outcome Studies
 Binary Outcome Studies
 Incidence Rate Outcome Studies
These areas and the tables in each category are explored below.
Continuous Outcome Studies
In the context of noninferiority and equivalence testing, the comparison of continuous outcomes using means is the most common situation to encounter. A wide range of design types and statistical methods are available for comparing this type of data depending on the assumptions and constraints relevant to proposed study. Common design types in this context would be onesample, paired, crossover and parallel studies. The most common statistical methods are based on assuming either that the data is normally distributed and comparing the difference in means (Additive model) or that the data is lognormally distributed and analysing the ratio of (geometric) means (Multiplicative model).
For noninferiority and equivalence testing of continuous data, the nQuery April 2018 release adds an additional 12 tables. These are as follows:
 Noninferiority for One Normal Mean
 Noninferiority for One LogNormal Mean
 Noninferiority for Paired Means Ratio
 Equivalence for One Mean
 Equivalence for Paired Means
 Equivalence for One LogNormal Mean
 Equivalence for Paired Means Ratio
 Noninferiority for crossover design
 Noninferiority for Twosample Mean Ratio on Logscale
 Noninferiority for Crossover Mean Ratio on Logscale
 Noninferiority for Twosample Mean Ratio on Original Scale
 Noninferiority for Crossover ratio on Original Scale
These tables expand upon the large number of preexisting tables for noninferiority and equivalence testing for means to give the largest number of options available to find the sample size.
Binary Outcome Studies
In the context of noninferiority and equivalence testing, the comparison of binary is less common but has grown in more recently as additional statistical methods have become popularised. Common design types in this context would be onesample, paired and parallel studies. In the context of binary data, one of the most noticeable aspects is wide variety of options available ranging from relatively simple normal approximation tests to more complex exact methods and sample size methods have followed this trend in regard to binary data analyses generally.
For noninferiority and equivalence testing of binary data, the nQuery April 2018 release adds an additional 2 tables. These are as follows:
 Noninferiority for a Single Binary Proportion
 Equivalence Test for Two Independent Proportions
Note that both these tables integrate more exact binomial enumeration methods as an option in addition to the typical normal approximation methods. They also include options for a wide range of proposed statistics with the main categories being chisquared tests (including option for continuity correction), Z and ttest approximations and several likelihood score statistics (Miettinen and Nurminen, Gart & Nam, Farrington and Manning). These tables expand upon the large number preexisting tables for the noninferiority and equivalence testing of binary proportions.
Incidence Rates Studies
In the context of noninferiority and equivalence testing, the comparison of incidence rates is a relatively uncommon scenario. Incidence rates are where the outcome of interest is the number of events which occur on average in a given time period (i.e. the event rate). The wider availability of software to analyse incidence rates directly rather than relying on normal approximations has seen a growth of interest in methods such as Poisson and Negative Binomial regression. This has naturally extended to the case of noninferiority and equivalence testing of incidence rate data. The timedependent nature of incidence rates means that models can integrate greater flexibility for time dependencies and this is reflected in the rapidly growing literature for sample size in the area.
For noninferiority and equivalence testing of incidence rates data, the nQuery April 2018 release adds 6 tables. These are as follows:
 Noninferiority for Two Rates using Poisson Model
 Equivalence for Two Rates using Poisson Model
 Equivalence for Negative Binomial Model (Equal Followup, Dispersion)
 Noninferiority for Negative Binomial Model (Equal Followup, Dispersion)
 Equivalence for Negative Binomial Model (Unequal Followup, Dispersion)
 Noninferiority for Negative Binomial (Unequal Followup, Dispersion)
These tables expand upon the preexisting options for analysing incidence rate data in the context of inequality testing. These methods represent the latest research and in the last two tables can integrate the effects of dispersion and unequal followup on the sample size estimate.
3. Correlation and Diagnostic Testing (ROC) Methods
Correlation, agreement and ROC methods are interested in characterising the strength of the relationship between a predictor (e.g. presence of treatment) and outcome (disease progression) of interest. These measures are often used in conjunction with models and statistical testing to characterise the nature of the relationship of interest. These methods are of interest when attempting to communicate the strength of a model or a relationship.
These types of measures are seen throughout statistical practise but are particularly prominent in areas such as diagnostic testing, the social sciences and biomarker studies.
In the nQuery April release, we will be adding an additional 9 additional tables in this area which fall in the following main categories:
 Correlation and Agreement Measures
 Diagnostic Screening Measures
These are summarised below.
Correlation and Agreement Measures
Correlation measures are used to characterise the strength of relationship between continuous and/or ordinal outcomes and measures such as Pearson’s correlation are ubiquitous in statistical practise. Agreement measures are used to analyse the strength of the ability of more than one rater (e.g. tester or test) to agree and correctly diagnose the condition of one or more entities (e.g. subject disease status). Both of these are common outcomes of interest in a wide variety of settings.
Due to the ubiquity of these methods, a wide range of measures have been proposed to adjust for scenarios which diverge from the most common correlation and agreement measures (e.g. Pearson correlation and Cohen’s Kappa). Common complications adjusted for are the presence of ordinal instead of continuous variables or divergences from common distributional assumptions (e.g. Normal)
In the nQuery April release, we are adding four additional options in this area. These are as follows:
 Confidence Interval for KendallTau Correlation
 Confidence Interval for Spearman Correlation
 Test for One Intracluster Correlation
 Test for One Cronbach Alpha
These add to the options available for other common correlation and agreement measures such as the Pearson Correlation, Lin’s Concordance Coefficient and Cohen’s Kappa Coefficient.
Diagnostic Screening Measures
Diagnostic screening measures are very common in clinical research. These measures are used to assess the performance of a diagnostic test to accurately predict a condition of interest in the population(s) of interest. Areas where this type of analysis have become particularly popular are biomarker studies, machine learning and predictive genetic tests.
Commonly, this strength is characterised by the Area under the Curve (AUC) of the Receiving Operating Curve (ROC) which provides a useful summary measure of screening performance over all potential cutoff points for a screening measure. However, a large number of other statistics may be of interest at specific cutoffs such as sensitivity (a.k.a. recall), specificity and positive predictive value (PPV, a.k.a. precision), among other.
In the nQuery April release, we are adding 5 new tables in this area. These are as follows:
 Confidence Interval for Area under the Curve (AUC)
 Confidence Interval for One Sensitivity
 Confidence Interval for One Specificity
 Simultaneous Interval for One Sensitivity and One Specificity
 Test for Paired Sensitivity
These add to the preexisting tables already present in nQuery for the testing of AUC values under varying design types and sensitivity and specificity.
4. Miscellaneous Tables
In the April release of the nQuery, 11 tables do not fit into the above categorisations. These cover areas such as the testing of lognormal means, the testing of variances and standard deviations and nonparametric tests. These tables are as follows:
 Confidence Interval for One Variance
 Confidence Interval for One Variance using Tolerance Probability
 Confidence Interval for One Variance using Relative Error
 Confidence Interval for One Standard Deviation
 Confidence Interval for One Standard Deviation using Tolerance Probability
 Confidence Interval for One Standard Deviation using Relative Error
 Confidence Interval for Ratio of Two Variances
 Confidence Interval for Ratio of Two Variances using Relative Error
 One Sample ttest for LogNormal data
 Paired ttest for Mean Ratio (logscale)
 One Sample/Paired Sample Wilcoxon SignRank Test
These options expand upon nQuery’s preexisting options in these areas.
