- Open Access
Missing steps in a staircase: a qualitative study of the perspectives of key stakeholders on the use of adaptive designs in confirmatory trials
Trials volume 16, Article number: 430 (2015)
Despite the promising benefits of adaptive designs (ADs), their routine use, especially in confirmatory trials, is lagging behind the prominence given to them in the statistical literature. Much of the previous research to understand barriers and potential facilitators to the use of ADs has been driven from a pharmaceutical drug development perspective, with little focus on trials in the public sector. In this paper, we explore key stakeholders’ experiences, perceptions and views on barriers and facilitators to the use of ADs in publicly funded confirmatory trials.
Semi-structured, in-depth interviews of key stakeholders in clinical trials research (CTU directors, funding board and panel members, statisticians, regulators, chief investigators, data monitoring committee members and health economists) were conducted through telephone or face-to-face sessions, predominantly in the UK. We purposively selected participants sequentially to optimise maximum variation in views and experiences. We employed the framework approach to analyse the qualitative data.
We interviewed 27 participants. We found some of the perceived barriers to be: lack of knowledge and experience coupled with paucity of case studies, lack of applied training, degree of reluctance to use ADs, lack of bridge funding and time to support design work, lack of statistical expertise, some anxiety about the impact of early trial stopping on researchers’ employment contracts, lack of understanding of acceptable scope of ADs and when ADs are appropriate, and statistical and practical complexities. Reluctance to use ADs seemed to be influenced by: therapeutic area, unfamiliarity, concerns about their robustness in decision-making and acceptability of findings to change practice, perceived complexities and proposed type of AD, among others.
There are still considerable multifaceted, individual and organisational obstacles to be addressed to improve uptake, and successful implementation of ADs when appropriate. Nevertheless, inferred positive change in attitudes and receptiveness towards the appropriate use of ADs by public funders are supportive and are a stepping stone for the future utilisation of ADs by researchers.
Traditionally, standard randomised controlled trials (RCTs) are designed with a fixed target sample size and recruit until this target is met. Recently, much attention has been paid to alternative types of RCTs, known as adaptive designs (ADs), in which prospectively planned modifications to the design are made based on accruing outcome data from an ongoing trial while preserving the scientific validity and integrity of that trial [1, 2]. This may mitigate the risk of making inaccurate design assumptions or potentially may shorten trial duration by allowing early stopping as soon as there is sufficient evidence to answer the research question(s) . However, despite potential promising benefits to clinical trials, patients and funders, the use of ADs in practice, particularly in the public sector, has been described by advocates as disappointing - with their uptake lagging far behind methodological developments . Moreover, the use of ADs is viewed as controversial, with the perception among some stakeholders that public funders and regulators have hindered their wider adoption .
Citing disappointing uptake, the pharmaceutical industry initiated a Pharmaceutical Research and Manufacturers of America (PhRMA) Adaptive Design Working Group with a vision to facilitate dialogue among key stakeholders in drug development and to establish a consensus position on the use of ADs . The group further investigated barriers and opportunities associated with the use of ADs across different trial phases, specifically in drug development. Although much related discussion and research has subsequently been undertaken [2, 5–10], it has been led and driven by the pharmaceutical industry, especially in the USA, with the public sector lagging behind.
Researchers have highlighted that the public sector has its own unique multifaceted challenges, which need to be explored in detail and addressed in order to improve uptake of ADs [2, 11, 12]. With this in mind, the NIH (National Institutes of Health, USA) and associates funded and facilitated a 2-day workshop to initiate some cross-industry discussions with representatives from the NIH, the Food and Drug Administration (FDA), the European Medicines Agency (EMA), the pharmaceutical industry, non-profit foundations, patient representatives and academia . Some recommendations have since been drawn up to enhance the use of ADs [11, 12]. Although this has been a significant milestone, the NIH workshop did not explore the perceptions and attitudes towards ADs of key stakeholders directly involved in the day-to-day conduct of clinical trials. Furthermore, some of the NIH findings may not be directly extrapolated to the UK setting, due to differences such as public funding and clinical trials infrastructure, capacity issues and underlying perceptions.
Little research has been undertaken to explore the use of ADs in the publicly funded confirmatory setting, particularly in the UK. In 2012, Morgan and colleagues  investigated the use of ADs and associated perceptions of barriers in the private and academic sector through a survey. The authors found change management, regulatory acceptance, lack of education and extra time and resources required for planning as major perceived barriers. Jaki  also investigated the use of ADs and Bayesian methods in early phase trials through a cross-sectional survey of registered UK Clinical Trials Units (CTUs), predominantly surveying statisticians. The poor application of these methods was attributed to five key barriers: lack of software, clinical investigators insisting on preferred methods, lack of expertise, inadequate funding structure and time required for trial design. These researches demonstrated the existence of barriers impeding the use of ADs. We have endeavoured to fill the gap in previous research [2, 3, 8, 9, 11–14] by incorporating nested qualitative interviews of key stakeholders with diverse roles in clinical trials research with a focus on publicly funded confirmatory trials prior to subsequent related surveys to be reported elsewhere.
This study is motivated by the belief that further related research and discussions are needed in the UK publicly funded confirmatory setting. Understanding perceptions towards ADs by researchers and decision-makers is key to unlocking potential benefits of ADs in this setting. We therefore aimed to explore key stakeholders’ experiences, perceptions and attitudes towards ADs in publicly funded confirmatory trials and their views on barriers and facilitators to the use of ADs. We believe our findings will inform researchers and decision-makers on key issues, in order to facilitate their preparedness to utilise acceptable ADs in publicly funded confirmatory trials where appropriate.
Study design and setting
This study valued the importance of understanding obstacles to AD use from the point of view of key stakeholders’ experiences, perceptions and attitudes, in order to generate facilitators to unlock barriers to appropriate use. This approach, which explores views, meaning and context, can be viewed within the phenomenological paradigm . Hence, we conducted cross-sectional, in-depth, semi-structured, one-to-one qualitative interviews of key stakeholders involved in clinical trials research . This approach encouraged participants to talk about pertinent issues about ADs through the use of open-ended questions. Some of these questions were a priori-designed based on topics from previous literature [7–9, 11, 14], and others were informed by researcher-driven hypotheses. Although we paid attention to the UK publicly funded setting, a cross-sector approach was undertaken by including participants with private sector experiences in order to explore diverse experiences, perceptions and attitudes. In particular, we purposively sought expertise in the private sector due to a perceived greater experience of ADs . In addition, four international participants were included in our sample in response to advice given by some participants during the interviews. We conducted interviews by telephone or through face-to-face conversations based on feasibility and the need to reach out to a wider geographical area of participants of interest.
Most qualitative studies base their sample size on reaching data saturation, which is unknown in advance because it depends on various factors such as: the scope and nature of the research subject, study design and resources available [17–20]. Some authors recommend up to 10 homogeneous interviews for phenomenological research . Bearing this and time constraints in mind, we intended to recruit six to eight participants per expertise category, yielding a minimum of 20 participants depending on the degree of overlap in expertise. We also adapted our sampling in some expertise categories guided by richness of information from previous interviews and the need for further exploration of certain phenomena. Overlapping of participant roles afforded an opportunity to explore wider views and experiences with a smaller sample.
Selection of participants
We purposively selected participants in a consecutive manner following informed consent agreement if they met the desired core duties and responsibilities in trials research. We adopted this cross-disciplinary approach to optimise maximum variation to capture diverse views and experiences . Core expertise for purposive sampling were UK CTU leaders (directors or deputy directors), public funding panel and board members (chairs or vice chairs including other ordinary members), Independent Data Monitoring Committee (IDMC) members, regulators, statisticians, health economists and chief investigators. We sent an invitation letter with an information sheet to target participants using various platforms; mass emailing to specialist network groups including the UK CRC Registered CTU Network  and the MRC Network of Hubs for Trial Methodology Research ; and personalised emails to referred contacts and hard to reach groups, such as private sector, regulators and public funding panel and board members.
We phrased the invitation letter to emphasise that participants would be eligible to participate regardless of their underlying experiences, perceptions and attitudes towards ADs in order to minimise potential responder bias due to oversampling of participants likely to express positive views. We asked responders to complete a short questionnaire detailing their demographic characteristics and key expertise and to return it with their signed informed consent form. We then sequentially selected participants until reaching the desired target sample size. We used interview guides tailored for participants’ expertise to prompt questions (see Additional file 1). We undertook five internal pilot interviews, four of which were face-to-face to test the appropriateness of the interview guides, prompts and interview duration. On completion of the interviews, we gave participants an option to verify their interview transcript and also to say anything relevant about ADs that they felt was not covered but worth contributing.
The lead author (MD) conducted the interviews, which were audio recorded and verbatim transcribed by experienced in-house transcribers. A favourable ethical opinion (0676) was received from the Research Ethics Committee of the School of Health and Related Research at the University of Sheffield, and all interviews were conducted with signed informed consent.
Analysis and reporting
Data were entered into NVivo10 , which was used to manage and organise the data analysis process. We employed the framework method [25, 26] to structure the analytical process, which includes the following key stages; familiarisation and annotation of transcripts, identifying a thematic framework , indexing, charting, mapping, and interpretation [28–30]. Mapping helped to identify relationships and clusters around themes, thereby facilitating understanding, communication and interpretation. Themes captured what is most important from the data in understanding views and experiences concerning the use of ADs. We adapted taxonomies developed and used by other authors in the field of evidence-based practice to classify barriers to the uptake of ADs into micro- and macro-level domains pertinent to key stakeholders at the individual and organisational level [31–33]. Some potential facilitators to perceived barriers are also presented. We paid attention to emerging themes and contrasted these with the existing literature. Supplementary interview data and case studies to support themes are provided (see Additional file 2). We utilised the COREQ checklist to guide the conduct, analysis and reporting of this study .
Description of participants
In total, 27 participants were interviewed between March and August 2014, predominantly based on a sampling frame of 45 registered UK CTUs (2012/2013). Previous AD experience of participants is shown in Table 1. One health economist agreed to be interviewed. Reasons for not taking part among 17 health economists who were directly invited were unfamiliarity with ADs (n = 5), busy schedule (n = 1), non-response (n = 10) and willing but incompatible schedule (n = 1).
Interviews were conducted by telephone (n = 17), face-to-face (n = 7), skype video call (n = 2) and skype telephone call (n = 1). Median duration (IQR) of interviews was 31 (26 to 38) minutes ranging from 13 to 51. Participants’ characteristics, demographics and diverse overlapping primary duties and responsibilities in clinical trials research are displayed in Tables 1 and 2.
Perceived value of ADs and opportunity for use in confirmatory trials
Perceived advantages of ADs
Participants stated potential advantages of ADs, which can be broadly classified into three categories: ethical benefits to patients, improving design efficiency to answer research question(s), and value for money. These perceived advantages are summarised with supporting data in Table 3 and Additional file 2.
Perceived therapeutic areas of opportunity to use ADs
In principle, most participants acknowledged that ADs could be applicable across a wide spectrum of therapeutic areas (see Additional file 2). However, some participants also believed ADs may be more appealing or appropriate for certain health conditions or populations, due to factors such as severity of the health condition, availability of standard care options and limitations of standard methods (such as in small populations).
Some participants mentioned the potential value of ADs in areas including, but not limited to oncology; emergency medicine; and respiratory, cardiovascular, infectious and rare or orphan diseases. Participants stated that the nature of the clinical outcome(s) of interest and study intervention may also influence relevance of ADs. A case study was shared on how ADs could be of value in evaluating interventions during outbreaks of rapidly evolving and fatal pandemics such as influenza or Ebola, due to the severity of the conditions, coupled with the need for urgent clinical decision-making by policymakers (see Additional file 2, Case Study A). Most importantly, participants stated that it is imperative that the proposed AD is fit for the purpose of robustly answering the research question(s) accompanied by a clear rationale.
Perceived types of ADs with potential in confirmatory setting
Participants mentioned that the following types of ADs have useful potential within the confirmatory setting, which are also reflected in the literature [1, 2, 8, 9, 35–38]. However, the receptiveness towards these ADs by the research community and policymakers varied considerably depending, for instance, on the type and scope of the adaptation. These types include the following:
Sample size re-estimation (SSR) conducted either in a blinded or unblinded manner.
Futility assessment based on stochastic curtailment (such as conditional power)
Standard group sequential design (GSD) allowing for early stopping (such as for futility and/or safety and/or efficacy and/or non-inferiority).
Strictly phase 3 multi-arm multi-stage design (MAMS) with treatment selection allowing for dropping of inferior or futile treatments and/or early trial stopping.
Operational or inferential seamless 2/3 design with treatment selection in phase 2.
Response adaptive randomisation (strictly based on primary outcome data).
Public funders’ perceived change in attitudes towards ADs
We inferred a paradigm shift in attitudes towards ADs by public funders, mainly motivated by value for money and the desire to make use of the public purse more efficiently. Funders expressed a clear positive will and receptiveness to fund adaptive trials and to encourage researchers to utilise ADs, whenever appropriate, to answer the research question(s). This view, supported by various initiatives relating to ADs funding opportunities (such as training fellowships and grant calls) [39–41], was acknowledged and welcomed by researchers.
‘I think generally speaking we are receptive to those ideas (of ADs) and in fact we, at [organisation] have held our own workshops on ADs last year or the year before in order to try and promote more use of ADs providing they are appropriate of course. So I think 10 years ago our attitudes were more towards traditional parallel group, it was a sort of traditional well-known pathway but I think now our modern thinking is that we welcome ADs when appropriate and it is very much for the applicants to make the case for why they want maybe four arms with interim analyses for dropping arms …’ (QL35 Chair - Public Funder)
Regulatory receptiveness and improving awareness and experiences
In principle, there appeared to be regulatory receptiveness towards the concept of ADs, but this was conditional on strong caveats, particularly in confirmatory trials on measures such as minimisation of operational bias during the conduct to preserve trial credibility and integrity, control of type I error, and use of appropriate statistical inference. These caveats and considerations are highlighted in regulatory guidance and reflection papers specific to drugs and biologics [42–44]. We inferred that regulatory awareness and experience is growing, particularly among statistical assessors due to the increasing number of AD-related scientific advice consultations and applications by researchers; especially on SSR, futility assessment and GSD trials, as also reflected in recent literature [14, 45].
‘I haven’t got the figures in front of me and I wouldn’t know how to get them but you see a lot more of them at the moment in the scientific advice arena, when people are coming saying ‘this is what we are going to do, what do you think?’ … I get a lot of them starting and not so many of them finished yet’. (QL19 Regulator, regulatory assessment experience)
Regulators advised researchers to engage them through scientific advice meetings and to adhere to their guidance when considering appropriate ADs from trial planning to the end.
Cross-disciplinary interest and positive clinical will
Most participants expressed widespread growing interest towards ADs even through there are existing concerns.
‘I guess there are a lot of concerns about them and so that’s perhaps why they’re not taken up so much. But it is interesting to see that there’s a lot more interest in the past few years and so maybe that is changing’. (QL26 Statistician, design and conduct experience)
‘… influential bodies like the FDA are now embracing ADs and there is probably an increasing number of ADs that are being utilised and will come through and report over the next 2/3/4 years …’ (QL21 Chief Investigator, design experience)
We also inferred a positive will by clinical investigators contradicting previous related findings in early phase trials, suggesting that clinical investigators insist on application of certain methods .
‘We definitely have an interest in advancing new methods in the field of sepsis and in particular there is probably room for improving clinical trial design and that is the focus of our group (ADs methods)’. (QL22 Chief Investigator, design experience)
Most importantly, the positive desire expressed by clinical investigators appears to depend mostly on how trialists market ADs to them and the availability of support during trial planning and conduct.
‘I think generally once you have explained it (AD concept), and you have said that it will be a very big expensive trial if we did it fully powered for as long as it would take, but say that it can be broken down to give different options to the funder for shorter time periods, and less cost, then they can see the advantages to it. … if we are happy to do it and design it and write that section up for them they will take it on’. (QL01 CTU Deputy Director, Proposal Developer, design experience)
‘Sometimes you need to sell it to them to get them to see its positives and advantages and in terms of the extra complication it takes to implement them’. (QL07 Statistician, design and conduct experience)
Cross-disciplinary interest appears to be influenced by the desire to improve design efficiency to answer clinical research questions, address ethical aspects and maximise value for money in research (Table 3).
‘… (ADs) makes a lot of sense from my point of view and in terms of optimising the design and feasibility of the study to address the particular research question. I think it is important that statisticians and clinicians discuss thoroughly the options that are available in clinical trial design to agree the best proposal because each will have particular insights with regard to how to address a research question and so communication is really essential’. (QL24 Chief Investigator, design experience)
‘…there is a lot of interest in them from a funder’s point of view, in that particularly difficulties in recruitment, when it has taken a long time to recruit for trials when recruitment is not up to its expected levels, it is very helpful to be able to have a design that allows you to have multiple looks at the data and to potentially stop early’. (QL04 Chief Investigator, Vice Chair - Public Funder, design experience)
Perceived potential barriers to the use of ADs in confirmatory trials
Cross-disciplinary lack of awareness and understanding
Some participants viewed the widespread lack of awareness and understanding of different types of ADs, circumstances when ADs are appropriate, and implementation resources, as barriers to appropriate use. Consequently, some participants believed that there are missed opportunities and underutilisation of ADs when appropriate in some trials. Some expressed worry that misunderstanding of when ADs are appropriate may lead to misuse, due to their growing prominence - even in certain circumstances when ADs are not superior to traditional fixed sample size designs.
Confusion over what is meant by an AD
One resonant finding we inferred is the potential for misunderstanding of what is meant by the term AD and its acceptable scope in the confirmatory phase. Most participants acknowledged a broadening of the scope of what is considered to be an AD in recent years. Consequently, the term is often loosely defined, but with broad contextual meaning prone to misinterpretation leading to confusion among researchers.
‘I would say, over the last three years, I’ve become aware of (the) detail of ADs. Prior to that, it was a sort of loosely bandied term … I could be in a room and everybody thinks they’re talking about the same thing and they’re talking about very different things’. (QL8 CTU Director, no experience)
‘So I am generally in favour (of ADs); however, convincing the community of that takes some work, so a big threat for ADs is just that it’s a cutesy word that means different things to different people, there’s misinformation about it and there are some existing biases in the community and so there really needs to be a lot of education’. (QL22 Chief Investigator, design experience)
Although this confusion has been partly addressed from a regulatory and industry perspective [8, 42, 43], some participants viewed it as a current problem in the public sector, where most study interventions do not require regulatory approval beyond standard ethics.
Most participants viewed cross-disciplinary conservatism as one of the major barriers to the usage of ADs, particularly in the confirmatory phase. This conservatism depended on: Most importantly, we inferred a complex multifaceted degree of conservatism, which appears to be influenced by many factors. Table 4 summarises subthemes inferred to influence conservatism and negative attitudes towards ADs.
Trial phase and nature of research objective(s),
Health condition or study population and nature of intervention under consideration,
Rationale put forward and completeness in description of the proposed AD(s),
Type and scope of proposed AD, the availability of well-established methods for statistical inference, and perceptions towards that AD by policymakers,
Perceived complexities associated with the AD and impact on implementation, potential introduction of operational bias during conduct, and interpretation of the findings, and
Underlying familiarity and understanding of the proposed AD.
Most participants stated that there is limited scope for ADs in confirmatory trials, due to the definitive nature of research objectives, with direct influence on policymakers’ decisions to approve new interventions into clinical practice. Moreover, some participants strongly advised against conducting too many adaptations in confirmatory trials, citing difficulties in the interpretation of the findings, which may undermine trial credibility.
‘… people should be cautious I guess in trying to do too much and having too many adaptations … We must still make sure we have that body of confirmatory evidence, so I think there might be a place in phase 3 for ADs, but only sort of minimal adaptations. We should sort of keep things under control in that particular setting …’ (QL19 Statistician, regulatory assessment experience)
Insufficient description of proposed ADs, with their statistical and operational characteristics supported with evidence (such as from simulation work or established references) was viewed to influence conservatism towards ADs. This view also reflected the FDA’s regulatory guidance position, which classified well-understood from less-understood ADs [8, 43].
We found that the MAMS AD attracted cross-disciplinary attention, particularly from policymakers, citing efficiency and value for money in testing multiple interventions in a single trial, allowing for dropping futile arms, as opposed to conducting multiple series of independent two arms trials [38, 46].
‘In terms of the multi-arm trials I’m much more comfortable now with the idea of maybe setting out, even on a phase 3 trial, with 4 or 5 potential interventions and dropping the ones that look least promising’. (QL14 Statistician, no experience)
Lack of knowledge and experience
Cross-disciplinary lack of knowledge and experience of ADs was perceived as a major barrier. Most participants viewed this to be intertwined with insufficient access to case studies to facilitate practical training, to raise the awareness of benefits and an understanding of when ADs are appropriate, and learning about barriers and facilitators to successful implementation. Certain participants raised concerns about deficiencies in current training approaches, which they viewed as more oriented towards statistical methodology rather than translational practical training. In addition, weaknesses in some current academic graduate training curricula, which do not tend to incorporate ADs as alternative designs, were articulated.
‘… the main challenge … I think it is a bit broader - is the lack of experience and knowledge within the bio-statistics community. There is a lack of understanding of adaptive methods, a lack of understanding of the opportunities, you know and a lack of familiarity’. (QL12 Clinical Research Leader, Trial Methodologist, design and conduct experience)
A number of participants conveyed a lack of familiarity and knowledge of alternative ethical and efficient designs among ethics and scientific review board members, which may hamper their ability to adequately review grant proposals. The lack of capacity and competency of peer reviewers of AD grant proposals in the public sector was also a perceived barrier. Similar concerns were also reflected in the United States [11, 12].
Degree of statistical and operational complexity
Amount of work and effort required and marketing of ADs to key stakeholders
Most participants stated that ADs, in general, require additional work and effort from a statistical and operational perspective compared to traditional fixed sample size designs during planning and implementation. Consideration of operational feasibility - how implementation of the AD is going to work in practice - was viewed to be vital. Operational feasibility encompasses aspects such as logistics, administration, resources, primary endpoint relative to the expected recruitment rate, implications of trial governance processes and collaborating sites, and intervention delivery . However, the level of statistical and operational challenges depends on the nature of the proposed AD and tends to increase with its complexity. A number of participants believed that more time and effort (depending on the type of AD) is required in marketing the rationale for the proposed AD to key stakeholders (such as funders, regulators and clinical collaborators) and in planning, compared to traditional fixed sample size designs.
Statistical simulation work at the design stage
Some participants mentioned that ADs require more effort and time (depending on the complexity of the AD) to conduct adequate simulation work under various scenarios, and to understand the statistical properties of the design and its implications on decision-making . Some of our interviewees mentioned the concern about inadequate simulation work and its consequences on statistical properties and decision-making. Regulators raised similar concerns about a response adaptive randomisation case study, concerning whether the simulation scenarios covered the entire domain of the desired sample space to guarantee control of the type I error (see Additional file 2, Case Study C). Some participants identified the need for applied training of statisticians on how to undertake adequate simulation work of ADs.
Robust data management infrastructure
Some participants viewed data management and related logistical challenges as potential barriers due to the need to minimise operational bias in the conduct of ADs and to provide clean, robust data to inform the adaptation process. The following considerations were raised:
Compatibility of data management infrastructure with collaborators;
Real-time data capturing, cleaning and processing. An example of a successful multi-centre case study, which used tablet computers for real time electronic data capturing in an African-based trial setting was shared (see Additional file 2, Case Study B);
Turnaround time of data management processes to inform adaptation; and
Systems, processes and procedures supported with audit trails to minimise potential operational bias encompassing the sort of information that should be disclosed and to whom, how the information should be transferred, and firewalls and clarity on who is doing what.
Confidentiality and implications of ADs on IDMC duties and responsibilities
The need to maintain confidentiality by the IDMC during communication and execution of their duties supported with documentation was viewed as paramount. It was advised that the training of and discussions with IDMC members prior to trial commencement regarding the proposed AD; related decision-making criteria; execution of their duties as guided by formalised documents, such as a charter ; communication protocol; and clarification on related issues are essential. Some participants stated that ADs, depending on complexity, may require more effort, time and expertise for the IDMCs in understanding the design, its decision rules and execution.
Additional statistical considerations
The availability of in-house statistical expertise supported with quality control, validated software or user written statistical codes to execute the AD, and delivery time of results to inform interim decision-making were some of the statistical obstacles raised. However, these depend on the type of proposed ADs. An experienced statistician shared a case study where they adapted the methods from another clinical area using a different endpoint to suit their research question but with additional statistical work and time commitment (see Additional file 2, Case Study B).
Concerns around trial credibility and integrity
Most participants expressed strong preference for planned ADs, with decision rules clearly pre-specified at the design stage: this facilitates adequate understanding of the design’s statistical properties through simulation and enhances proper planning. Most importantly, pre-planning of ADs is a regulatory necessity to safeguard trial credibility, integrity and validity, especially in the confirmatory setting [10, 42]. A resonant view was that ADs are not a remedy for poor planning, and most participants were concerned about ad hoc (unplanned) adaptations, which they viewed with great suspicion, regarding such activity as cherry-picking and potentially hiding negative findings to advance the hidden personal agendas of some researchers. Preference for ‘prospectively-planned adaptation’ or the ‘adaptation by design’ concept, particularly in the confirmatory phase, is reinforced in the literature and regulatory guidance [8, 9, 35, 43].
Fear of compromising the trial by potential introduction of bias during its conduct and potential population drift during adaptation were viewed as major concerns, which could be due to dissemination of the interim results [8, 35]. The need for safeguards and firewalls to minimise leaking of interim results, with clear processes, procedures and documentation with audit trails was reinforced [10, 14]. These shared views are in agreement with some of the considerations highlighted in the EMA reflection paper  and the FDA guidance document .
‘I think it (ADs) will always raise an element of suspicion if there have been some decisions made along the way that have been data driven. And the key thing is just to have all the documentation in place; it has to be set out precisely in the protocol how it will be done and you need the right mechanisms in terms of the monitoring committee or steering committee makes the decision and make sure you comply with all the mechanisms. I mean it’s like GCP (Good Clinical Practice); it’s not enough to do the right thing, you’ve actually got to be able to prove you’ve done the right thing… with adaptive trials it’s that much harder to prove that you’ve done it legitimately. So you’ve got to be very careful about the process and got to be able to demonstrate through documentation that you have followed true process’. (QL14 Statistician, no experience)
Although most participants acknowledged routine monitoring as part of every trial, some expressed concern about the lack of understanding of the impact of ad hoc changes on the statistical properties of the design, introduction of bias, interpretation and credibility of the findings . They also highlighted the need for some minimal flexibility as part of routine monitoring in case of unexpected events within a planned AD framework. Anticipation of possible scenarios as much as possible at the planning stage was viewed as imperative.
Concerns around trial validity
A number of statisticians and regulators expressed anxiety about the use of appropriate statistical inference following an AD, arguing that little attention is paid by researchers to the impact on trial results (estimates of treatment effects, confidence intervals (CIs) and P-values). However, they seemed to acknowledge that the awareness regarding control of the type I error rate has improved. More so, some participants highlighted the need for adequate transparency in the conduct and reporting of ADs, and opinion seems divided on whether the current CONSORT guidance is fit for purpose in the case of ADs.
Public sector perspective
Worry about impact of ADs on research staff employment contracts
Some participants stated that the existing public funding models for fixed trial designs create financial uncertainty for research staff employment contracts when trials are stopped early . Consequently, there is nervousness among some UK CTU directors to support certain ADs with options for early stopping. However, some participants stated that this problem is not unique to ADs because some fixed designed trials are stopped early, mainly due to poor recruitment [48, 49]. In contrast, they also stated that design flexibility is somewhat inevitable, due to a paradigm shift by some UK public funders towards risk assessment within an internal pilot framework, with associated staggered research contracts. In addition to the UK CTU reputation and experience, some participants viewed that concerns about the impact of the funding model on staff contracts depends on factors such as the following:
Type of AD proposed - some ADs such as SSR and MAMS are less likely to be affected.
Size of the research group and trial portfolio - large UK CTUs can more easily reassign staff to other trials in the pipeline when an AD is stopped early.
Remit of the public funder - some have more flexible funding models than others.
‘… Because of the size of the trials unit there are many trials that are taking place so we look very closely at people’s contracts and what studies are taking place, it’s not just based on 1 study. We have a lot of different trials at the trials unit so the infrastructure allows for -if the trial stops early then they would be able to work on another trial. So it is not driven by the fact that the contracts or by whether or not it would stop early on this particular trial because of the other trials taking place requiring statistical, trial management, data management support’. (QL27 Statistician, design and conduct)
Many participants acknowledged the need for public funders to draw up standardised, flexible funding agreements compatible with key research partners: UK CTUs, Universities, sites and UK CRN . Some suggested this could be achieved through modification of the current staggered research contracts employed for studies with internal pilots.
Lack of capacity within UK CTUs and time limitation
Most participants stated that there is a lack of expertise and capacity, particularly a dearth of statisticians and proposal developers with knowledge to support complex ADs. However, they acknowledged that capacity and expertise varies across UK CTUs. The majority of participants mentioned that they have limited time to support design work of complex ADs - citing the extra work required against pressure to deliver on competing priorities based on simpler, traditional, fixed sample size designs.
‘One is just the lack of expertise within the unit, so it is easier when you are very busy to put forward a design you know rather than one you don’t. It is also easier because if you put forward a design that does not look the same to clinicians who expect straightforward designs you have to be very confident in that design to be able to convince them to some extent’. (QL9 CTU Director, Statistician, design experience)
Lack of bridge funding for UK CTUs to support planning
Some UK CTU directors voiced concern about the lack of a business case  - citing the amount of time required to support the design of complex ADs, which is unpaid for, betting on uncertain future success of grant applications. Similar concerns have been raised in the US publicly funded setting [11, 12]. UK CTU directors called for funding opportunities in the form of developmental grants to support adequate design work of complex ADs, conditional on meeting research and funding priorities.
‘I think for some of the really complex ADs it would be good if there was availability to go for some small trial development grants so that you could say ‘look this is a convincing clinical question, we think it should be approximately this sort of design but actually we need £20,000 or whatever to properly work it up and design it’ and that type of trial development grant I think would help unlock some of that’. (QL09 CTU Director, Statistician, design experience)
Although this was acknowledged, a funder expressed a contrasting view; citing that in the UK, bridge funding is partly addressed through the NIHR infrastructure support funding accessible to over 25 accredited UK CTUs on a rolling contract basis . However, UK CTU directors’ views appear to suggest that this funding is insufficient, and there is high risk attached to supporting the design work of complex ADs. Funders suggested that researchers may need to consider applying for small grants within the remit of other NIHR funding streams to support developmental work of ADs.
‘Typically for complex ADs then you have to do quite a lot of modelling -that could take 12 or 18 months. Ideally, there should be grants to cover that early development work. Yes, I have sympathy to the idea that there needs to be additional funding but on the other hand I suppose all work that CTUs do prior to a trial application is done at risk. When I was CTU director, typically you are talking about 2 years work before you applied to do a definitive trial. I could say there ought to be more grants to help with all of that and the reality is that we in [organisation] in a sense do pay that upfront because we have a scheme whereby we support CTUs, we give them £250,000 per year if you like, like a front loaded loan, which they use to buy core staff in order to develop new projects. So in a way I think we are doing it already’. (QL25 Chair - Public Funder)
Limitations of the grant application process
In the case of complex ADs, some participants suggested the need to increase the proposal development time prior to submission deadlines to give researchers adequate planning time, particularly for commissioned calls. Some authors suggested a minimum of 3 months for design and planning for ADs . Moreover, a slight modification to the grant application form depending on the funding remit was suggested, to give more space for researchers to describe the rationale, design and its properties, decision scenarios and variable costs adequately.
‘From a practical point of view when you are designing adaptive trials there is more work involved for the application in planning the trial and working out the timelines … you have to do it for a number of different scenarios. So the work involved in that is more from the trialist and statistician’s point of view, the statistician has to do various modelling and look at different scenarios and we have to do all of the different planning and you are usually on a fairly tight deadline for applications because of the way that NIHR funding works so if you only have 6 weeks to work with the team, trying to fit in time to do lots of different scenarios can be quite tricky and can make it more difficult’. (QL01 CTU Deputy Director, Proposal Developer, design experience)
‘There is not an existing section in grant submissions that says ‘if you are doing an adaptive trial design please provide the following information’, so I just don’t know that it’s well organised yet and that could be a good thing or a bad thing …’ (QL22 Chief Investigator, design experience)
Contributions of this study and implications for practice and future research
We found the following cross-sector perceived barriers to the use of ADs in confirmatory trials among the stakeholders we interviewed:
Lack of practical knowledge and applied training coupled with insufficient access to case studies to facilitate practical learning.
Time constraints to support planning relative to other competing priorities based on traditional designs.
Lack of awareness of opportunities about when ADs are appropriate in conjunction with the lack of understanding of their acceptable scope in confirmatory trials.
Statistical and operational complexities during planning and implementation of ADs.
A cross-disciplinary degree of conservatism influenced by various factors.
Specific to the public sector the barriers included lack of bridge funding accessible to UK CTUs to support the design work of complex Ads, difficulties in marketing ADs to collaborators, anxiety about the impact of early trial stopping on full-time researcher employment contracts and lack of capacity to support ADs. Some of these barriers have already been previously reported [2, 5, 11, 13, 14].
Practical education tailored to trialists is paramount to address the lack of practical knowledge. Activities such as educational seminars or webinars and practice-oriented workshops can facilitate translational knowledge sharing. The content of such activities should cover the practical, statistical and logistical issues that need to be addressed in planning and conduct of adaptive trials with the aid of case studies where possible.
We strongly encourage accessible publication of ‘successful’ and ‘unsuccessful’ case studies of ADs previously undertaken, addressing aspects beyond the primary results, such as practical barriers and facilitators, which will complement the educational resources. These should include positive and negative lessons learnt to help the design and conduct of future adaptive trials. Adequate reporting and indexing of these AD-related publications of case studies is important. We are now witnessing such publications in the literature [52–56].
The establishment of small design developmental grants accessible to UK CTUs could encourage trialists wishing to undertake time-consuming and complex ADs, which are efficient to address research questions. Such small grants could be a collaborative initiative among public funders such as the NIHR and MRC. There is an issue that needs to be addressed of funding for statistical AD design work being granted before the merits of the scientific clinical question have been fully addressed. This may entail trialists going through a multi-stage grant application process. An initial ‘Outline’ stage could encompass putting forward the research question, justifying its importance, design rationale, explaining the gap in the design requiring further work, and time and resources required to undertake such work. Trialists could then be sign-posted to apply for small grants to develop the design conditional on the ‘Outline’ proposal meeting research and funding priorities. Further funding of the main trial could then be available conditionally, subject to the outcome of the design work. In addition, there should be mandatory open access publication of the initial stage design-related outputs such as statistical software or implementation codes to help the design of future similar studies.
We found evidence of a complex, multifaceted cross-disciplinary degree of conservatism, which appears to influence perceptions and attitudes towards use and acceptability of ADs among our study participants. We uncovered some of the factors influencing this conservatism, thereby aiding our understanding to address barriers to use and acceptability of ADs. For instance, concerns that the research community, clinical community and policymakers may struggle to accept findings from ADs to influence policy underscores the strong need for methodological assurances and effective communication regarding robustness in decision-making. We believe adequate description of the proposed AD, with clear rationale, scope, and its operational and statistical properties supported with tangible simulation evidence where necessary, may alleviate some cited concerns [2, 7, 11, 36]. This should include the appropriateness of the proposed AD to address the research question(s). Such description must also encompass the use and adequate reporting of appropriate, established statistical methods to control type 1 error and power and to obtain unbiased or bias-adjusted results (estimates of treatment effects, CIs and P values).
The use of retrospectively planned case studies aided with simulation work may help illustrate lost opportunities and provide assurance of the robustness of ADs in decision-making. This may also facilitate practical learning and highlight some pitfalls during the implementation of ADs. In addition, a review of undertaken ADs published in ‘high impact’ peer-reviewed journals and their publicity may help to improve acceptability of ADs in research to change clinical practice. Reassurance of the rigour in the science and conduct of ADs enhanced through transparent, adequate and accessible trial reports is paramount. Such accessible related trial materials include protocols (and amendments), simulation protocols and reports, open and closed IDMC minutes and interim results reports. Consumers of research findings should be able to make informed judgements about the quality of the AD in front of them. This can only be achieved through adequate transparent reporting that has been improved by the advent of CONSORT statements [57, 58]. Recent studies have suggested some modifications to the CONSORT guidance to accommodate ADs [12, 59], without, however, underpinning evidence regarding the state of reporting of ADs. We propose cross-disciplinary and cross-sector discussions to draw recommendations for a modified adaptive CONSORT statement.
We also recommend some form of standardised, consensus guidance toolkit tailored for the public sector (where interventions are so variable) similar to the guidance for the evaluation of complex interventions  that will address appropriate scope, benefits, statistical and practical considerations for successful implementation of ADs in confirmatory trials. This should be carefully crafted so as not to stifle design innovation. In addition, we propose the development of a troubleshooting toolkit tailored for trialists on important general and design-specific questions they should ask themselves when considering ADs at the planning stage. In our opinion, poor communication and dissemination deficiencies of AD methods to researchers are important perceived barriers, which require addressing. A generalised complexity message - ‘ADs are complex to design, implement and analyse’ - could act as a barrier, by deterring researchers from implementing even simple ADs because the level of complexity varies considerably across types of ADs. Hence, we believe there is scope for a toolkit addressing practical and statistical issues specific to particular forms of ADs, rather than the most common generic qualitative statements. We believe too much generalisation of AD-related issues is becoming a syndrome and a communication barrier scaring off some researchers.
We advocate for ADs to be considered alternative trial designs in the ‘toolbox’ of trialists and incorporated as part of MSc academic training curricula in Universities - targeting aspiring trialists and not just statisticians. It is also vital to raise awareness regarding the acceptable scope of Ads, when appropriate, among UK CTU trialists. We encourage collaboration among UK CTUs and industry organisations and the creation of a knowledge-sharing platform on AD-related issues to facilitate problem solving and learning from pacesetters. We support initiatives by UK public funders such as the NIHR and MRC on AD-related capacity building. We hope these efforts will enhance the capacity of experts with practical knowledge to help UK CTUs wishing to undertake efficient, but time consuming and complex ADs, when appropriate. These experts should also be available to undertake scientific peer-review of AD-related grant applications.
We support initiatives by public funders to communicate their receptiveness towards appropriate use of efficient designs such as ADs. We encourage more effort through outreach events to communicate this shift in position to trialists and launching of related funding research opportunities. In addition, we encourage knowledge and experience sharing among public funders to facilitate problem solving on AD-related issues, such as drawing up flexible contractual agreements suited to ADs. Refresher training of public funding panels and board members prior to their grants review meetings may improve their awareness on AD-related issues. More so, it is important for trialists to describe aspects of the proposed AD and related decision-making scenarios adequately in grant applications in order to help public funders to make informed decisions.
Strengths and limitations of the study
The main strength of this study is that we have built upon previous research by incorporating interviews of key stakeholders and maximised sampling variation to capture perceptions and experiences that are more diverse to inform subsequent surveys to be reported elsewhere. This enhanced our robust exploration of barriers and related facilitators to the uptake of ADs. We also used an additional experienced qualitative researcher (JB) to validate consistency in annotation and coding of 7(26 %) of the transcripts, which is recommended as good practice in some qualitative research good practice checklists . In addition, mapping of themes was discussed independently with experienced qualitative researchers (JB and Alicia O’Cathain).
One of the study limitations is the poor participation by health economists, which limited the exploration of AD-related issues among this stakeholder group. Non-participation could be due to a lack of basic understanding of ADs and their implications for health economic evaluation when trials are stopped early, and to some extent, health economists may feel on the periphery of clinical trial design. Thus, there is need for research to understand the implications of ADs on health economic evaluation. Finally, our study participants were predominantly from the UK public-funded setting - thus generalisation to other settings, particularly in relation to organisational barriers may be limited.
There are still considerable, multifaceted individual and organisational barriers perceived to be hampering the appropriate use of ADs in publicly funded confirmatory trials Nevertheless, widespread interest and UK public funders’ perceived positive changes in attitudes and receptiveness towards ADs when appropriate are supportive, and a platform for the future use of ADs in the UK public sector. Our findings have been used to inform the design of follow-up surveys to be reported elsewhere, aiming to generalise these findings, rank barriers with respect to importance for prioritisation and to contrast themes on barriers between stakeholders based in private and public settings.
Clinical Research Collaboration
Clinical Research Networks
Clinical Trials Unit
Doctoral Research Fellowship
European Medicines Agency
Food and Drug Administration
group sequential design
Health Technology Assessment
Independent Data Monitoring Committee
Medical Research Council
National Institute for Health Research
National Health Service
National Institutes of Health
Pharmaceutical Research and Manufacturers of America
Sample size re-estimation
Trial Steering Committee
Bretz F, Koenig F, Brannath W, Glimm E, Posch M. Adaptive designs for confirmatory clinical trials. Stat Med. 2009;28:1181–217.
Kairalla JA, Coffey CS, Thomann MA, Muller KE. Adaptive trial designs: a review of barriers and opportunities. Trials. 2012;13:145.
Morgan CC, Huyck S, Jenkins M, Chen L, Bedding A, Coffey CS, et al. Adaptive design: results of 2012 survey on perception and use. Ther Innov Regul Sci. 2014;48:473–81.
Millard WB. The gold standard’s flexible alloy: adaptive designs on the advance. Ann Emerg Med. 2012;60:22A–7.
Gallo P, Chuang-Stein C, Dragalin V, Gaydos B, Krams M, Pinheiro J. Adaptive designs in clinical drug development--an executive summary of the PhRMA working group. J Biopharm Stat. 2006;16:275–83. discussion 285–91, 293–8, 311–2.
Krams M, Burman C-F, Dragalin V, Gaydos B, Grieve AP, Pinheiro J, et al. Adaptive designs in clinical drug development: opportunities, challenges, and scope reflections following PhRMA’s November 2006 workshop. J Biopharm Stat. 2007;17:957–64.
Chang M, Chow S-C, Pong A. Adaptive design in clinical research: issues, opportunities, and recommendations. J Biopharm Stat. 2006;16:299–309. discussion 311–2.
Chow S-C, Corey R. Benefits, challenges and obstacles of adaptive clinical trial designs. Orphanet J Rare Dis. 2011;6:79.
Coffey CS, Kairalla JA. Adaptive clinical trials: progress and challenges. Drugs R&D. 2008;9:229–42.
Gallo P. Confidentiality and trial integrity issues for adaptive designs. Drug Inf J. 2006;40:445–50.
Coffey CS, Levin B, Clark C, Timmerman C, Wittes J, Gilbert P, et al. Overview, hurdles, and future work in adaptive designs: perspectives from a National Institutes of Health-funded workshop. Clin Trials. 2012;9:671–80.
Connecting Non-Profits to Adaptive Clinical Trial Designs. Themes and Recommendations from the Scientific Advances in Adaptive Clinical Trial Designs Workshop. https://custom.cvent.com/536726184EFD40129EF286585E55929F/files/2627e73646ce4733a2c03692fab26fff.pdf. Date accessed 03/09/2014.
Jaki T. Uptake of novel statistical methods for early-phase clinical studies in the UK public sector. Clin Trials. 2013;10:344–6.
Quinlan J, Gaydos B, Maca J, Krams M. Barriers and opportunities for implementation of adaptive designs in pharmaceutical product development. Clin Trials. 2010;7:167–73.
Englander M. The interview: data collection in descriptive phenomenological human scientific research*. J Phenomenological Psychol. 2012;43:13–35.
Legard R, Keegan J, Ward K. In-depth Interviews. In: Ritchie J, Lewis J, editors. Qualitative research practice: a guide for social science students and researchers. 1st ed. London, UK: Sage; 2003. p. 138–69.
O’Reilly M, Parker N. Unsatisfactory Saturation: a critical exploration of the notion of saturated sample sizes in qualitative research. Qual Res. 2013;13(2):190–197.
Mason M. Sample Size and Saturation in PhD Studies Using Qualitative Interviews. Forum Qualitative Sozialforschung/Forum: Qualitative Social Research, 11(3), Art. 8. 2010.
Creswell JW. Qualitative Inquiry and Research Design: Choosing Among Five Traditions. 2nd Ed. Thousand Oaks, CA, US: SAGE Publications, Inc; 2007.
Morse JM. Determining sample size. Qual Health Res. 2000;10:3–5.
Coyne I. Sampling in qualitative research. Purposeful and theoretical sampling; merging or clear boundaries? J Adv Nurs. 1997;26:623–30.
UKCRC Registered Clinical Trials Units Network. http://www.ukcrc-ctu.org.uk/. Date accessed 03/06/2014.
MRC Network of Hubs for Trials Methodology Research. http://www.methodologyhubs.mrc.ac.uk/. Accessed 03/06/2014.
NVivo: Qualitative Data Analysis Software. http://www.qsrinternational.com/products_nvivo.aspx. Date accessed 03/06/2014.
Gale NK, Heath G, Cameron E, Rashid S, Redwood S. Using the framework method for the analysis of qualitative data in multi-disciplinary health research. BMC Med Res Methodol. 2013;13:117.
Smith J, Firth J. Qualitative data analysis: the framework approach. Nurse Res. 2011;18:52–62.
Braun V, Clarke V. Using thematic analysis in psychology. Qual Res Psychol. 2006;3:77–101.
Pope C, Ziebland S, Mays N. Qualitative research in health care. Analysing qualitative data. BMJ (Clin Res ed). 2000;320:114–6.
Pope C, Ziebland S, Mays N. Analysing Qualitative Data. In: Pope C, Mays N, editors. Qualitative Research in Health Care. 3rd ed. Oxford, UK: Blackwell Publishing Ltd; 2006. p. 63–81.
Ritchie J, Lewis J. Qualitative research practice: a guide for Social Sience Students and Researcher. 1st Ed. London: SAGE Publications, Inc; 2003.
Cabana MD, Rand CS, Powe NR, Wu AW, Wilson MH, Abboud PA, et al. Why don’t physicians follow clinical practice guidelines? A framework for improvement. JAMA. 1999;282:1458–65.
Funk SG, Champagne MT, Wiese RA, Tornquist EM. BARRIERS: the barriers to research utilization scale. Appl Nurs Res. 1991;4:39–45.
Gifford WA, Graham ID, Davies BL. Multi-level barriers analysis to promote guideline based nursing care: a leadership strategy from home health care. J Nurs Manag. 2013;21:762–70.
Tong A, Sainsbury P, Craig J. Consolidated criteria for reporting qualitative research (COREQ): a 32-item checklist for interviews and focus groups. International J Qual Health Care. 2007;19:349–57.
Chow S-C, Chang M. Adaptive design methods in clinical trials - a review. Orphanet J Rare Dis. 2008;3:11.
Quinlan J, Krams M. Implementing adaptive designs: logistical and operational considerations. Drug Inf J. 2006;40:437–44.
Chow S-C. Adaptive clinical trial design. Annu Rev Med. 2014;65:405–15.
Wason JMS, Jaki T. Optimal design of multi-arm multi-stage trials. Stat Med. 2012;31:4269–79.
Encouraging adaptive designs in NIHR funded clinical trials. http://goo.gl/OIpGq7. Date accessed 03/06/2014.
Funding for Primary Research Using Efficient Study Designs to Evaluate Clinical and Public Health Interventions for the NHS. http://www.wales.nhs.uk/sites3/documents/970/NIHR. HTA Researcher Led (efficient study) Specification.pdf. Date accessed 03/09/2014.
MRC funded PhD projects for 2015 entry. http://web.archive.org/web/20141008134746/http://www.methodologyhubs.mrc.ac.uk/about_us/phd2015/phd2015-projects.aspx. Date accessed 03/09/2014.
CHMP. Reflection Paper on Methodological Issues in Confirmatory Clinical Trials Planned with an Adaptive Design. 2007. http://www.ema.europa.eu/docs/en_GB/document_library/Scientific_guideline/2009/09/WC500003616.pdf. Date accessed02/12/2014.
FDA. Guidance for Industry: Adaptive Design Clinical Trials for Drugs and Biologics. 2010. http://www.fda.gov/downloads/Drugs/.../Guidances/ucm201790.pdf. Date accessed 02/12/2014.
Cook T, DeMets DL. Review of draft FDA adaptive design guidance. J Biopharm Stat. 2010;20:1132–42.
Elsäßer A, Regnstrom J, Vetter T, Koenig F, Hemmings RJ, Greco M, et al. Adaptive clinical trial designs for European marketing authorization: a survey of scientific advice letters from the European medicines agency. Trials. 2014;15:383.
Parmar MKB, Carpenter J, Sydes MR. More multiarm randomised trials of superiority are needed. Lancet. 2014;384:283–4.
DAMOCLES Study Group. A proposed charter for clinical trial data monitoring committees: helping them to do their job well. Lancet. 2005;365:711–22.
Sully BGO, Julious SA, Nicholl J. A reinvestigation of recruitment to randomised, controlled, multicenter trials: a review of trials funded by two UK funding agencies. Trials. 2013;14:166.
McDonald AM, Knight RC, Campbell MK, Entwistle VA, Grant AM, Cook JA, et al. What influences recruitment to randomised controlled trials? a review of trials funded by two UK funding agencies. Trials. 2006;7:9.
The Clinical Research Network: delivering research to make patients, and the NHS, better. http://www.crn.nihr.ac.uk/. Date accessed 03/09/2014.
NIHR Clinical Trials Unit (CTU) Support Funding. http://www.nets.nihr.ac.uk/programmes/ctu. Date accessed 03/09/2014.
Chen YHJ, Gesser R, Luxembourg A. A seamless Phase IIB/III adaptive outcome trial: design rationale and implementation challenges. Clinical trials (London, England). 2014.
Baraniuk S, Tilley BC, del Junco DJ, Fox EE, van Belle G, Wade CE, et al. Pragmatic randomized optimal platelet and Plasma ratios (PROPPR) Trial: design, rationale and implementation. Injury. 2014;45:1287–95.
Pritchett Y, Jemiai Y, Chang Y, Bhan I, Agarwal R, Zoccali C, et al. The use of group sequential, information-based sample size re-estimation in the design of the PRIMO study of chronic kidney disease. Clin Trials. 2011;8:165–74.
Sydes MR, Parmar MKB, James ND, Clarke NW, Dearnaley DP, Mason MD, et al. Issues in applying multi-arm multi-stage methodology to a clinical trial in prostate cancer: the MRC STAMPEDE trial. Trials. 2009;10:39.
Bratton DJ, Phillips PPJ, Parmar MKB. A multi-arm multi-stage clinical trial design for binary outcomes with application to tuberculosis. BMC Med Res Methodol. 2013;13:139.
Turner L, Shamseer L, Altman DG, Schulz KF, Moher D. Does use of the CONSORT Statement impact the completeness of reporting of randomised controlled trials published in medical journals? A Cochrane review Syst Rev. 2012;1:60.
Altman DG, Moher D, Schulz KF. Improving the reporting of randomised trials: the CONSORT Statement and beyond. Stat Med. 2012;31:2985–97.
Detry M, Lewis R, Broglio K, Connor J. Standards for the Design, Conduct, and Evaluation of Adaptive Randomized Clinical Trials. 2012. http://www.pcori.org/assets/Standards-for-the-Design-Conduct-and-Evaluation-of-Adaptive-Randomized-Clinical-Trials.pdf. Date Accessed 24/11/2014
Craig P, Dieppe P, Macintyre S, Michie S, Nazareth I, Petticrew M. Developing and evaluating complex interventions: the new Medical Research Council guidance. BMJ (Clin Res ed). 2008;337:a1655.
Stallard N, Hamborg T, Parsons N, Friede T. Adaptive designs for confirmatory clinical trials with subgroup selection. J Biopharm Stat. 2014;24:168–87.
Freidlin B, Korn EL. Biomarker enrichment strategies: matching trial design to biomarker credentials. Nat Rev Clin Oncol. 2014;11:81–90.
MD is fully funded by NIHR as part of a DRF (Grant Number: DRF-2012-05-182). The University of Sheffield funds SAJ and JPN. ST and JB are funded by the University of Reading and University of Hertfordshire respectively. We express our gratitude to Mike Bradburn, Alicia O’Cathain, Cindy Cooper and Daniel Hind for their useful comments and advice during this study as part of the fellowship advisory panel. We also thank Kylie Cross, Helen Wakefield and Lauren O’Hara for interview transcription support. We are grateful for the generous contribution of participants to our understanding of the subject given their busy schedules. Finally, we would like to thank the three reviewers for their constructive comments and suggestions, which improved our manuscript.
This paper presents independent research funded by the NIHR. The views expressed are those of the authors and not necessarily those of the National Health Service, the NIHR, the Department of Health or organisations affiliated to or funding them. Perceptions and attitudes shared by participants are their personal views not necessarily the policy of their organisation.
MD is fully funded by the public sector through NIHR; however, his views do not necessarily represent those of the NIHR, the National Health Service, or the Department of Health. All other authors have no competing interests to declare.
MD led the research, conducted interviews and analysis, and wrote the manuscript. SAJ, JPN and ST supervised the fellowship research project. JB is a qualitative advisory panel member of the fellowship who advised MD during protocol development, conduct and analysis of interviews, and reporting. JB conducted analysis quality control on a proportion of interviews. All authors read and approved the final manuscript.
About this article
Cite this article
Dimairo, M., Boote, J., Julious, S.A. et al. Missing steps in a staircase: a qualitative study of the perspectives of key stakeholders on the use of adaptive designs in confirmatory trials. Trials 16, 430 (2015). https://doi.org/10.1186/s13063-015-0958-9
- Adaptive designs
- Flexible designs
- Qualitative interviews
- Confirmatory trials
- Phase 3
- Clinical trials
- Publicly funded trials
- Early stopping
- Interim analyses