Sunday, July 01, 2018

Geriatric Investigational Plan - Clinical Trials for Elderly

Every clinical trial should have a protocol and the protocol needs to include the inclusion/exclusio criteria to define the study population (i.e., who can participate in the trial). The first inclusion criterion is usually the age limit. In many clinical trials sponsored by AstraZeneca, the 'Ages Eligible for Study' was defined as "18 Years to 130 Years (Adult, Older Adult)". Here are links to entries in clinicaltrials.gov for two of these studies: NEPTUNE study and MYSTIC study The upper limit of 130 years old for the study triggered me to look into the age range for clinical trials in adult population (especiaily the upper limit in elderly patients).

The clinical trials in the pediatric population and the age groups in the pediatric population have been discussed in previous posts.


The reason to have an upper limit of age for study participants is to have a homogeneous population for the study to maximize the chance to have a positive study.  Nowadays, many trials do not set an upper limit for age. An 80-85-year-old individual can be very healthy and an 80-85 years old patient can participate in clinical trials without problems.

For pediatric population, in order to obtain the indication or label, a separate set of clinical trials need to be conducted. There are many requirements and rules for pediatric clinical trials. Regulatory agencies encouraged the sponsors to do clinical trials in pediatric population – pediatric investigational plan (PIP).

Similarly, the safety and efficacy of the specific drug product should also be studied in the geriatric population. The safety and efficacy of a product may be different between the geriatric population and the general adult population. There was an ICH E7 STUDIES IN SUPPORT OFSPECIAL POPULATIONS:GERIATRICS and its Questions and Answers. FDA has a guidance on "Content and Format for Geriatric Labeling". Below are some additional discussions about the clinical trials in the geriatric population.

For a study in adult including elderly patients, the typical criteria for age will be 18 years and older with no upper limit for age. In this way, the geriatric subjects can be enrolled into the study if the subject meets all other inclusion/exclusion criteria and the very older patients can be enrolled into the study if they are healthy enough. However, if we look at the various studies, different sponsors use different criteria for age in their studies.

Using Phase III studies in NSCLC, here are some examples from clinicaltrials.gov. We can see that different clinical trials use different ranges for age. 

Study 
Inclusion Criterion for Age 
18 Years and older (Adult, Older Adult) 
18 Years to 70 Years (Adult, Older Adult) 
18 Years to 75 Years (Adult, Older Adult) 

18 Years to 80 Years (Adult, Older Adult) 
22 Years and older (Adult, Older Adult) 
20 Years and older (Adult, Older Adult) 

Whether or not the upper limit is set to 70, 75, 80 years old or have no upper limit, if the drug is approved, the indication for the drug will be the same and the only difference may be the descriptions in section 8.5 of the drug label per FDA's guidance "Labeling for Human PrescriptionDrug and Biological Products –Implementing the PLR Content and Format Requirements".

Saturday, June 16, 2018

Sample Size Calculation Without Considering Dropout Rate

When planning for clinical trials, sample size calculation is a two-step or three-step process. For clinical trials with continuous variable or categorical variable as primary efficacy endpoint, the sample size calculation is usually a two-step process: the step 1 is to calculate the sample size based on the effect size / standard deviation for continuous variable; difference in rate/proportions and the control group rate/proportion for categorical variable. The step 2 is to apply the dropout rate to calculate the number of subjects needed to be randomized. I had an earlier post “sample size considering the dropout rate” to state that the calculation to account for the dropout rate should be by dividing (1-dropout rate), not by multiplying (1+dropout rate).

For clinical trials with time to event variables, the sample size calculation is usually involved in three steps:
  • Step 1: Estimate the number of events needed based on hazard ratio, median survival time, or event rate at a specified time frame
  • Step 2: Estimate the number of subjects needed to obtain the required number of events based on the accrual time, the follow-up, or the overall study duration
  • Step 3: Estimate the total sample size by considering the dropout rate.
Applying the dropout rate in clinical trials with time to event endpoint is not straightforward. The dropout rate is usually not a constant throughout the study (during the accrual period and during the follow-up period).

To calculate the sample size for studies with time to event variables, I usually use Cytel’s EAST software. To deal with the dropout rate, the EAST software user manual suggests a ‘trial and error’ method. With this method, we will need to provide an initial dropout rate (expressed in ‘probability of dropout’ or ‘hazard for dropout’). We then check the summary of the sample size estimation. The number of dropouts is displayed, and the dropout rate can easily be calculated by dividing the number of dropouts by the number of subjects. If the calculated dropout rate is different from the assumptions, we can then go back to revise the initial input – do this several times until the calculated dropout rate from EAST matches the assumed dropout rate.

I recently run into two scenarios where the sample size calculation does not need to consider the dropout rate.
  • Sample size calculation for an animal study (pigs for example) with a time to event variable. The pigs used in the experiment will be confined in a facility. There is no lost-to-follow-up or anything situation like that. There is no need for us to consider the dropout rate if we try to estimate the number of pigs needed for an experiment. 
  • Sample size calculation for a metastatic non-small cell lung cancer (NSCLC) study with overall survival as the primary efficacy endpoint. The median survival time for metastatic NSCLC patients is usually short. In other words, the mortality rate for NSCLC patients is high. In this situation, the number of subjects who lost to follow-up is generally small (less than 5%) and we may not need to apply the dropout rate in calculating the number of subjects to be randomized. A friend of mine who is an expert in NSCLC clinical trials told me that in their studies with OS as primary efficacy endpoint, they will calculate the number of death events needed, and then simply add 25-30% more patients on top of the number of death events to account for a small proportion of subjects who may live much longer. For example, based on the hazard ratio, accrual time, and follow-up time, if the calculated number of death events is 200 events, we can simply have a sample size of 250-260 subjects randomized (25-30% more than the number of death events). 
Here are some recent trials from New England Journal of Medicine where there was no mention of dropout rate in sample size calculations:



Saturday, June 09, 2018

Expectedness Assessment: Expected / Unexpected Adverse Events and SUSAR

In clinical trials, reporting of adverse events is critical to ensure the safety of participants. When an adverse event is reported, it is also assessed for the severity, seriousness, causality, and outcome. All of these are the standard fields that are supposed to be collected on the case report forms and on serious adverse event (SAE) forms. 

Expectedness of an adverse event is also critical; however, the assessment of expectedness is usually not collected on the case report form or SAE form because the responsibility of the expectedness evaluation is not on investigator’s side, but on the sponsor’s side. SUSAR (suspected unexpected serious adverse reaction) must be reported to regulatory agencies and IRBs in expedited way.

Definition of Expectedness:

According to ICH E2A “CLINICAL SAFETY DATA MANAGEMENT: DEFINITIONS AND STANDARDS FOR EXPEDITED REPORTING”, the unexpected adverse drug reaction is defined as the following:
3. Unexpected Adverse Drug Reaction An adverse reaction, the nature or severity of which is not consistent with the applicable product information (e.g., Investigator's Brochure for an unapproved investigational medicinal product). (See section III.C.)
In each individual study protocol, the definition may be a little bit different, but essentially the same.
Unexpected: – Not listed in Investigator Brochure or is not listed at the specificity or severity that has been observed, or, if an investigator brochure is not required or available, is not consistent with the risk information described in the general investigational plan or elsewhere in the current application.

An unexpected adverse reaction has a nature or severity of which is not consistent with the study intervention description (e.g. Investigator's Brochure for an unapproved investigational product or package insert/summary of product characteristics for an approved product). The unexpected AE must be reported, whether related to the study intervention or not, with as much detail as is available

Expected: - listed in Investigator Brochure.
The Purpose of Expedited Reporting of the Suspected Unexpected Serious Adverse Events (SUSAR)

According to ICH E2A “CLINICAL SAFETY DATA MANAGEMENT: DEFINITIONS AND STANDARDS FOR EXPEDITED REPORTING E2A”, the purpose of the SUSAR reporting is defined as:
C. Expectedness of an Adverse Drug Reaction The purpose of expedited reporting is to make regulators, investigators, and other appropriate people aware of new, important information on serious reactions. Therefore, such reporting will generally involve events previously unobserved or undocumented, and a guideline is needed on how to define an event as "unexpected" or "expected" (expected/unexpected from the perspective of previously observed, not on the basis of what might be anticipated from the pharmacological properties of a medicinal product). As stated in the definition (II.A.3.), an "unexpected" adverse reaction is one, the nature or severity of which is not consistent with information in the relevant source document(s). Until source documents are amended, expedited reporting is required for additional occurrences of the reaction.
The following documents or circumstances will be used to determine whether an adverse event/reaction is expected: 1. For a medicinal product not yet approved for marketing in a country, a company's Investigator's Brochure will serve as the source document in that country. (See section III.F. and ICH Guideline for the Investigator's Brochure.) 2. Reports which add significant information on specificity or severity of a known, already documented serious ADR constitute unexpected events. For example, an event more specific or more severe than described in the Investigator's Brochure would be considered "unexpected". Specific examples would be (a) acute renal failure as a labeled ADR with a subsequent new report of interstitial nephritis and (b) hepatitis with a first report of fulminant hepatitis.
Expectedness/Unexpectedness Not Collected in Case Report Forms or SAE Forms

CDISC/CDASH “Clinical Data Acquisition Standards Harmonization (CDASH) User Guide” excluded the collection of ‘expected criteria’, citing that it is “handled in Clinical Investigative Brochure”



In FDA’s Guidance for Industry and Investigators “Safety Reporting Requirements for INDs and BA/BE Studies”, the reporting responsibility for unexpected adverse events is specified for sponsors (not the investigators). 

Two Types of Unexpectedness and Handling the List of Unexpected AEs

According to a blog post "Seriousness, Expectedness and Investigator Brochures", there are actually two types of unexpectedness 
The first is “regulatory expectedness”. This refers to the SAEs that the company considers likely/possibly or probably related to the study drug. This list is used to determine whether an SAE is a SUSAR (Suspected, unexpected serious adverse reaction) and thus expeditable to FDA, EMA and other health agencies.

The second we can call “clinical expectedness” which is a listing of SAEs that the investigator and patient may encounter during the trial and should be aware of. They may be due to the drug, the disease, comedications, concomitant illnesses (e.g. the flu) or other causes. These may or may not be the same as the “regulatory expectedness” list of SAEs but are important for the treating physician to be aware of and look for. It may not be possible yet to determine whether the particular SAE is due to the drug or the disease or comedications etc. This may become clearer later in the drug’s lifespan as more data becomes available; but sometimes it does not ever become clear.
Due to different understanding of the expectedness, different companies may act differently in handling the expectedness assessment. Some companies (especially the European companies) may want to add as many AEs as possible to the Investigator Brochure so that less AEs would meet the unexpected criteria – specifically the SUSAR criteria for expedite reporting.

Other companies may want to add as few AEs as possible to the Investigator Brochure because two many AEs listed in the Investigator Brochure would give the investigators an impression that the investigational product is not safe.

The right approach should be that the list of AEs/SAEs should be carefully reviewed by one or more medically qualified persons to decide if terms should be included (added) in the Investigator Brochure.

Expectedness assessment is more for fulfilling the sponsor’s reporting responsibility and expected / unexpected AEs are evaluated by the sponsor (not the investigators) through comparing to the Investigator Brochure or product label.

Expected AEs versus AEs of Special Interest (AESI)

Sometimes, the study protocol may include a list of AEs of Special Interest. According to FDA guidance for industry: E2F Development Safety Update Report AEs of Special Interest are defined as following.
“Adverse event of special interest: An adverse event of special interest (serious or non-serious) is one of scientific and medical concern specific to the sponsor’s product or program, for which ongoing monitoring and rapid communication by the investigator to the sponsor can be appropriate. Such an event might warrant further investigation in order to characterize and understand it. Depending on the nature of the event, rapid communication by the trial sponsor to other parties (e.g., regulators) might also be warranted. (Based on CIOMS VI)”
Here are some examples of AESI: distal emboli events in clinical trials using thrombolytic agents, syncope events in pulmonary arterial hypertension studies, diarrhea in Irritable Bowel Syndrome studies) . I had a previous post "Adverse Event of Special Interest (AESI), Standardized MedDRA Query (SMQs), Customer Queries (CQs), and SAS Programming"

AEs of Special Interest are usually the expected AEs.

An event can be an unexpected AE in the early development stage, but become the expected AE and AE of Special Interest in late stage. For example, during the TYSABRI® (natalizumab) (for multiple sclerosis) drug development, a rare brain infection—called progressive multifocal leukoencephalopathy (PML)— was unexpected in early stage, and then become an AE of Special Interest.

Sunday, May 13, 2018

Grading the Severity of AEs and its Impact on AE Reporting


For all adverse events including serious adverse events in clinical trials, severity (or intensity) should be assessed and recorded. AE severity used to be called AE intensity. Nowadays, severity is more commonly used. The assessment of severity is based on the investigator’s clinical judgement, therefore, there are lot of subjective judgement in the AE severity assessment/reporting.

There seems to be three different grading scale in assessing/recording the severity:

Mild, Moderate, and Severe
This is commonly used in non-oncology studies. The definition of the mild, moderate, and severe may be different from one study protocol to another. The severity (intensity) of each AE including SAE recorded in the CRF should be assigned to one of the following categories:
  • Mild: An event that is easily tolerated by the subject, causing minimal discomfort and not interfering with everyday activities.
  • Moderate: An event that is sufficiently discomforting to interfere with normal everyday activities.
  • Severe: An event that prevents normal everyday activities.

 or 
  • Mild: awareness of sign or symptom, but easily tolerated
  • Moderate: discomfort sufficient to cause interference with normal activities
  • Severe: incapacitating, with inability to perform normal activities

CTCAE 
In oncology clinical trials, the AE severity is usually graded according to NCI’s AE Severity Grading Scale -  Common Terminology Criteria for Adverse Events (CTCAE). CTCAE can also be used to grade the AE for non-oncology studies, but generally not appropriate for studies using healthy volunteers.
  • Grade 1 Mild; asymptomatic or mild symptoms; clinical or diagnostic observations only; no intervention indicated
  • Grade 2 Moderate; minimal, local or noninvasive intervention indicated; limiting age-appropriate instrumental ADL
  • Grade 3 Severe or medically significant but not immediately lifethreatening; hospitalization or prolongation of hospitalization indicated; disabling; limiting self care ADL
  • Grade 4 Life-threatening consequences; urgent intervention indicated.
  • Grade 5 Death related to AE.

Vaccine's Trials
In FDA’s guidance on vaccine trials “Toxicity GradingScale for Healthy Adult and Adolescent Volunteers Enrolled in PreventiveVaccine Clinical Trials”, the AE severity based on clinical abnormalities and laboratory abnormalities was graded as
  • Mild (grade 1)
  • Moderate (Grade 2)
  • Severe (Grade 3)
  • Potentially Life Threatening (Grade 4)

In statistical summaries, the grade 1 is counted as ‘mild’, the grade 2 as ‘moderate’, >= grade 3 will be counted as ‘severe’.  

During the course of an adverse event, the severity may change – which may have impact on how we report the adverse event.

In one of the previous posts ‘SAE Reconciliation and  Determining / recording the SAE Onset Date’, we discussed that an AE with the change in seriousness might need to be split into two events for recording: one non-serious AE with onset date of the first sign/symptom and one serious AE with onset date of the event meeting one of the SAE criteria. The similar issue arises when we try to record the AE with severity change.

The most common instruction for AE recording is that when there is severity change, a new AE should be recorded. Here are some example instructions:
Start Date
Record the date the adverse event started. The date should be recorded to the level of granularity known (e.g., year, year and month, complete date) and in the specified format. If a previously recorded AE worsens, a new record should be created with a new start date. There should be no AE start date prior to the date of the informed consent. Any AE that started prior to the informed consent date belongs instead in the medical history. If an item recorded on the medical history worsens during the study, the date of the worsening is entered as an AE with the start date as the date the condition worsened.
End Date
Record the date the adverse event stopped or worsened.  The date should be recorded to the level of granularity known (e.g., year, year and month, complete date) and in the specified format.  If an AE worsens, record an end date and create a new AE record with a new start date and severity. 
If the AE increases in severity per the DAIDS Grading Table, a new AE Log CRF should be completed to document this change in severity.
the eCRF Completion Guidelines for adverse events:  Enter a new event if action taken, seriousness, causality, severity (intensity), etc. changes over the course of an adverse event.  A timestamp for any changes in events can be seen in the data via event start/stop dates.
However, this way of recording the adverse events may result in splitting the single event into multiple adverse events and may result in over reporting in the number of adverse events.

Suppose the subject experienced a headache adverse event, the event started with mild intensity, then progressed to moderate, and then went back to the mild intensity. Should this headache be reported as three separate adverse events (two with mild severity and one with moderate severity)? or Should it be reported as single event with moderate severity?

This question was submitted to FDA and the FDA response (see the link below) suggested that this should be reported as one event (with the maximum severity)


The second question and answer explicitly stated:

Question 2:


[Redacted] is the sponsor of the study. We have been advised by our data coordinating center to record an AE that changes in severity as two AEs instead of 1 AE - starting a new AE each time the severity changes. This convention is different than that of our previous coordinating center and has caused us great concern.

Answer 2:

We have concerns that an approach to adverse event reporting as you described below (i.e., a change in severity of an adverse event necessitates a new adverse event report) may inaccurately reflect the adverse event profile for the product. Therefore, we strongly recommend that you contact the FDA review division regulating this clinical investigation for additional input on the most scientifically and medically sound approach to the adverse event reporting specifically for this trial.


I recently submitted this same question to FDA’s OC GCPQuestions and Answers and got the following response:

Question: 
We constantly run into the issue how to record the adverse event in the database in the situation there is a severity change or seriousness change during the course of the adverse event.
 A subject in clinical trial reported a mild headache. Two days later, the headache became moderate in severity. Then headache became mild in severity again.
 In this case, shall we record this as one headache event with moderate severity or record as three headache events (a new event is record whenever there is a severity change)?
 Similarly, a subject in clinical trial reported a non-serious adverse event. Several days later, subject needs to be hospitalized for this adverse event – now the event meets the seriousness criteria.
 In a situation of a non-serious adverse event becoming serious, shall we record it as a single AE with seriousness or shall we record as two separate AEs (one non-serious AE and one serious AE)?

OC-GCP Response:
Given your brief description that the subject's headache is ongoing, it would seem that this adverse event would best be reported as a single event with variable severity. However, the clinical judgment of the principal investigator (or, if the principal investigator is not a clinician, then a physician consultant to the research) would be helpful in clarifying the symptoms and hence the reporting of the adverse event(s). There are several cogent clinical scenarios the understanding of which would require more information than you have supplied. For example, the subject's symptomatology could represent an unremitting headache of several days duration or episodic headaches of finite duration with varying intensities or a symptom of another event altogether such as a change in blood pressure, etc. The same would apply for the hospitalization event.
 To best sort out the adverse event(s) itself and therefore the appropriate reporting, I would recommend a clinical assessment of the headache. In addition, the protocol may have detailed how adverse events should be reported. As well, the sponsor (I'm not sure of [Redacted] status in this trial, i.e., is/is not the sponsor) may have specifications for adverse event reporting that could guide you. If you still feel uncertain, I would strongly recommend contacting the FDA review division regulating this trial.
 Lastly, if it becomes apparent that this same "fact pattern" recurs, it may be advisable for the sponsor to clearly articulate standards for adverse event reporting such that there can be consistency in reporting of headaches.
From the statistical analysis standpoint, whether or not it is recorded as one event with maximum severity or multiple events with various seventies do not have impact on our calculation of the incidence of AEs. However, it will have great impact on the calculation of the number of AEs.

It is the common understanding that if an event recorded on the medical history worsens during the study or after the initiation of the study drug, a new AE should be recorded and the date of the worsening is entered as the new AE onset date

Sunday, April 08, 2018

Clinical Trials Using Historical Control in Rare Disease Drug Development


While the randomized, control clinical trial has become and remains to be the golden standard in drug development, we also see the increased use of non-randomized, single-arm study where the effectiveness of the testing drug is compared to the historical control.

When a pivotal study is a single arm and has no concurrent control, the results from the study will need to be compared to the historical control or a common standard that has been accepted by the medical community or regulatory agencies. This seems to be more common in the oncology area and in rare disease areas.

Without any specific statistics, I can only say this is my impression that the historical control seems to be more accepted by the US FDA in its approvals in oncology and rare disease areas. Here are three examples of recent drug approvals based on historical control:

Venetoclax in Relapsed / Refractory Chronic Lymphocytic Leukemia (CLL)

The pivotal study is a single-arm study without concurrent control and the primary efficacy endpoint is objective response rate (ORR). The result is compared with the historical / standard rate of 40%.

eteplirsen in Duchenne Muscular Dystrophy (DMD)

The original study was a randomized, double-blind, placebo control study with three arms (eteplirsen 30 mg/kg weekly; eteplirsen 50 mg/kg weekly, and placebo) – 4 subjects in each arm for a total 12 subjects. All subjects including placebo subjects were rolled over to an open-label extension study for long-term assessment.

The results from double-blind portion of the study did not provide the strong evidence for efficacy. The sponsor conducted a post-hoc comparison with a historical control.

FDA was not convinced with eteplirsen’s efficacy and conducted the advisory committee. In the end, the eteplirsen was approved as the first treatment in Duchenne Muscular Dystrophy with a lot of controversies. The comparison with historical control (while hotly debated) was a big part of the evidence contributing to the approval.

Brineura for Batten Disease

The entire clinical program included: 
  • A natural history study with 69 subjects (42 evaluable)
  • A Phase 1/2 FIM single-arm study with 24 subjects (23 completed)
  • A long-term follow-up study with 23 subjects

Natural history study was based on registry data; Provided the basis as the historical control group
Comparability between natural history study and phase 1/2 study were extensively debated during the review process.

FDA finally approved Brineura for Treating Batten disease in 2017.

The use of historical control is not new. It was stated in the ICH guideline E10 “CHOICE OF CONTROL GROUP AND RELATED ISSUES IN CLINICAL TRIALS”
Historical control was again mentioned in FDA’s guidance for industry “Rare Diseases: Common Issues in Drug Development”. FDA encourages the natural history study to establish the historical control.

During the FDA advisory committee meeting, Dr Temple gave a presentation about "Historically Controlled Trials": see FDA's presentation slides (Dr Temple's presentation started from page 20).

In FDA's statistical review of eteplirsen in DMD, the following comments were made on the use of historical control: 
Historical Control Cohort
The comparison of eteplirsen with historical controls was not part of an adequate and wellcontrolled study. The applicant obtained historical data after observations were made for the eteplirsen patients. Historical data were obtained from 2 DMD patient registries (Italian DMD Registry and the Leuven Neuromuscular Reference Center – NMRC) for comparison to eteplirsen-treated patients
……
According to the ICH E10 guidance on Control Group and Related Issues in Clinical Trials, the major and well-recognized limitation of externally controlled (including historical control) trials is inability to control bias. The best group and control group can be dissimilar with respect to a wide range of observable and unobservable factors that could affect outcome. It may be possible to match the historical control group to the test group in observed factors but there is no assurance for any unobserved factors. “The lack of randomization and blinding, and the resultant problems with lack of assurance of comparability of test group and control group, make the possibility of substantial bias inherent in this design and impossible to quantitate.”
 Because of the serious concern about the inability to control bias, the use of the external control design is restricted only to unusual circumstances.
  • ICH E10 states that “an externally controlled trial should generally be considered only when prior belief in the superiority of the test therapy to all available alternatives is so strong that alternative designs appear unacceptable…” However, such prior belief does not exist for eteplirsen.
  • ICH E10 states that “use of external controls should be limited to cases in which the endpoints are objective…” however, performance on the 5-minute walk test can be influenced by motivation. Patients may not achieve maximal 6MWT due to concerns of falling or injury, or patients could try harder with encouragement and with the expectation that the drug might be effective.
  • Pocock’s criteria for acceptability of a historical control group require that “the methods of treatment evaluation must be the same,” and “the previous study must have been performed in the same organization with largely the same clinical investigators.” This is especially important when assessing endpoints such as 6MWT, in contrast to hard endpoints such as mortality. For this NDA, these requirements are not met.
Moreover, the historical control group was identified post-hoc in this NDA, leading to potential selection bias that cannot be quantified. If a historical control is to be utilized, selection of the control group and matching on selection criteria should be prospectively planned without knowing the outcome of the drug group and control group.
 Based on ICH E10, “a consequence of the recognized inability to control bias is that the potential persuasiveness of findings from externally controlled trials depends on obtaining much more extreme levels of statistical significance and much larger estimated differences between treatments than would be considered necessary in concurrently controlled trials.” The success criteria for this historical control study were not discussed or pre-specified in the protocol.
 Given all these concerns, including issues of comparability of eteplirsen-treated patients and historical control cohort patients, the fact that 6MWT is not a “hard” efficacy endpoint, the potential of selection bias due to the post-hoc identification of the control cohort by the applicant, and all the known pitfalls with the use of historical controls, the comparison of the eteplirsen with historical control is not statistically interpretable. 


However, even though the statistical review on the use of historical control was very negative, the eteplirsen was still approved as the first drug treating the DMD and presumably the results from the comparison to historical control played the pivotal role in decision. 

In general, the use of historical control can be accepted in some situations especially in rare disease area where there is no approved drug available. Whenever the historical control is used, the following factors need to be considered:
  • If the proposed historical control cohort is a priori or post hoc. It is encouraged to collect historical control information from natural history studies. 
  • if the patient population from historical control is comparable
  • If the outcome measurement is comparable
  • if the outcome measurement is a hard endpoint (such as death) or soft endpoint (such as 6MWD)
  • If the endpoint measure is easily affected by other factors
  • if the endpoint is a soft endpoint (such as objective response rate ORR), whether or not any approach is implemented to avoid the bias (such as using central reader)

Saturday, April 07, 2018

Generating graph / figure in publication quality

As I was recently preparing a poster for ATS Internal Conference, I was told that the plots I provided were not in high quality. When placing the plots on the poster, they became blurry. I realized that the issue was with the DPI. 

DPI is used to describe the resolution number of dots per inch in a digital print and the printing resolution of a hard copy print dot gain. High DPI = High Resolution.

The journal may have a requirement for the minimum resolution for the graphs and figures, for example, the PLOT One requires the resolution in the range of DPI 300-600. The Science magazine has the following requirement: 
Resolution. For manuscripts in the revision stage, adequate figure resolution is essential to a high-quality print and online rendering of your paper. Raster line art should have a minimum resolution of 600 dots per inch (dpi) and, preferably, should have a resolution of 1200 dpi. Grayscale and color artwork should have a minimum resolution of 400 dpi, and a higher resolution if possible.
Wiley had a paper discussing the challenges the authors might face for providing the high resolution figures. See the editorial: How to meet dots per inch requirements for images

I used SAS procedure sgplot to create the plots. The default DPI is 100, which is too low for publication or poster. Fortunately, there are easy ways to change the DPI for the output plots. Below are some programs for doing so: 

*default DPI=600; low resolution;
ods listing gpath='c:Temp\';
ods graphics on;
proc sgplot data=sashelp.stocks (where=(date >= "01jan2000"d
                                 and date <= "01jan2001"d
                                 and stock = "IBM"));
   title "Stock Volume vs. Close";
   vbar date / response=volume;
   vline date / response=close y2axis;
run;
ods graphics off;
ods listing close;


*set DPI=400;
ods listing gpath="c:\Temp\" dpi=400;
ods graphics on;
proc sgplot data=sashelp.stocks (where=(date >= "01jan2000"d
                                 and date <= "01jan2001"d
                                 and stock = "IBM"));
   title "Stock Volume vs. Close";
   vbar date / response=volume;
   vline date / response=close y2axis;
run;
ods graphics off;
ods listing close; 

*set DPI =400 and also use style=journal;
ods listing gpath="c:\Temp\" style=journal dpi=400;
ods graphics on;
proc sgplot data=sashelp.stocks (where=(date >= "01jan2000"d
                                 and date <= "01jan2001"d
                                 and stock = "IBM"));
   title "Stock Volume vs. Close";
   vbar date / response=volume;
   vline date / response=close y2axis;
run;
ods graphics off;

ods listing close;

When high DPI is chosen, the size of the file will increase. For the same plot, the file sizes for three plots (in png format) above are 26, 160, and 158 kb. 

The following program will service the same purpose, but use ods pdf commend. 

ods pdf file="c:\Temp\test.pdf";
proc sgplot data=sashelp.stocks (where=(date >= "01jan2000"d
                                 and date <= "01jan2001"d
                                 and stock = "IBM"));
   title "Stock Volume vs. Close";
   vbar date / response=volume;
   vline date / response=close y2axis;
run;
ods pdf close;


ods pdf file="c:\temp\test.pdf" dpi=600;
proc sgplot data=sashelp.stocks (where=(date >= "01jan2000"d
                                 and date <= "01jan2001"d
                                 and stock = "IBM"));
   title "Stock Volume vs. Close";
   vbar date / response=volume;
   vline date / response=close y2axis;
run;
ods pdf close;


ods pdf file="c:\Temp\test.pdf" style=journal dpi=600;
proc sgplot data=sashelp.stocks (where=(date >= "01jan2000"d
                                 and date <= "01jan2001"d
                                 and stock = "IBM"));
   title "Stock Volume vs. Close";
   vbar date / response=volume;
   vline date / response=close y2axis;
run;
ods pdf close;

The same issue with graph quality is also true when we use R. There is a RES = option to select the desired DPI. Please see the blog post by Daniel Hocking "High Resolution Figures in R".

Monday, March 05, 2018

Handling Randomization Errors in Clinical Trials with Stratified Randomization

The stratified randomization is very common in randomized, controlled clinical trials. The usage of the stratified randomization has been discussed in previous posts. 
While the stratified randomization has its benefits, it does not mean the more stratification factors are  better. The more stratification factors we have, the more easily the randomization error of using a wrong stratum can occur. 

It becomes common to utilize the interactive response technology (IRT) system such as interactive response system (IVR) or interactive web response (IWR) systems for implementing the randomization and treatment assignments. The IRT system usually has to go through extensive quality control (QC) and user acceptance test (UAT) before the implementation, therefore the randomization errors can be minimized. Comparing to the manual randomization process, the randomization error rate is lower in studies with IRT system for implementing the randomization. 

However, the use of IRT system requires the investigation site staff (pharmacist, investigator, or study coordinator) to enter the stratification information at the time of randomization. The site staff can enter the incorrect stratification information into the IRT system, the treatment assignment will then be pulled from the wrong stratum. The randomization error due to choosing a wrong stratum is probably the most common randomization error we see in clinical trials with stratified randomization. The more stratification factors we have, the more likely incorrect stratum can be chosen. 

In addition to the number of stratification factors, ambiguous description / definition of the randomization stratum and lack of clarity about source of measurement (for example, the local lab or central lab results for a lab related stratification factor) can all contribute to choosing an incorrect stratum for randomization. 

For example, in a clinical trial in neurology area, the sponsor plan to have patients stratified by their use of cholinesterase inhibitors, corticosteroid, immunosuppressant/immunomodulator. The following stratification factor is constructed.
  • Regimen includes only cholinesterase inhibitors
  • Regimen includes corticosteroid (CS) as the only
  • immunosuppressant/immunomodulator, alone or in combination with other MG medications (e.g., a subject on prednisone plus a cholinesterase inhibitor would be in this stratum)
Without appropriate training, it is likely that the site staff will choose a wrong category for the randomization.

It is also common that the stratification factor is based on one of the laboratory measures. The original laboratory measure is a continuous result and it is then categorized for the stratification purpose. In this case, the protocol must be clear whether or not the stratification will be based on the lab results from the local lab or central lab because the results from local versus central labs can be different. 

When a wrong stratification stratum is chosen for the randomization (the randomization error occurs), the natural reaction is trying to fix it. However, with the IRT system, it is not easy to go back to the system to fix the randomization error. Actually it is strongly encouraged not to try to fix the issue. 

"...the safest option is to accept the randomisation errors that do occur and leave the initial randomisation records unchanged. This approach is consistent with the ITT principle, since it enables participants to be analysed as randomised, and avoids further problems that can arise when attempts are made to correct randomisation errors. A potential disadvantage of accepting randomisation errors is that imbalance could be introduced between the randomised groups in the number of participants or their baseline characteristics. However, any imbalance due to randomisation errors is expected to be minimal unless errors are common. Imbalance can be monitored by an independent data monitoring committee during the trial and investigated by the trial statistician at the analysis stage."
It is true that if randomization errors can skew the analyses especially when the occurrence of the randomization errors is not infrequent. In a paper by Ke et al "On Errors in Stratified Randomization", the impact of the randomization errors on treatment balance and properties of analysis approaches was evaluated. 

If there are a lot of randomization errors, the study quality and integrity will be questioned. From the statistical analysis standpoint, the strict intention-to-treat analysis may not be appropriate. With significant number of randomization errors with incorrect treatment assignment, we may need to analyze the data using 'as treated' instead of 'asrandomized'. With significant number of randomization errors due to incorrect selection of the randomization stratum, we may need to base the stratum information from the clinical database (assuming it is correctly recorded) instead of from the information used in IRT system. 

When randomization errors are identified during a study, the root cause of the error should be investigated. Additional training may be needed to prevent the further occurrence of the randomization error.