Rational Selection, Criticality Assessment, and Tiering of Quality Attributes and Test Methods for Analytical Similarity Evaluation of Biosimilars
- 3.3k Downloads
Leading regulatory agencies recommend biosimilar assessment to proceed in a stepwise fashion, starting with a detailed analytical comparison of the structural and functional properties of the proposed biosimilar and reference product. The degree of analytical similarity determines the degree of residual uncertainty that must be addressed through downstream in vivo studies. Substantive evidence of similarity from comprehensive analytical testing may justify a targeted clinical development plan, and thus enable a shorter path to licensing. The importance of a careful design of the analytical similarity study program therefore should not be underestimated. Designing a state-of-the-art analytical similarity study meeting current regulatory requirements in regions such as the USA and EU requires a methodical approach, consisting of specific steps that far precede the work on the actual analytical study protocol. This white paper discusses scientific and methodological considerations on the process of attribute and test method selection, criticality assessment, and subsequent assignment of analytical measures to US FDA’s three tiers of analytical similarity assessment. Case examples of selection of critical quality attributes and analytical methods for similarity exercises are provided to illustrate the practical implementation of the principles discussed.
KEY WORDSbiosimilars comparability CQA QTPP tiering
Medicines of biological origin are complex mixtures containing a diversity of chemical variations of a therapeutic protein and other substances as a result of their production in living systems. The specific composition of such a product is controlled within carefully defined limits of variation by the design and control of its manufacturing process. Because the details of product manufacture are proprietary knowledge, biosimilar product developers cannot precisely replicate the manufacturing process of a reference product. Due to the inherent manufacturing variability associated with cell culture and the inability to precisely replicate the originator’s manufacturing process, even the best reproductions of protein therapeutics can at best be highly similar to a reference product based on current technologies.
Once satisfactory biosimilar material has been produced, analytical assessment of biosimilarity can proceed. The analytical assessment (and corresponding analytical study design) is viewed as a critical first step in biosimilar development by both US FDA and EMA, since the scope and range of the requisite clinical studies needed for approval of the biosimilar will be, in part, determined by the degree of analytical similarity. The assessment can be broken down into several distinct steps. First, molecular attributes to be considered are defined based on the drug’s structure, mechanism of action, safety, and efficacy, and appropriate analytical methods to assess those attributes are selected. Second, the criticality of each attribute is determined based on its effect on clinical outcome, as assessed from a number of angles. Both of these steps build on the work that has ideally been initiated at, or prior to, the start of product development. US FDA specifically expects a third step, in which the attributes are assigned to different tiers of statistical evaluation, each of which is associated with a specific methodology of data analysis and interpretation. The statistical methodology for evaluation of analytical similarity data was discussed in a previous publication (1) and is the subject of recent draft regulatory guidance documents (2,3).
This white paper provides possible methodologies distilled from the combined experience of a group of subject matter experts from different organizations regarding the selection of quality attributes and their criticality assessment, as a prerequisite to the design of an analytical similarity study program meeting the current expectations of leading regulatory agencies. It also discusses the assignment of proposed analytical similarity measures to the different tiers of assessment recommended by US FDA (2,4).This document was prepared under the auspices of the AAPS Biosimilars Focus Group’s Chemistry Manufacturing and Controls Analytical (CMC-A) subcommittee and is intended to share current practices within the CMC-A community to aid the development of biosimilar products.
SELECTION OF ATTRIBUTES AND ANALYTICAL PROCEDURES
Any publicly available information from the reference product manufacturer that provides information on QAs that are tested and tightly controlled within a predefined range for the reference product. Notably, Certificates of Analysis (COAs), when available, may only be accessible for the reference drug product (DP) not for the originator drug substance (DS). Release tests that are only performed on DS may not be listed on the DP COA.
Reported information on the reference product and similar products from other sources, e.g., information from standard setting organizations (product-specific or molecular type, e.g., monoclonal antibodies), regulatory guidelines (e.g., ICH Q6B), pharmacovigilance reports, European public assessment reports (EPARs) on the reference product, and health agency communications to the reference product sponsor (post marketing commitments or findings).
Literature reports on the reference product and related products (research papers, mechanism of action, and clinical findings).
Results from detailed analytical characterization of multiple lots of the reference product.
Case study example: attribute selection
Avastin (bevacizumab) is known to exert its clinical effect by binding and neutralization of vascular endothelial growth factor (VEGF), preventing association of VEGF with its receptor by steric hindrance. Original research did not reveal a contribution of Fc-related bioactivity (antibody-dependent cell-mediated cytotoxicity (ADCC), complement-dependent cytotoxicity (CDC)) to the observed clinical effect of bevacizumab (5). However, bevacizumab may also bind with membrane-bound VEGF (expressed by certain carcinomas, or in VEGF ligand-receptor complexes) and can hypothetically mediate Fc-related effector functions. Salvador et al. (6) report in vitro CDC activity for bevacizumab in human ovarian carcinoma SKOV-3 cells. This example illustrates how all plausible biological functions of the molecule of interest, even those reported to be clinically irrelevant for the reference product, must be included in the list of attributes for similarity assessment. Publically available information on the development of ABP215, Amgen’s biosimilar version of bevacizumab, confirms this expectation (7).
Suitability of the analytical method to measure existing differences between reference product and biosimilar product.
Ability of the method to yield quantitative data (e.g. qualitative vs. quantitative methods such as SDS-PAGE vs. capillary gel electrophoresis with sodium dodecyl sulfate (CE-SDS)).
Orthogonality. Each attribute may be interrogated by one or more analytical methods. Conversely, the same method may inform on several attributes. For example, aggregates are observed, e.g., by SEC, AUC (analytical ultracentrifugation), or AF4 (asymmetrical flow field flow fractionation), to name a few. At the same time, SEC can also be used to differentiate other size variants, such as dimers or clips.
As the biosimilar product manufacturing process and the overall biosimilar product understanding matures, test methods dedicated to testing particular QAs are developed, qualified, validated, and implemented. In addition, as the analytical tools are developed and optimized, the improved assays can be used to provide feedback for the manufacturing process to achieve a high degree of similarity. While a formal validation study is not strictly required for methods used in similarity assessments, their suitability for the intended purpose should be demonstrated. Sponsors also need to consider the composition of the test samples for analytical similarity studies: regulatory agencies usually expect analysis of the drug product, except for quality attributes that cannot be adequately measured in the drug product, e.g., due to low protein concentration or the presence of interfering excipients. In such cases, appropriate deformulation or other sample processing strategies have to be developed.
Proposed Quality Attributes and Possible Techniques for Analytical Similarity Assessment of a Protein Therapeutic
Attribute (method or methods to query that attribute)
• Primary sequence (e.g., UPLC peptide map, LC-MS/MS, amino acid analysis, Edman degradation, carboxypeptidase sequencing)
• Disulfide structure (e.g., LC-MS of non-reduced protein digest)
• Intact mass (e.g., LC-MS)
• Isoelectric point (e.g., IEF gels, cIEF, iCE)
• Extinction coefficient (e.g., UV/AAA or UV/RI)
Secondary and tertiary structure
• Low-resolution secondary structure or indirect tertiary structure measurements (e.g., CD, DSC, FTIR, fluorescence)
• High-resolution measurements of higher order structure (e.g., 2D-NMR, HDX-MS, X-ray crystallography)
• Glycosylation (e.g., HILIC, MS (MALDI, ESI), exoglycosidase sequencing, HPLC-FLD, HPAEC-PAD, CE-LIF)
• Glycosylation site mapping/site occupancy (e.g., peptide mapping by LC-MS)
• Protein content (e.g., UV A280, RP-HPLC)
• Deliverable volume (extractable volume)
• Sub-visible particles (e.g., light obscuration, MFI, NTA)
• Biological activity (e.g., for mAb: proliferative bioassay, cytotoxicity assay, ADCC, and CDC; other assays may be appropriate for other proteins, e.g., enzyme kinetics for biosimilar enzyme)
• Receptor and/or ligand binding (e.g., SPR, ELISA)
Product variants (product-related substances and impurities)
• High molecular weight species (e.g., SEC-MALS, AF4/HF5, AUC, DLS)
• Covalent dimers (e.g., SDS PAGE, CE-SDS)
• Purity and impurities (oxidation, deamidation, glycation, isomerization, fragmentation, disulfide reduction, e.g., RP-HPLC, CEX, SEC, IEX, IEF, cIEF, LC-MS)
• Amino acid misincorporations (e.g., LC-MS/MS)
• Microsequence heterogeneity (e.g., LC-MS)
• C- and N-terminal modifications (e.g., LC-MS, Edman degradation)
Expectations for comparison of levels of process-related impurities (e.g., host-cell protein (HCP), host-cell deoxyribonucleic acid (DNA), Protein A) differ between jurisdictions. Some agencies, such as EMA, do not expect comparative analysis, arguing that these impurities are process-dependent and are therefore not expected to be similar, although they must be controlled according to acceptable standards (8). US FDA, on the other hand, does instruct sponsors to compare levels of process-related impurities, and to perform a risk-based assessment in case of differences in-process-related impurities (9).
Knowledge of the product’s specific mode of action;
Publicly available information on the reference product, including scientific literature and regulatory approval history;
Knowledge of similar products, in particular information on effector functions and interactions with biological targets that are relevant for the proposed biosimilar, and information on the relationship between structural features and clinical outcomes; and
The sponsor’s own experimental data, e.g., results of biological characterization (e.g., bioassays, in vivo models or clinical studies) of isolated product variants (or variants obtained by enzymatic processing, forced degradation, etc.).
Different models have been used successfully in the criticality assessment of product quality attributes. Examples of such models have been presented at conferences, by representatives of companies who have received approval for their biosimilar products in the USA (10, 11, 12). Based on these examples, and the collective experience of the authors, we conclude that most models appraise at least two out of three dimensions of criticality. First and foremost, the relationship between the attribute of interest and the clinical performance of the product is evaluated, to determine how the attribute may affect clinical outcome. The second dimension commonly employed assesses the evidence supporting the attribute’s clinical impact, which can range from purely academic (no data available) to unequivocal proof derived from (multiple) clinical studies or other clinical experience. Finally, a risk score may be factored into the assessment of attribute criticality, such as the risk of the attribute truly affecting clinical performance. One example is when a product variant is found to be biologically inactive (or less active), and therefore viewed as critical (risk of reduced efficacy). If the product variant in question is only formed in minute quantities (independently from any clearance during downstream processing), it may be judged to be clinically irrelevant based on its low abundance. An additional consideration for industry may be the likelihood of the attribute being dissimilar based on knowledge of process performance and level of process control. Notably, certain quality attributes may have a critical impact on product efficacy or safety if not controlled within a predefined range, but are so tightly controlled by process design, in-process, and quality control testing that the risk of the attribute being out of range is negligible, reducing the overall criticality ranking for this attribute. Thus, a tightly controlled process can mitigate product risk, may impact the need for release testing as part of the control strategy, and may also impact the tier of an attribute in a similarity assessment. One such example is protein dose (i.e., the absolute amount of therapeutic protein administered to the patient) which can be tightly controlled by manufacturing processes. However, protein dose remains a critical attribute for similarity assessment because of potential impact on the patient who may be receiving the reference product and/or the biosimilar product as part of their treatment. Importantly, consideration of the product control strategy for determination of attribute criticality is a practice that is not endorsed by all companies and is not accepted by regulatory agencies.
Quantitative or qualitative assessment techniques can be used to determine attribute criticality. Quantitative models use predefined scoring roles, whereby each assessment dimension is broken down into several numerical scores with an associated definition, ideally in sufficient detail to avoid ambiguity and subjectivity. An aggregate score for the attribute is calculated by simple arithmetic, usually by multiplying the individual scores assigned to each assessment dimension. In a next step, each attribute is assigned to a criticality category (e.g., low-moderate-high, each associated with a predefined scoring range) based on the attribute’s aggregate score. Examples of quantitative risk scoring techniques are described in the A-Mab case study (13). A variation of Tool 1 described in this case study is used by Sandoz (11,12), the first company to receive approval for a biosimilar product in the USA and EU. Qualitative approaches equally define different levels within each assessment dimension (although these are not associated with a numerical score), but do not make use of a calculation for determination of the overall criticality “score” (or category). Instead, visual approaches (such as heat maps based on relative rankings) may be used (10). Alternatively, rules can be defined that clarify how the assessment dimensions must be integrated to determine the overall criticality category. Whichever approach is chosen, the end-result will consist of the complete list of quality attributes grouped into a number of criticality categories.
Overall, each criticality assessment model has its strengths and weaknesses. Quantitative models have the advantage of offering a rank order of all quality attributes, meaning that attributes within a specific criticality category are further ranked in order of criticality. This approach allows sub-sampling of attributes from a category for specific purposes, which may prove useful when tiering the attributes. A disadvantage may be that the overall criticality score (and thus criticality category) is very sensitive to the individual scores assigned to each assessment dimension (due to multiplication of dimension scores) and a slight difference in interpretation may therefore result in a different categorization of the attribute. Consider a case where attributes are assessed using the assessment dimensions impact and uncertainty (refer to the A-Mab case study, quality attribute assessment Tool #1 (13), for details on this methodology), with criticality defined in categories of low (scores 0–29), moderate (scores 30–79), and high (scores 80–140). Take a quality attribute for which nonclinical data demonstrate this attribute decreases the biological activity of the product. In this example, an uncertainty score of 3 will be assigned (given the availability of nonclinical data for this molecule), but the impact may be scored differently depending on how critical the impact is judged to be by the company (e.g., low (“acceptable change,” score = 4) or moderate (“moderate change,” score = 12))—arguably a subjective decision. The resulting aggregate score (12 or 36) can therefore either lead to a designation of low or moderate criticality, which determines attribute control (including the tier it is assigned to for similarity assessment). A qualitative model may be more intuitive and less laborious, but bears the risk of less substantiated judgment: experience shows that qualitative assessments are often based on “expert opinions,” whereas quantitative tools tend to promote more intense research of information sources. However, this distinction is not absolute, and qualitative models can equally rely on substantial documentation of proposed criticality assignments.
We choose not to recommend a specific criticality assessment method, but emphasize that whichever model is preferred by the developer, it must be fit for purpose, compatible with the company’s internal practices, demonstrate comprehensive justification of scoring decisions, and ensure an appropriate assessment hierarchy for linked attributes. The latter point refers to several quality attributes that relate to a single critical attribute, whereby each individual quality attribute is independently scored as critical because of its link to the critical attribute. A typical example is aggregation: aggregates are often viewed as highly critical, particularly for therapeutic proteins that intend to substitute endogenous proteins (e.g., erythropoetin, insulin, somatotropin, G-CSF). Different quality attributes may increase the propensity of aggregation if inadequately controlled, such as protein conformation, protein modifications such as oxidation and deamidation, and intermolecular disulfide bonds. The criticality assessment should differentiate between attributes that are directly associated with an adverse clinical impact (i.e., aggregation—which can be measured by various analytical techniques), and attributes that have a relationship with the critical attribute (which should be assessed independently of their potential to promote aggregation). Another example of such linkage is the correlation between (a) ADCC (a direct measure of biological activity) and (b) the quality attributes FcγR3a binding (one of the effector functions involved in ADCC) and afucosylated glycan content (correlated with FcγR binding) (14). The developer should consider all the roles of these linked QAs, their correlation with each other, and the specific assays being used in designing a similarity strategy.
Case study example: Sandoz’ approach to attribute ranking (15)
The Sandoz approach to criticality assessment, as published for its candidate biosimilar product for etanercept, starts with assigning all quality attributes to one of the following categories: (a) product-related variants; (b) process-related impurities; and (c) potency, identity, strength and composition, and appearance and description. The criticality of each attribute is then assessed by different risk assessment tools. Specifically, Tools A and B calculate the criticality score as a function of impact and uncertainty. Tool C is used for attributes that are considered highly critical by default.
Tool A calculates the impact and uncertainty scores for product-related variants. The impact ranking assesses the known or potential influence on, or consequences for, the clinical performance, and generates a numerical impact factor (ranging from 2 to 20), by considering the attribute’s effect on the following:
1. Efficacy, either through clinical experience or results using the most relevant potency assay(s)
2. Pharmacokinetics/pharmacodynamics (PK/PD)
3. Immunogenicity (against the drug substance)
In addition to the impact factor, an uncertainty factor is assigned (ranging from 1 to 7). The uncertainty factor depends on the type and source of knowledge available for a quality attribute, e.g., sponsor or literature data on in vitro or in vivo studies, clinical data relating to the same attribute in a related molecule or even clinical data on the molecule of interest containing the variant in question. The category with highest calculated score determines the overall criticality score for the quality attribute.
Tool B assesses the criticality of process- and excipient-related impurities by a similar approach to that taken for active-ingredients in Tool A above. Tool B considers only the relevance to safety/immunogenicity for impact scoring (ranging from 2 to 20). Similarly, uncertainty scoring can take only three different default values: 1 (none/no uncertainty, e.g., generally regarded as safe), 3 (low, e.g., component used in previous processes), or 7 (high, e.g., no information available, new impurity).The Sandoz approach differs from the approach described in Tool 1 of the A-Mab case study for calculation of the overall criticality score. The A-Mab case study tool ranks quality attributes with known (regardless whether positive or negative) impact on clinical performance as lower than those for which there is uncertainty with respect to their impact on, e.g., efficacy or safety. This approach is difficult to justify from the perspective of patient interest, safety and well-being: those attributes certain to have an impact on clinical performance should be assigned the highest ranking, to ensure that they are well controlled and receive the highest attention in the similarity assessment and control strategy. In addition, product development would be mainly uncertainty-driven if those attributes with higher uncertainty received greater attention than those for which the impact is known. This uncertainty dilemma in the A-Mab case study Tool 1 is caused by the approach of simple multiplication of the criticality sub-scores to calculate the overall criticality score (criticality score = impact score × uncertainty score). Therefore, if two attributes have exactly the same impact score, the attribute with higher uncertainty would be ranked higher. To overcome this issue, Sandoz modified the equation and implemented a full-factorial function ensuring that those attributes with low uncertainty and limited impact on clinical performance receive the lowest score, followed by those with uncertainty about their low impact, attributes with high uncertainty about their high impact, and the highest scores reserved for attributes with low/no uncertainty about their known high impact. For example, consider two quality attributes for the same product. For attribute 1, clinical evidence exists of its very high impact on safety. This would translate in an impact score of 20 and an uncertainty score of 1. Attribute B is a new glycoform of the product, for which a very high impact on immunogenicity is assumed on theoretical grounds only (no experimental data available); an impact score of 20 and an uncertainty score of 7 are assigned. In this example, the A-Mab approach would rank attribute 2 (total score 140) higher in criticality than attribute 1 (total score 20). Sandoz’ approach used for criticality assessment of Erelzi’s quality attributes results in an inversed ranking: attribute 1 receives a score of 140, attribute 2 a score of 90.
The scoring methodology for Tool A and Tool B are depicted in Fig. 2 below.
In contrast to product-related variants and process- and excipient-related impurities, quality attributes related to potency, identity, strength and composition, and appearance and description are considered critical quality attributes by default, unless justified otherwise (Tool C). Different criticality categories (and therefore criticality scores) are assigned to these quality attributes depending on their individual impact on safety and efficacy and a written justification included for each.
Case study example: Amgen’s approach to attribute ranking (10)
Amgen’s approach of risk ranking involves relative prioritization of attributes that have the potential to impact potency, pharmacokinetics, safety, and immunogenicity. It also considers the potential risk for a difference based on process design (e.g., cell line differences, formulation, and process steps). Amgen employs a two-dimensional grid (Fig. 3) as a tool for criticality mapping. One axis scores the potential impact on similarity of potency, safety, or pharmacokinetics. The other axis considers risk to similarity due to differences in-process design. For example, attributes which could change during stability are considered of higher risk (e.g., HMWs, LMWs, disulfide modifications) than attributes highly controlled by process design (e.g., HCP, DNA).
TIER ASSIGNMENT (US FDA ONLY)
FDA recommends assigning each quality attribute to one of the three tiers, each of which is associated with a specific methodology of (statistical) assessment of the analytical data. Tiers are assigned based on a risk assessment that takes into consideration potential clinical impact and uncertainty (2). Tier assignment starts with criticality. Each criticality category should be mapped to one of the three assessment tiers defined by US FDA, whereby tier 1 is reserved for the most critical tests with direct impact to the mode of action, tier 2 is used for tests of moderate criticality, and tests for attributes of the lowest criticality rank are assigned to tier 3. Several additional factors should be considered to determine the ultimate tier assignment of a quality attribute or analytical test.
Before discussing these factors, we need to first emphasize that analytical measures, not quality attributes, are assigned to the different assessment tiers. We define analytical measure as the result of an analytical procedure that is proposed as a basis for determination of similarity between the proposed biosimilar and its reference product. For example, oligosaccharides profiling by HPLC will yield quantitative information on all enzymatically released glycan species that are quantifiable by the method, in addition to a chromatographic fingerprint of the glycan population. Each individual glycan, predefined groups of glycans (sialylated, afucosylated, etc.), and the chromatogram itself, all could be defined as analytical measures. This distinction is important: most quality attributes will be measured by several orthogonal analytical techniques (in line with US FDA’s recommendations (9)), but the agency does not expect that each individual test (or measure from that test) is subsequently assigned to the same tier.
Tests which only yield qualitative information, such as spectra or other analytical fingerprints (with no possibility of deconvolution), or are binary in nature (pass/fail, e.g., protein sequence), cannot be assigned to tier 1 or tier 2, regardless of their criticality. Attribute content levels are another factor of consideration. For example, if several or all results from analysis of the proposed biosimilar and/or reference product are below the quantitation limit of the most sensitive analytical method available for measurement of that attribute, then the data cannot be processed by US FDA’s recommended statistical tests, necessitating assignment to tier 3 (regardless of the criticality of the associated quality attribute). Tier assignment may also be influenced by analytical considerations. If, for example, different analytical measures controlling the same highly critical quality attribute were all assigned to tier 1, the probability of at least one false negative result (to be understood as the chance that non-equivalence is concluded, when the products are truly equivalent, i.e., a type II error) greatly increases. Therefore, a sponsor should propose to assign only one analytical measure per attribute to tier 1, while other analytical measures of the same attribute are assigned to tier 2, based on criteria such as method performance (sensitivity, precision), known method limitations; and the clinical relevance of the analytical information.
Once the tiers have been assigned and the study conducted, the results are evaluated. An earlier commentary by the AAPS Biosimilars Focus Group’s CMC-A subcommittee addresses this final step in the analytical similarity exercise, the statistical evaluation of analytical similarity results (1). In brief, tier 1 tests are assessed using mean equivalence testing while tier 2 tests are evaluated using a quality range approach. Results from tests assigned to tier 3 are compared using various forms of visual displays.
In conclusion, meaningful assignment of analytical similarity data to one of the US FDA’s recommended three tiers for assessment requires detailed knowledge of the type of information that is offered by each method, mapped against a predefined and criticality-ranked list of quality attributes. Each analytical measure used to demonstrate analytical similarity should be assigned to the appropriate tier through the application of a set of rules, based on criteria such as attribute criticality, analytical data type, attribute levels, and analytical considerations.
The selection and criticality ranking of quality attributes constitutes an essential step in early biosimilar development, informing both process development and controls as well as the design of the pivotal analytical similarity study. The attributes and associated analytical methods are selected based on careful review of the various sources of knowledge for the product or related molecules, including the biosimilar manufacturer’s own data. Subsequent appraisal of the known or suspected relationship between each quality attribute and the clinical performance of the product yields a rank order of attribute criticality. A number of approaches to criticality ranking have been described, each with their own set of advantages and disadvantages. Case examples of products approved by US FDA show that varying approaches to criticality ranking have been employed in their development, indicating that successful programs can take different, equally valid, paths.
In case of analytical similarity assessment programs intended to support US registration, individual tests need to be assigned to one of the three tiers, each associated with a specific methodology of data evaluation, as a final step in program design. Attribute criticality is only one among several factors that determine the ultimate assignment of each analytical test to the appropriate tier.
With the emergence of detailed regulatory standards and the wide availability of increasingly powerful analytical technology, the structural and functional comparison of a proposed biosimilar with its reference product continues to gain in importance. Sponsors are therefore encouraged to spend careful attention to the design of their in vitro study program and discuss their proposal with regulatory agencies prior to execution.
The authors wish to thank Thomas Stangler for his contributions to develop CQA assessment concepts and helpful discussions related to this manuscript.
Compliance with Ethical Standards
This article reflects the views of the authors and should not be construed to represent FDA’s views or policies.
Amino acid analysis
Antibody-dependent cell-mediated cytotoxicity
Asymmetrical flow FFF
Critical quality attribute
Dynamic light scattering
Differential scanning calorimetry
Enzyme-linked immunosorbent assay
Field flow fractionation
Fourier transform infrared
Hydrogen deuterium exchange
Hollow fiber flow FFF
Hydrophilic interaction chromatography
High-performance anion exchange chromatography
High-performance liquid chromatography
Imaged capillary isolectric focusing
Matrix-assisted laser desorption/ionization
Multi-angle light scattering
Tandem mass spectrometry
Nuclear magnetic resonance
Nanoparticle tracking analysis
Pulsed amperometric detection
Polyacrylamide gel electrophoresis
Sodium dodecyl sulfate
Size exclusion chromatography
Ultra-high-pressure liquid chromatography
- 2.U.S. FDA. Guidance for industry (draft): statistical approaches to evaluate analytical similarity. September 2017. https://www.fda.gov/downloads/Drugs/GuidanceComplianceRegulatoryInformation/Guidances/UCM576786.pdf. Accessed 1 Feb 2018.
- 3.EMA. EMA/CHMP/138502/2017: Reflection paper on statistical methodology for the comparative assessment of quality attributes in drug development (draft). March 23, 2017. http://www.ema.europa.eu/docs/en_GB/document_library/Scientific_guideline/2017/03/WC500224995.pdf. Accessed 22 Jan 2018.
- 7.U.S. FDA. Oncologic drugs advisory committee (ODAC) meeting briefing document ABP215 (proposed biosimilar to Avastin). July 13, 2017. https://www.fda.gov/downloads/AdvisoryCommittees/CommitteesMeetingMaterials/Drugs/OncologicDrugsAdvisoryCommittee/UCM566365.pdf. Accessed 26 Jan 2018.
- 8.EMA. EMA/CHMP/BWP/247713/2012: Guideline on similar biological medicinal products containing biotechnology-derived proteins as active substance: quality issues (revision 1). May 22, 2014. http://www.ema.europa.eu/docs/en_GB/document_library/Scientific_guideline/2014/06/WC500167838.pdf. Accessed 20 Dec 2017.
- 9.U.S. FDA. Guidance for industry: quality considerations in demonstrating biosimilarity of a therapeutic protein product to a reference product. April 2015. https://www.fda.gov/downloads/Drugs/GuidanceComplianceRegulatoryInformation/Guidances/UCM291134.pdf. Accessed 20 Dec 2017.
- 10.Karow, M. Applying risk ranking for similarity into the QTPP for antibody biosimilars. 2016 PDA Biosimilars Conference, Baltimore, MD, USA. June 20–21, 2016.Google Scholar
- 11.Sonderegger C. QTPP for biosimilars: from reference product data to biosimilarity criteria. 2016 PDA Biosimilars Conference, Baltimore, MD, USA. June 20–21, 2016.Google Scholar
- 12.Stangler T. What to control? CQAs and CPPs. 2011 BWP workshop on setting specifications, London, UK. September 9, 2011. http://www.ema.europa.eu/docs/en_GB/document_library/Presentation/2011/10/WC500115824.pdf. Accessed 15 Dec 2017.
- 13.CMC Biotech Working Group. A-Mab: a case study in bioprocess development (version 2.1). October 30, 2009. http://c.ymcdn.com/sites/www.casss.org/resource/resmgr/imported/A-Mab_Case_Study_Version_2-1.pdf. Accessed 1 Dec 2017.
- 15.U.S. FDA CDER. Quality review BLA761042 - GP2015 (Etanercept), Sandoz. 2016. Application number 761042Orig1s000 chemistry review(s). https://www.accessdata.fda.gov/drugsatfda_docs/nda/2016/761042Orig1s000ChemR.pdf. Accessed 29 Jan 2018.
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.