An Introduction to the Method Lifecycle Management Approach

In this Q&A from our popular webinar, learn about an analytical procedure implemented by the pharmaceutical industry and how it can benefit your lab

17 Apr 2019
Sarah Thomas
Associate Editor

The pharmaceutical industry relies on data generated by analytical methods for many critical decisions. Traditional approaches consist of distinct exercises — method development, validation, transfer, verification — with limited understanding of the effect of variation on method performance.

The analytical Method Lifecycle Management (MLCM) approach is a holistic approach encompassing all activities, from method development to validation, routine use, change control and retirement of the method. This enhanced approach improves the method understanding and performance, leads to fewer out-of-specifications results (OOS), facilitates method transfer and has the potential to lessen the regulatory burden.

In one of our most popular webinars to date, now available on demand, experts from the USP and the pharmaceutical industry provide an overview of the MLCM approach and how it is being implemented.

Dr. Horacio Pappa, Director of General Chapters, United States Pharmacopeia, shares key concepts of MCLM and the current thinking of the USP, FDA and ICH, while Dr. Alexander Schmidt, Founder and CEO, and Mijo Stanic, Cofounder, General Manager and Technical Director, Chromicent Pharmaceutical Services, present three case studies of how the approach is already being implemented in a pharmaceutical research setting, utilizing UHPLC and convergence chromatography technology from Waters Corporation.

If you missed the webinar, don’t worry! You can now watch it on demand at a time that suits you.

During the live webinar we received many questions from our audience on this trending topic. Read on for insights into the MCLM approach from Dr. Horacio Pappa, Dr. Alexander Schmidt, Mijo Stanic and Greg Martin, Vice Chair of General Chapters, United States Pharmacopeia and President, Complectors Consulting, LLC.

With regards to analytical target profile (ATP), do you think it's important to qualify the method and determine target measurement uncertainty (TMU) acceptance criteria at each step of the pharmaceutical development?

Dr. Horacio Pappa, Director of General Chapters, United States Pharmacopeia

HP: Keep in mind that the USP usually focuses on those methods that are used in the finalization of the development, i.e., when the product is already on the market. Having said that, the answer is usually no. You don't need to put a lot of effort into methods that are developed during the process of product development. It's the same pilot that we use now for classical validation. The big effort in the validation is when the method is already defined with the product already developed.

How do I define acceptable TMU for Class I (Assay) and Class II (Impurity) analytical procedures?

HP: This would depend on how critical the method is. I recommend reading this article.

ICH Q12 defines a concept called Established Conditions. How is this concept related to ATP?

HP: Established Conditions is a regulatory term. At this point, it is outside the USP discussion.

GM: What I can say is that the analytical target profile (ATP) will essentially establish some system suitability requirements. For example, an ATP for a degradant method might state: “The method is capable of separating the active drug from degradant A and degradant B with a resolution of NLT 1.5, with the ability to quantitate peaks at or above the reporting threshold of 0.10% of claim.”

  • In this example, the system suitability test should include resolution requirements for active, deg A and deg B, and sensitivity requirements for a 0.10% std (probably signal-to-noise ratio of NLT 10).

I guess it could be interpreted that the MLCM establishes the system suitability requirements.

In Quality-by-Design (QbD) for methods, how do you define precision for samples and accuracy?

HP: This is not just for Quality-by-Design, but also included in the approach we took in our new Chapter, <1210>, which proposes a theoretical method for procedure validation. We promote the use of the total error approach where you can have accuracy and precision all together as one parameter. So, no matter how high or low your precision or accuracy is, when you put them together, the total error is within what you define in your ATP.

How is QbD considered in the change permitted for the Chapter <621>?

HP: That's a good question. In the far future, when the methods are developed and validated using QbD concept, the information we will obtain at the pharmacopeia in terms of the performance of the method — the potential accuracy and precision of the method, meeting the appropriate system strategy requirements — will be very handy and of high level in the process of creating the monograph.

So, in other words, all the adjustments that we proposed in <621> will not be needed anymore, because the monograph will contain all the information necessary for these acceptable adjustments in the method. The adjustments in <621> were introduced as a temporary solution, because of the lack of specific information on the design space of the method.

If there are three laboratories in one firm (R&D and QC in two different locations), do you agree that methods validated out of house should be verified in all three laboratories, or, alternatively, if validated in R&D, they should be transferred and verified in both QC labs?

HP: I would say yes. The basic concept is that you will validate the method that is not in the pharmacopeia, as it's not been validated before; you verify the methods that are in the compendia; and you transfer the methods that are not in the compendia.

However, when you are moving compendial methods from one lab to the other, it's not a clear line between verification and transfer. What is important, I think, is that the verification activity is laboratory-dependent because, within the verification process, you try to demonstrate that the conditions where the method will be performed are suitable. And that includes not just the analytical metrics of your sample, but also the laboratory conditions.

It was mentioned that linearity quantification limit (QL) is part of the method development and that only precision and accuracy are needed for the validation. Does that mean that the official validation protocol would only include tests for precision and accuracy? And, what would be included in the final validation report needed for filing of the method?

HP: First, I need to clarify that the approach discussed today is an idea/model. There is nothing on the near horizon that indicates that you will have to change your filings or even the validation protocol that you have to submit to the pharmacopeia.

What some groups believe is that when you define the total error in the range that the method needs to operate, through accuracy and precision, there is nothing else needed to define the performance of the method in that region. Things like linearity and understanding linearity as the response function of the method are secondary parameters of the main accuracy and precision combination. So, you define those during the method development based on your ATP and then the validation or the qualification, depending on what would be the final terminology, would be just accuracy and precision.

Again, as I say in my presentation, this is a work in progress and not everybody agrees on the same concept, so I tried to present all the current ideas.

GM: Theoretically, all testing, such as linearity, robustness, etc. (as stated in ICH Q2) are performed as part of method understanding and development in stage 1. So, when the method is ready to be in stage 2, i.e. validation, accuracy and precision are only needed to confirm the method is performing consistently.

What parameters should I consider for impurity method verification?

HP: In general, accuracy, precision and LoQ.

What should I consider when monitoring method performance during MLCM? What parameters should be measured and compared, and how often?

AS/MS: Please see USP chapter <1010> – Analytical Data, Interpretation and Treatment for details of which parameters can be used for monitoring.

GM: It depends on the method, but examples of parameters to trend are; resolution, RT, peak area, peak tailing, efficiency, peak resolution, peak response, system pressure.

Are draft versions of ICH Q2(R2) and ICH Q14 available?

HP: Not yet. The expectation is that the first readable version will be discussed at the next meeting in June in Amsterdam. And, after that, I think that you should expect to see the document on the ICH website in probably a couple of months.

In what time frame will USP <1220> be implemented?

HP: A new expert panel will start its work in April, with the goal of presenting in PF <1220> in 2019-2020.

Dr. Alexander Schmidt, Founder and CEO, Chromicent Pharmaceutical Services

For a newly developed method, how should method validation be performed with the MCLM approach? What about compendial methods? What about methods that have been validated but are being transferred?

AS/MS: Currently, USP chapters <1224>, <1225> and <1226> as well as the ICH Q2(R) guideline are valid and should be followed. A revision of Q2 is planned and will have details on how to proceed.

How would one apply QbD for very early stage method development projects?

AS/MS: Please see our publication “Lifecycle Management for analytical methods” in Journal of Pharmaceutical and Biomedical Analysis 147 (2018) 506–517 for details.

If we are following a vendor method for impurities, how do I find the limit of quantitation (LoQ)?

AS/MS: Just make a solution of the API of 0.05% (or whatever your LoQ is) and inject. If the s/n is >1:10 the requirement of LoQ is fulfilled.

How does the continued method improvement fit with the requirement that every change in the validated method requires a risk assessment (RA) submission, which often impacts the supply chain for several months?

AS/MS: Please refer to the concept of Established Conditions and post-approval change management protocol of the ICH Q12 guideline which will guide you through this process.

Mijo Stanic, Cofounder, General Manager and Technical Director, Chromicent Pharmaceutical Services

What is your strategy to qualify for method operable design region (MODR)? The validation at each extreme point of the MODR is challenging for pharmaceutical companies.

MS: After we receive the design space, it's often only a model. We then perform verification points – the working point and six points around the working point. Sometimes we also perform verification points outside of the MODR or design space also, to check if the knowledge space shows not only where it's good, but also points where the method is not good. After the validation, we don't perform any robustness tests because we have performed them at the end of the development.

How do you design a MODR when the method is new? I.e., for an innovative drug that has no existing compendial monograph.

AS/MS: Design of Experiments will help you to define a method operable design region. From there, you would then choose a working point within the MODR and perform robustness testing.

Are you able to peak track unknown impurities/unknown masses, at the 0.05% level relative to the main peak with the Waters QDa Mass Detector?

MS: It depends, of course, on the analyte and the method. Usually, yes. But sometimes it also happens that higher peaks are not detectable with MS. They have to be ionized, and everything depends on their ionization.

What Analytical Quality by Design (AQbD) software do you use?

AS/MS: We use Fusion software, DryLab software, and Design Expert for AQbD.

Can you import chromatographic data directly into DryLab to account for accurate modeling of peak widths?

AS/MS: Yes, you can. Please export a chromatogram in AIA format from your CDS and import directly into Drylab.

Which modeling system do you recommend, DryLab or Fusion?

MS: That's a difficult question. It depends what you want to do. All software has its benefits. In DryLab, you can change many parameters, for example, radiant time, per cent start, per cent end. You can implement gradient steps, and you can change your column dimensions, the flow rate, the dwell volume. After you receive knowledge space and you can see what the chromatogram ran. In Fusion, you have the benefit that you can implement every result you receive in your chromatographic system. So, it doesn't matter if it is tailing factor; peak height, recovery, resolution. Overall, it depends on the user, and the focus.

How much total analyst or manager time was required to address the issues in the case studies presented?

AS: For the Avastin case study, we worked on it for two weeks. And for the pramipexole case study, it was a little bit quicker – I think it was done in two days or so.

MS: For the third case study, I needed two weeks. All three cases studies presented are very easy. Sometimes, we need up to two months for very difficult method development.

Missed the webinar? Watch it on demand now>>

SelectScience runs 3-4 webinars a month across various scientific topics, discover more of our upcoming webinars>>

ACQUITY UPC² System

Waters

Convergence Chromatography with the ACQUITY UPC² System Bring streamlined workflow, expanded selectivity, and an orthogonal technique to your laboratory. Convergence chromatography (CC) is a broad-based, complementary analytical platform that is taking its place alongside of LC and GC as one of the three essential separation technologies for modern laboratory analysis. This technique provides another, reliable choice as an analytical platform that is orthogonal to reversed-phase LC (RPLC), while significantly streamlining the entire analytical workflow. In convergence chromatography, the separation is achieved by manipulating the density and composition of a supercritical fluid-based mobile phase. Because of the very high diffusivity of the mobile phase, can be achieved. Additionally, the diversity of stationary phase and mobile phase (co-solvent) options provides access to the largest selectivity space available to any separation technique. The ACQUITY UPC² System is a holistically designed UltraPerformance Convergence Chromatography (UPC²) system that utilizes liquid CO2 as the primary mobile phase. The system leverages the chromatographic principles and selectivity of normal-phase chromatography while providing the ease-of-use of RPLC, enabling you to address routine and complex separations challenges. ACQUITY UPC² System: Discover what's been missing from your analytical toolbox With the ACQUITY UPC² System, scientists can now separate, analyze, and understand compounds that have challenged all other LC and GC technologies including: Hydrophobic and chiral compounds Lipids Thermally-labile samples Polymers It's also the perfect complement to MS due to its low solvent load and high resolution, narrow peaks and fast separations. ACQUITY UPC² System Features: Compressed carbon dioxide (CO2), the primary mobile phase, is less costly and less toxic than liquid mobile phases or carrier gases. Variable volume injection delivers 0.5 µL to 10 µL as well as the ability to match column and injection volume with minimal sample loss and need to change loops. Higher throughput and productivity allows you to analyze more samples per day with faster re-equilibration of columns and shorter cycle times, increasing laboratory productivity. Co-solvent and column switching provides flexibility in method development for faster solvent screening and faster column screening.

(18)

ACQUITY QDa Mass Detector

Waters

The ACQUITY QDa® Mass Detector gives you certainty above all else. You can minimize the risk of unexpected coelutions or components and confirms trace components with the analytical confidence of mass detection. The ACQUITY QDa® Detector is a mass detector built around the needs of analytical scientists for chromatographic analysis. Robust, reliable and requiring no sample adjustments, it just integrates with your current LC, ACQUITY UPLC, ACQUITY UPC2 and purification systems. You'll enhance the analytical value and productivity of each analysis, and you won't need to run all the additional assays or time-consuming alternative techniques or even wait for results from specialist labs that cost your laboratory productivity. With the ACQUITY QDa® Detector, you now have information-rich mass spectral data to complement data from your current Waters® optical detectors, including the ACQUITY UPLC PDA, TUV, ELS, and FLR detectors, as well as the ACQUITY UPC² PDA Detector. This mass spectral information integrates seamlessly into the same workflow; routinely giving you more complete separation characterization. Features: Intuitive operation More information from every sample Effortless integration Increased efficiency Resolving complexity Applications: Food & Environment Life Sciences and Pharmaceuticals Chemical

(20)

Links

Tags