(652b) Building Confidence in Machine Learning Models: A Case Study in Deep Learning-Based Image Analytics for Characterization of Pharmaceutical Samples
AIChE Annual Meeting
2023
2023 AIChE Annual Meeting
Pharmaceutical Discovery, Development and Manufacturing Forum
Pharma 4.0 (Advanced Controls, Process Automation, Data Analytics, etc.) - Technologies & Infrastructure
Wednesday, November 8, 2023 - 12:51pm to 1:12pm
In the highly regulated pharmaceuticals industry, applying such innovative solutions go hand in hand with regulatory influence. This reinforces the need that these novel AI/ML tools and their internal decision-making mechanisms should be accessible, as much as possible, to various stakeholders including scientists, process engineers, and regulatory agencies. Historically, chemical engineers have dealt with this for a long time by incorporating physics into their models. However, model transparency is not a given in some modern AI/ML cases where models can be highly complex and accurate but at the expense of interpretability, including obscured correlations to underlying physicochemical phenomena. A key question that we must ask about every model is: What is the confidence in the model? The answer to which might go beyond simply reporting accuracy on some test data.
In this talk, we use a common case, developing ML models for data analysis in image-based characterization of particulate samples (e.g., amorphous solid dispersions), to highlight some of the aforementioned model confidence issues and provide suggestions for tackling them. We use this example to discuss some of the opportunities and pitfalls in using state-of-the-art models such as deep neural networks, including the common misunderstanding of them being fully âblack boxâ and âunexplainableâ models. We illustrate how model interpretation and competency methods such as Integrated Gradients can help explaining model predictions and provide insight into the modelâs internal decision-making processes. We argue that such insight is crucial for understanding model limitations, anticipating its potential future failure modes, and building confidence and trust for future wide-spread use. This understanding also allows incorporating the necessary changes in the model development process to appropriately deal with challenging future test cases.