1. Home
  2. Member Resources
  3. Councils and Committees
  4. Artificial Intelligence (AI) in Pathology

Artificial Intelligence (AI) in Pathology

For More Information

Send us an email with any comments, inquiries and questions related to AI in pathology.

Send an email Right Arrow

The College of American Pathologists (CAP) is working to stay abreast of AI advancements in pathology. Here, we explore the transformative role of artificial intelligence in the field of pathology, offering valuable insights, educational materials, and practical guidance for pathologists at all levels. As AI technology continues to evolve, our commitment to advancing pathology practice through innovative tools and techniques remains at the forefront. Dive in to discover how AI can enhance diagnostic accuracy, streamline workflows, and ultimately improve patient care.

CAP Activities

The CAP is engaged in several activities targeting Artificial Intelligence and Machine Learning (AI/ML). The Informatics Committee has formed a Machine Learning Working Group focused on education and technical issues particularly related to verification and performance monitoring. This group is sharing its technical work with the FDA. The Information Technology Leadership Committee has formed an AI Project Team to ensure coordination and alignment of AI/ML activities across the organization and to provide reports to the BOG. An AI in Anatomic Pathology Work Group, reporting to the Council on Scientific Affairs, is developing use cases for AI/ML in pathology that may evolve into PT programs.

Externally, the CAP participates in a several organizations including the Alliance for Digital Pathology, a collaborative group interested in the evolution of regulatory science as it applies to digital pathology and AI. The CAP also works with the American College of Radiology Data Science Institute, a resource in understanding how radiologists are developing and using AI systems. In addition, the CAP is the Primary Secretariat to the Integrating the Healthcare Enterprise's International Pathology and Lab Medicine domain as well as DICOM Working Group 26: Pathology. These standards organizations are developing technical profiles for incorporation of AI/ML systems into healthcare that will be available to developers of AI/ML tools and systems.

Latest News

Peter McCaffrey, MD, MS; Ronald Jackups, MD, PhD; Jansen Seheult, MB, BCh BAO, MSc, MS, MD; Mark A. Zaydman, MD, PhD; Ulysses Balis, MD; Harshwardhan M. Thaker, MD, PhD; Hooman Rashidi, MD, MS; Rama R. Gullapalli, MD, PhD. Archives of Pathology & Laboratory Medicine.

Generative artificial intelligence (GAI) technologies are likely to dramatically impact health care workflows in clinical pathology (CP). Applications in CP include education, data mining, decision support, result summaries, and patient trend assessments. Read more in the Archives of Pathology & Laboratory Medicine.

Cecchini, M. J., Borowitz, M. J., Glassy, E. F., Gullapalli, R. R., Hart, S. N., Hassell, L. A., Homer, R. J., Jackups, R., McNeal, J. L., & Anderson, S. R. (2024). Harnessing the power of generative artificial intelligence in pathology education. Archives of Pathology & Laboratory Medicine

Generative AI presents a powerful tool kit for enriching pathology education, offering opportunities for greater engagement, accessibility, and personalization. Careful consideration of ethical implications, potential risks, and appropriate mitigation strategies is essential for the responsible and effective integration of these technologies. Future success lies in fostering collaborative development between AI experts and medical educators, prioritizing ongoing human oversight and transparency to ensure that generative AI augments, rather than supplants, the vital role of educators in pathology training and practice. Read more in Archives of Pathology & Laboratory Medicine. 

Jackson, B. R., Rashidi, H. H., Lennerz, J. K., & de Baca, M. E. (2024). Ethical and regulatory perspectives on generative artificial intelligence in pathology. Archives of Pathology & Laboratory Medicine.

The literature on the ethical management of artificial intelligence in medicine is extensive but is still in its nascent stages because of the evolving nature of the technology. Effective and ethical integration of GenAI requires robust processes and shared accountability among technology vendors, health care organizations, regulatory bodies, medical professionals, and professional societies. As the technology continues to develop, a multifaceted ecosystem of safety mechanisms and ethical oversight is crucial to maximize benefits and mitigate risks. Read more in Archives of Pathology & Laboratory Medicine. 

In the September 2024 issue of CAP Today, From the President's Desk discusses the rapid expansion of artificial intelligence (AI) in medicine, particularly in pathology. While AI tools have been approved since the mid-1990s, many pathologists feel apprehensive about their implications. The author compares current anxieties about AI to past concerns over technologies like immunohistochemical stains and molecular genomic analysis, which ultimately enhanced rather than replaced the role of pathologists. Read From the President's Desk in CAP Today. 

Artificial intelligence is an exciting technology impacting the practice of pathology. When applied to the area of precision medicine, especially to treat oncologic disease, even more frontiers open for providing better patient care. This episode, a joint effort between the Personalized Health Care Committee and Digital and Computational Pathology Committee, features Dr. Marilyn Bui and Dr. Eric Walk discussing machine learning developments and the future of AI in precision medicine. Listen to the podcast. 

Peter McCaffrey, MD, FCAP and M. E. de Baca, MD, FCAP discuss the AI and Readiness course they are leading at the 2023 Pathologists Leadership Summit. They share some of the potential benefits, limitations, and implications of AI's growth in pathology testing. Hear it here.

Artificial intelligence and machine learning techniques have the potential to transform the diagnostics of inflammatory bowel disease. Automated image analysis, disease classification, predictive modeling, and risk stratification are just a few of the emerging applications that can enhance diagnostic accuracy, improve patient management, and guide treatment decisions. However, it is important to acknowledge that AI algorithms should not replace the expertise of pathologists but rather complement their work, serving as decision support tools. Read the article. 

AI already is being explored in several pathology settings with substantial potential to enhance and personalize patient care. It is therefore reasonable to expect that AI integration into the clinical cytometry laboratory will have a similar impact for diagnosis, risk stratification, or outcome prediction for patients with hematologic disorders. Read the article. 

Definitions of Artificial Intelligence and Machine Learning

Artificial intelligence (AI) is the ability of computer software to mimic human judgment. Current AI systems carry out only very specific tasks for which they are designed, but they may integrate large amounts of input data to carry out these tasks quickly and accurately. The current excitement about AI is focused on machine learning (ML) systems and this domain is sometimes referred to as AI/ML. AI/ML systems may be trained using defined input data sets, which may include images, to associate patterns in data with clinical contexts such as diagnoses or outcomes. Once trained, AI/ML systems are used with new data to predict diagnosis or outcome in specific cases or carry out other useful tasks. To date, systems are limited in the range of diagnoses, predictions, and tasks covered, but can be impressively accurate within their defined scope.

  1.  Yu K-H, Beam AL, Kohane IS. Artificial intelligence in healthcare. Nature Biomedical Engineering. 2018;2:719-731.
  2. Topol EJ. High-performance medicine: The convergence of human and artificial intelligence. Nature Medicine. 2019;25:44-56.

Concept of Augmented Intelligence

The American Medical Association has popularized the term Augmented Intelligence to represent the use of AI/ML as a tool to enhance rather than replace human healthcare providers. The Augmented Intelligence concept is based on studies that integrate AI/ML with human experts in a synergistic workflow that achieves higher performance than either separately. In the pathology context, Augmented Intelligence brings the computational advantages of AI/ML into the clinical and laboratory setting in the form of supportive tools that can enhance pathologists’ diagnostic capabilities by, for example, suggesting regions of interest or counting elements on a slide, or providing decision support to inform clinical judgment.

  1. AMA. Artificial Intelligence in Medicine
  2. Colling R, Pitman H, Oien K et al. Artificial intelligence in digital pathology: a roadmap to routine use in clinical practice. J Pathol. 2019;249:143-150.

How AI/ML may be used in Pathology

Pathologists who are interested in AI/ML envision a variety of tools that may provide increased efficiency and diagnostic accuracy in the pathologist’s daily diagnostic workflow. As noted above, tools for the pathologist could scan slides to count elements such as lymph node metastases, mitoses, inflammatory cells, or pathologic organisms, presenting results at sign-out and flagging examples for review. AI/ML tools could also flag regions of interest on a slide or prioritize cases based on slide content. Studies to date have shown promise for automated detection of foci of cancer and invasion, tissue/cell quantification, virtual immunohistochemistry, spatial cell mapping of disease, novel staging paradigms for some types of tumors, and workload triaging. Future systems may be able to correlate patterns across multiple inputs from the medical record, including genomics, allowing a more comprehensive prognostic statement in the pathology report.

  1. Pantanowitz L, Quiroga-Garza GM, Bien L et al. An artificial intelligence algorithm for prostate cancer diagnosis in whole slide images of core needle biopsies: a blinded clinical validation and deployment study. The Lancet Digital Health. 2020;2:e407-e416.
  2. Colling R, Pitman H, Oien K et al. Artificial intelligence in digital pathology: a roadmap to routine use in clinical practice. J Pathol. 2019;249:143-150.
  3. Campanella G, Hanna MG, Geneslaw L, et al. Clinical-grade computational pathology using weakly supervised deep learning on whole slide images. Nat Med. 2019;25(8):1301–1309.
  4. Rashidi HH, Tran NK, Betts EV, Howell LP, Green R. Artificial intelligence and machine learning in pathology: The present landscape of supervised methods. Acad Pathol. 2019;6:2374289519873088.
  5. Mezheyeuski A, Bergsland CH, Backman M, et al. Multispectral imaging for quantitative and compartment-specific immune infiltrates reveals distinct immune profiles that classify lung cancer patients. J Pathol. 2018;244(4):421–431.
  6. Wilkes EH, Rumsby G, Woodward GM. Using Machine Learning to Aid the Interpretation of Urine Steroid Profiles. Clin Chem. 2018;64:1586-1595.
  7. Arnaout R. Machine Learning in Clinical Pathology: Seeing the Forest for the Trees. Clin Chem. 2018;64(11):1553–1554.
  8. Cabitza F, Banfi G. Machine learning in laboratory medicine: waiting for the flood?. Clin Chem Lab Med. 2018;56(4):516–524.

Ethical use of AI in Healthcare

The need for large sets of patient data to train AI/ML algorithms raises issues of patient consent, privacy, data security, and data de-identification in the production of AI/ML systems. There is also an ethical duty to review algorithms prior to implementation and verify their performance at deployment to ensure that they are safe, efficacious, and reliable. Recent experience has shown that subtle biases may be incorporated into training data and influence the performance of the resulting systems; these must be mitigated, and training data must reflect the diversity of the patient population that the AI/ML systems are intended to serve. An algorithm trained without using best practices for representing ethnic groups, socioeconomic classes, ages, and sex may limit system generalizability to these patient populations in real-world settings and exclude (or harm) these groups inadvertently. The "black box" nature of some popular algorithms (not revealing the data patterns associated with particular predictions) combined with the natural proprietary orientation of system vendors may lead to transparency problems and difficulty checking the algorithms by independent interpretation. Finally, the human resource toll of AI/ML must be considered: deskilling of the workforce through dependence on AI/ML must be mitigated and there will be a need to repurpose job roles to adapt to increasing automation.

  1. Jackson BR, Ye Y, Crawford JM et al. The Ethics of Artificial Intelligence in Pathology and Laboratory Medicine: Principles and Practice. Academic Pathology. 2021;8:237428952199078.
  2. Keskinbora KH. Medical ethics considerations on artificial intelligence. J Clin Neurosci. 2019;64:277-282.
  3. O’Sullivan S, Nevejans N, Allen C et al. Legal, regulatory, and ethical frameworks for development of standards in artificial intelligence (AI) and autonomous robotic surgery. Int J Med Robot. 2019;15:e1968.

Regulation of Artificial Intelligence and Machine Learning

The training and use of AI/ML algorithms introduce a fundamentally new kind of data analysis into the healthcare workflow that requires an appropriate regulatory framework. By virtue of their influence on pathologists and other physicians in the selection of diagnoses and treatments, the outputs of these algorithms can critically impact patient care. The data patterns identified by these systems are often not exact: there is no perfect separation of classes or predictions. Thus, there are analogies with sensitivity, specificity, and predictive value of other complex tests performed by clinical laboratories. However, in machine learning the patterns in data are identified by software and often are not explicitly revealed. Biases or subtle errors may be incorporated inadvertently into machine learning systems, and these must be identified and mitigated prior to deployment. Naturally occurring changes in healthcare context, such as case mix changes, updated tests or sample preparation, or new therapies, may also change the input data profile and reduce the accuracy of a previously well-functioning machine learning system. 

An effective and equitable regulatory framework for machine learning in healthcare will 1) define requirements based on risk, i.e., tailored to the likelihood and magnitude of possible harm from each machine learning application, 2) require best practices for system development by vendors including bias assessment and mitigation, 3) define appropriate best practices for verification of system performance at deployment sites, ie, local laboratories, 4) define best practices for monitoring the performance of machine learning systems over time and mitigating performance problems that develop, and 5) clearly assign responsibility for problems if and when they occur.

The development of this framework is in its early stages. To date, the White House has released draft guidance for regulation of artificial intelligence applications that provides a set of high-level principles to which a regulatory framework in any domain should adhere. Specific to healthcare, the FDA has released proposals for processes leading to approval or clearance of machine learning software for use as a medical device. None of these proposals yet addresses best practices for local performance verification and monitoring of machine learning systems analogous to CLIA-mandated laboratory test performance requirements. The CAP regards this omission as a gap in current regulatory planning for machine learning in healthcare and is promoting the development of a more complete regulatory framework that will include guidance, approved methods, and best practices for local laboratories in deploying machine learning tools as they become available.

  1.  Shulz WL, Durant TJS, Krumholz HM. Validation and regulation of clinical artificial intelligence. Clin Chem 2019;65:1336-1337.
  2. Allen TC. Regulating artificial intelligence for a successful pathology future. Arch Pathol Lab Med 2019;143(10):1175.
  3. Office of Management and Budget. Guidance for regulation of artificial intelligence applications. White House Memo. 2020;Jan 7:1-15.
  4. FDA. Proposed regulatory framework for modifications to artificial intelligence/machine learning (AI/ML)-based software as a medical device (SaMD): Discussion paper and request for feedback. 2019;1-20.
  5. FDA. Artificial Intelligence/Machine Learning (AI/ML)-Based Software as a Medical Device (SaMD) Action Plan. 2021; Jan 12:1-7.