K Number
K210666
Device Name
Chest-CAD
Date Cleared
2021-07-20

(137 days)

Product Code
Regulation Number
892.2070
Panel
RA
Reference & Predicate Devices
N/A
AI/MLSaMDIVD (In Vitro Diagnostic)TherapeuticDiagnosticis PCCP AuthorizedThirdpartyExpeditedreview
Intended Use

Chest-CAD is a computer-assisted detection (CADe) software device that analyzes chest radiograph studies using machine learning techniques to identify, categorize, and highlight suspicious regions of interest (ROI). Any suspicious ROI identified by Chest-CAD is assigned to one of the following categories: Cardiac, Mediastinum/Hila, Lungs, Pleura, Bones, Soft Tissues, Hardware, or Other. The device is intended for use as a concurrent reading aid for physicians. Chest-CAD is indicated for adults only.

Device Description

Chest-CAD is a computer-assisted detection (CADe) software device designed to assist physicians in identifying suspicious regions of interest (ROIs) in adult chest X-rays. Suspicious ROIs identified by Chest-CAD are assigned to one of the following categories: Cardiac, Mediastinum/Hila, Lungs, Pleura, Bones, Soft Tissues, Hardware, or Other. Chest-CAD detects suspicious ROIs by analyzing radiographs using deep learning algorithms for computer vision and provides relevant annotations to assist physicians with their interpretations.

For each image within a study, Chest-CAD generates a DICOM Presentation State file (output overlay). If any suspicious ROI is detected by Chest-CAD in the study, the output overlay for all images includes the text "ROI(s) Detected:" followed by a list of the category/categories for which suspicious ROI(s) were found, such as "Lungs, Bones". In addition, if suspicious ROI(s) are detected in the image, bounding boxes surrounding each detected suspicious ROI are included in the output overlay. If no suspicious ROI is detected by Chest-CAD in the study, the output overlay for each image will include the text "No ROI(s) Detected" and no bounding boxes will be included. Regardless of whether a suspicious ROI is detected, the overlay includes text identifying the X-ray study as analyzed by Chest-CAD and a customer configurable message containing a link to or instructions for users to access labeling. The Chest-CAD overlay can be toggled on or off by the physician within their Picture Archiving and Communication System (PACS) viewer, allowing for concurrent review of the X-ray study.

AI/ML Overview

Acceptance Criteria and Device Performance for Chest-CAD (K210666)

The Chest-CAD device is a computer-assisted detection (CADe) software that analyzes chest radiographs to identify, categorize, and highlight suspicious regions of interest (ROI). It is intended as a concurrent reading aid for physicians and is indicated for adults only.

1. Acceptance Criteria and Reported Device Performance

The acceptance criteria for Chest-CAD were established through both a standalone performance assessment and a multi-reader, multi-case (MRMC) comparative effectiveness study. No explicit numerical acceptance criteria were stated as targets for the standalone performance metrics (sensitivity, specificity, AUC). Instead, the performance demonstrated in the standalone test formed a basis for comparison and validation. For the MRMC study, the primary objective was to demonstrate superiority of aided reading over unaided reading in terms of AUC.

Table 1: Acceptance Criteria and Reported Device Performance

MetricAcceptance Criteria (Internal/Implicit)Standalone Device PerformanceMRMC Study Performance (Aided vs. Unaided)
Standalone Performance
Overall SensitivityHigh sensitivity is expected0.908 (95% CI: 0.905, 0.911)N/A
Overall SpecificityHigh specificity is expected0.887 (95% CI: 0.885, 0.889)N/A
Overall AUCHigh AUC is expected0.976 (95% CI: 0.975, 0.976)N/A
Clinical Performance (MRMC Study)Aided vs. Unaided
Reader AUCAided AUC > Unaided AUCN/AAided: 0.894 (95% CI: 0.879, 0.909)
Unaided: 0.836 (95% CI: 0.816, 0.856)
Reader SensitivityAided Sensitivity > Unaided SensitivityN/AAided: 0.856 (95% CI: 0.850, 0.862)
Unaided: 0.757 (95% CI: 0.750, 0.764)
Reader SpecificityAided Specificity > Unaided SpecificityN/AAided: 0.870 (95% CI: 0.866, 0.873)
Unaided: 0.843 (95% CI: 0.839, 0.847)

2. Sample Sizes and Data Provenance

Standalone Test Set:

  • Sample Size: 20,000 chest radiograph cases.
  • Data Provenance: From 12 hospitals, outpatient centers, and specialty centers in the United States. The study was retrospective.

MRMC Study Test Set:

  • Sample Size: 238 cases.
  • Data Provenance: From 9 hospitals, outpatient centers, and specialty centers in the United States. The study was retrospective.

3. Number and Qualifications of Experts for Ground Truth (Test Set)

  • Standalone Test Set: Not explicitly stated for the standalone test set. The document indicates "Suspicious ROIs were manually annotated and categorized by board-certified radiologists before the images were used for benchmarking." This implies that the ground truth for standalone testing was established by expert radiologists, but the number and specific qualifications (years of experience) are not detailed.
  • MRMC Study Test Set: "Each case was previously evaluated by a panel of U.S. board-certified radiologists who assigned a ground truth binary label indicating the presence or absence of a suspicious ROI for each Chest-CAD category." The exact number of experts in the panel is not specified, nor is their specific level of experience (e.g., 10 years).

4. Adjudication Method (Test Set)

  • The document implies a consensus-based adjudication method for the ground truth of the MRMC study cases, stating that "a panel of U.S. board-certified radiologists who assigned a ground truth binary label..." This strongly suggests that multiple radiologists reviewed the cases and reached a consensus for the ground truth. It does not explicitly state a 2+1 or 3+1 method but refers to a "panel" establishing the ground truth.

5. Multi-Reader Multi-Case (MRMC) Comparative Effectiveness Study

  • Yes, a fully-crossed MRMC retrospective reader study was conducted.
  • Effect Size of Human Reader Improvement with AI vs. Without AI Assistance:
    • Reader AUC: Improved from 0.836 (Unaided) to 0.894 (Aided). Effect size = 0.058 increase.
    • Reader Sensitivity: Improved from 0.757 (Unaided) to 0.856 (Aided). Effect size = 0.099 increase.
    • Reader Specificity: Improved from 0.843 (Unaided) to 0.870 (Aided). Effect size = 0.027 increase.
      The study concluded that the accuracy of readers was superior when aided by Chest-CAD.

6. Standalone (Algorithm Only) Performance Study

  • Yes, a standalone performance assessment was conducted.
  • Reported Performance:
    • Overall Sensitivity: 0.908 (95% CI: 0.905, 0.911)
    • Overall Specificity: 0.887 (95% CI: 0.885, 0.889)
    • Overall AUC: 0.976 (95% CI: 0.975, 0.976)
    • AUC by Category: Ranged from 0.921 (Mediastinum/Hila) to 0.994 (Hardware).
    • Sensitivity by Category: Ranged from 0.854 (Bones) to 0.967 (Hardware).
    • Specificity by Category: Ranged from 0.830 (Mediastinum/Hila) to 0.960 (Hardware).

7. Type of Ground Truth Used (Test Set)

  • Expert Consensus: For both the standalone testing and the MRMC study, the ground truth was established by board-certified radiologists who manually annotated and assigned binary labels for suspicious ROIs.

8. Sample Size for the Training Set

  • The document does not explicitly state the sample size for the training set. It refers to "deep learning algorithms for computer vision" but does not detail the dataset used for training or validation of these algorithms.

9. How the Ground Truth for the Training Set was Established

  • The document does not explicitly state how the ground truth for the training set was established. It only mentions that the device uses "machine learning techniques" and "deep learning algorithms." Typically, for such devices, the training set ground truth would also be established by expert radiologists, usually through a process similar to or more extensive than that used for the test sets. However, this is not detailed in the provided text.

§ 892.2070 Medical image analyzer.

(a)
Identification. Medical image analyzers, including computer-assisted/aided detection (CADe) devices for mammography breast cancer, ultrasound breast lesions, radiograph lung nodules, and radiograph dental caries detection, is a prescription device that is intended to identify, mark, highlight, or in any other manner direct the clinicians' attention to portions of a radiology image that may reveal abnormalities during interpretation of patient radiology images by the clinicians. This device incorporates pattern recognition and data analysis capabilities and operates on previously acquired medical images. This device is not intended to replace the review by a qualified radiologist, and is not intended to be used for triage, or to recommend diagnosis.(b)
Classification. Class II (special controls). The special controls for this device are:(1) Design verification and validation must include:
(i) A detailed description of the image analysis algorithms including a description of the algorithm inputs and outputs, each major component or block, and algorithm limitations.
(ii) A detailed description of pre-specified performance testing methods and dataset(s) used to assess whether the device will improve reader performance as intended and to characterize the standalone device performance. Performance testing includes one or more standalone tests, side-by-side comparisons, or a reader study, as applicable.
(iii) Results from performance testing that demonstrate that the device improves reader performance in the intended use population when used in accordance with the instructions for use. The performance assessment must be based on appropriate diagnostic accuracy measures (
e.g., receiver operator characteristic plot, sensitivity, specificity, predictive value, and diagnostic likelihood ratio). The test dataset must contain a sufficient number of cases from important cohorts (e.g., subsets defined by clinically relevant confounders, effect modifiers, concomitant diseases, and subsets defined by image acquisition characteristics) such that the performance estimates and confidence intervals of the device for these individual subsets can be characterized for the intended use population and imaging equipment.(iv) Appropriate software documentation (
e.g., device hazard analysis; software requirements specification document; software design specification document; traceability analysis; description of verification and validation activities including system level test protocol, pass/fail criteria, and results; and cybersecurity).(2) Labeling must include the following:
(i) A detailed description of the patient population for which the device is indicated for use.
(ii) A detailed description of the intended reading protocol.
(iii) A detailed description of the intended user and user training that addresses appropriate reading protocols for the device.
(iv) A detailed description of the device inputs and outputs.
(v) A detailed description of compatible imaging hardware and imaging protocols.
(vi) Discussion of warnings, precautions, and limitations must include situations in which the device may fail or may not operate at its expected performance level (
e.g., poor image quality or for certain subpopulations), as applicable.(vii) Device operating instructions.
(viii) A detailed summary of the performance testing, including: test methods, dataset characteristics, results, and a summary of sub-analyses on case distributions stratified by relevant confounders, such as lesion and organ characteristics, disease stages, and imaging equipment.