K Number
K223754
Device Name
BraveCX
Manufacturer
Date Cleared
2023-11-09

(329 days)

Product Code
Regulation Number
892.2080
Panel
RA
Reference & Predicate Devices
AI/MLSaMDIVD (In Vitro Diagnostic)TherapeuticDiagnosticis PCCP AuthorizedThirdpartyExpeditedreview
Intended Use

BraveCX is a radiological computer-assisted triage and notification software that analyzes adult (≥18 years old) chest Xray images for the presence of pre-specified suspected critical findings (pleural effusion and/or pneumothorax). BraveCX uses an artificial intelligence algorithm to analyze images for features suggestive of critical findings and provides caselevel output available in the PACS/workstation for worklist prioritization or triage. As a passive notification for prioritization-only software tool within standard of care workflow, BraveCX does not send a proactive alert directly to the appropriately trained medical specialists. BraveCX is not intended to direct attention to specific portions of an image or to anomalies other than pleural effusion and/or pneumothorax. Its results are not intended to be used on a stand-alone basis for clinical decision-making.

Device Description

BraveCX is a Deep Learning Artificial Intelligence (AI) software that analyzes adult (≥18 years old) chest X-ray images for the presence of pre-specified suspected critical findings (pleural effusion and/or pneumothorax. It uses deep learning to analyze each image to identify features suggestive of pleural effusion and/or pneumothorax. Upon image acquisition from other radiological imaging equipment (e.g. X-ray systems), Anteroposterior (AP) and Posteroanterior (PA) chest X-Rays are received and processed by BraveCX. Following receipt of an image, BraveCX de-identifies a copy of each DICOM file and analyses it for features suggestive of pleural effusion and/or pneumothorax. Based on the analysis result, the software notifies PACS/workstation for the presence of the critical findings, indicated by "flag" or "(blank)". This allows the appropriately trained medical specialists to group suspicious exams together with potential for prioritization. Chest radiographs without an identified anomaly are placed in the worklist for routine review, which is the current standard of care. The intended user of the BraveCX software is a health care professional such as radiologist or another appropriately trained clinician. The software does not alter the order or remove cases from the reading queue. The software output to the user is a label of "flag" or "(blank)" that relates to the likelihood of presence of pneumothorax and/or pleural effusion. BraveCX platform ingests prediction requests with either attached DICOM images or DICOM UIDs referencing images already uploaded to DICOM storage. The results will be made available via a newly generated DICOM that is stored in DICOM storage or as a JSON file. The DICOM storage component may be a Picture Archiving and Communications (PACS) system or some other local storage platform. BraveCX works in parallel to and in conjunction with the standard of care workflow to enable prioritized review by the appropriately trained medical specialists who are qualified to interpret chest radiographs. As a passive notification for prioritization-only software tool within standard of care workflow, BraveCX does not send a proactive alert directly to the appropriately trained medical specialists who are qualified to interpret chest radiographs. BraveCX is not intended to direct attention to specific portions or anomalies of an image and it should not be used on a standalone basis for clinical decision-making. BraveCX automatically runs after image acquisition. It prioritises and displays the analysis results through the worklist interface of PACS/workstation. An on-device, technologist notification is generated within 15 minutes after interpretation by the user, indicating which cases were prioritized by BraveCX in PACS. The technologist notification is contextual and does not provide any diagnostic information. The on-device, technologist notification is not intended to inform any clinical decision, prioritization, or action.

AI/ML Overview

The provided text describes the BraveCX device, a radiological computer-assisted triage and notification software that analyzes adult chest X-ray images for the presence of suspected critical findings (pleural effusion and/or pneumothorax).

Here's a breakdown of the acceptance criteria and the study that proves the device meets them:

1. Table of Acceptance Criteria and Reported Device Performance

The acceptance criteria for the BraveCX device are not explicitly listed in a separate table as "acceptance criteria." However, based on the "Summary of results" in the "9. Non-Clinical Performance Data" section and the comparison to the predicate device, the implied acceptance criteria are:

  • For Pleural Effusion:
    • ROC AUC > 0.95
    • Sensitivity > 0.85 (implied by "lower bounds of both sensitivity and specificity are above 0.85")
    • Specificity > 0.85 (implied by "lower bounds of both sensitivity and specificity are above 0.85")
  • For Pneumothorax:
    • ROC AUC > 0.95
    • Sensitivity > 0.85 (implied by "lower bounds of both sensitivity and specificity are above 0.85")
    • Specificity > 0.85 (implied by "lower bounds of both sensitivity and specificity are above 0.85")
  • Device Performance Time (Time-to-notification): Comparable to the predicate device.

Reported Device Performance (External Independent Testing):

MetricPleural Effusion (BraveCX)Pneumothorax (BraveCX)
ROC AUC0.988 (95% CI: 0.9885-0.9887)0.972 (95% CI: 0.9727-0.9729)
Sensitivity92.62% (95% CI: 90.67%-94.27%)93.38% (95% CI: 92.23%-94.40%)
Specificity98.11% (95% CI: 97.33%-98.71%)97.27% (95% CI: 96.49%-97.92%)
Time-to-notification4.8-10.4 seconds (95% CI: 4.2-10.41s) for simultaneous prediction4.8-10.4 seconds (95% CI: 4.2-10.41s) for simultaneous prediction

Predicate Device Performance (Lunit INSIGHT CXR Triage, K211733):

MetricPleural Effusion (Predicate)Pneumothorax (Predicate)
ROC AUC0.9686 (95% CI: 0.9547 - 0.9824)0.9630 (95% CI: 0.9521 - 0.9739)
Sensitivity89.86% (95% CI: 86.72 - 93.00)88.92% (95% CI: 85.60 - 92.24)
Specificity93.48% (95% CI: 91.06 - 95.91)90.51% (95% CI: 88.18 - 92.83)
Time-to-notification20.76 seconds (95% CI: 20.23 - 21.28)20.45 seconds (95% CI: 19.99 - 20.92)

The BraveCX device’s performance metrics for ROC AUC, sensitivity, and specificity exceed the implied acceptance criteria (all > 0.95 for AUC and > 0.85 for sensitivity/specificity). The time-to-notification is also comparable to, and even faster than, the predicate device.

2. Sample Sizes and Data Provenance for the Test Set

  • Sample Size:
    • Pleural Effusion: n=2,509 images (with n=867 positive cases)
    • Pneumothorax: n=3,245 images (with n=2,114 positive cases)
  • Data Provenance: The study used the MIMIC Chest X-ray (MIMIC-CXR) Database v2.0.02, NIH Chest X-Ray dataset (NIH-CXR), and CheXpert dataset (Stanford Hospital). These datasets represent the US population. The specific institutions mentioned are Beth Israel Deaconess Medical Center in Boston, MA, NIH Clinical Center, and Stanford Hospital. The data is retrospective.

3. Number of Experts and Qualifications for Ground Truth

  • Number of Experts: Three.
  • Qualifications: Board-certified Radiologists with at least 10 years of experience in specialty radiology training.

4. Adjudication Method for the Test Set

The document states, "All images were manually labelled by three board-certified Radiologists with at least 10 years of experience in specialty radiology training." It does not explicitly specify an adjudication method like 2+1 or 3+1, but implies that the agreement among these three experts established the ground truth. There is no mention of a specific tie-breaking rule or consensus process beyond all three being involved in labeling.

5. Multi-Reader Multi-Case (MRMC) Comparative Effectiveness Study

No, a multi-reader multi-case (MRMC) comparative effectiveness study comparing human readers with AI vs. without AI assistance was not explicitly mentioned or performed. The study described is a standalone performance evaluation of the AI algorithm.

6. Standalone Performance (Algorithm Only) Study

Yes, a standalone study was performed. The "Non-Clinical Performance Data" section describes an "external independent testing to assess the performance of BraveCX." This is a standalone evaluation of the algorithm's performance without a human in the loop. The results (ROC AUC, sensitivity, specificity, and time-to-notification) are reported for the algorithm itself.

7. Type of Ground Truth Used

The ground truth used for the test set was expert consensus (manual labeling by three board-certified radiologists).

8. Sample Size for the Training Set

The document mentions "Model training, validation, and testing sets were generated by stratified random partitions of 80%, 10%, and 10% respectively." While the exact total number of images used for training across all datasets is not explicitly stated, it implies that the training set constituted 80% of the total dataset used for development.

For the internal independent testing set, it contained n=1,209 cases for pleural effusion and n=1,387 cases of pneumothorax. Assuming these numbers are part of the 10% split for testing for the dataset from NHS Greater Glasgow and Clyde, the training set for that specific dataset would be significantly larger (e.g., if 1209 cases were 10%, training would be 8x that).

The external validation (MIMIC, NIH, CheXpert) had a sample size for testing (2509 for pleural effusion, 3245 for pneumothorax), but the specific training set size for the model that produced these results is not directly stated in terms of an absolute number, only the proportion (80%).

9. How the Ground Truth for the Training Set Was Established

The ground truth for the training set was established through manual curation by three board-certified Radiologists with at least 10 years in specialist radiology training. This is explicitly stated: "Images used in the training, validation, and testing of the subject device were all manually-curated ground truths provided by three board-certified Radiologists with at least 10 years in specialist radiology training."

§ 892.2080 Radiological computer aided triage and notification software.

(a)
Identification. Radiological computer aided triage and notification software is an image processing prescription device intended to aid in prioritization and triage of radiological medical images. The device notifies a designated list of clinicians of the availability of time sensitive radiological medical images for review based on computer aided image analysis of those images performed by the device. The device does not mark, highlight, or direct users' attention to a specific location in the original image. The device does not remove cases from a reading queue. The device operates in parallel with the standard of care, which remains the default option for all cases.(b)
Classification. Class II (special controls). The special controls for this device are:(1) Design verification and validation must include:
(i) A detailed description of the notification and triage algorithms and all underlying image analysis algorithms including, but not limited to, a detailed description of the algorithm inputs and outputs, each major component or block, how the algorithm affects or relates to clinical practice or patient care, and any algorithm limitations.
(ii) A detailed description of pre-specified performance testing protocols and dataset(s) used to assess whether the device will provide effective triage (
e.g., improved time to review of prioritized images for pre-specified clinicians).(iii) Results from performance testing that demonstrate that the device will provide effective triage. The performance assessment must be based on an appropriate measure to estimate the clinical effectiveness. The test dataset must contain sufficient numbers of cases from important cohorts (
e.g., subsets defined by clinically relevant confounders, effect modifiers, associated diseases, and subsets defined by image acquisition characteristics) such that the performance estimates and confidence intervals for these individual subsets can be characterized with the device for the intended use population and imaging equipment.(iv) Stand-alone performance testing protocols and results of the device.
(v) Appropriate software documentation (
e.g., device hazard analysis; software requirements specification document; software design specification document; traceability analysis; description of verification and validation activities including system level test protocol, pass/fail criteria, and results).(2) Labeling must include the following:
(i) A detailed description of the patient population for which the device is indicated for use;
(ii) A detailed description of the intended user and user training that addresses appropriate use protocols for the device;
(iii) Discussion of warnings, precautions, and limitations must include situations in which the device may fail or may not operate at its expected performance level (
e.g., poor image quality for certain subpopulations), as applicable;(iv) A detailed description of compatible imaging hardware, imaging protocols, and requirements for input images;
(v) Device operating instructions; and
(vi) A detailed summary of the performance testing, including: test methods, dataset characteristics, triage effectiveness (
e.g., improved time to review of prioritized images for pre-specified clinicians), diagnostic accuracy of algorithms informing triage decision, and results with associated statistical uncertainty (e.g., confidence intervals), including a summary of subanalyses on case distributions stratified by relevant confounders, such as lesion and organ characteristics, disease stages, and imaging equipment.