Search Filters

Search Results

Found 1 results

510(k) Data Aggregation

    K Number
    K203517
    Device Name
    Saige-Q
    Manufacturer
    Date Cleared
    2021-04-16

    (137 days)

    Product Code
    Regulation Number
    892.2080
    Reference & Predicate Devices
    Why did this record match?
    Device Name :

    Saige-Q

    AI/MLSaMDIVD (In Vitro Diagnostic)TherapeuticDiagnosticis PCCP AuthorizedThirdpartyExpeditedreview
    Intended Use

    Saige-Q is a software workflow tool designed to aid radiologists in prioritizing exams within the standard-of-care image worklist for compatible full-field digital mammography (FFDM) and digital breast tomosynthesis (DBT) screening mammograms. Saige-Q uses an artificial intelligence algorithm to generate a code for a given mammogram, indicative of the software's suspicion that the mammogram contains at least one suspicious finding. Saige-Q makes the assigned codes available to a PACS/EPR/RIS/workstation for worklist prioritization or triage.

    Saige-Q is intended for passive notification only and does not provide any diagnostic information beyond triage and prioritization. Thus, it is not intended to replace the review of images or be used on a stand-alone basis for clinical decision-making. The decision to use Saige-Q codes and how to use those codes is ultimately up to the interpreting radiologist. The interpreting radiologist is reviewing each exam on a diagnostic viewer and evaluating each patient according to the current standard of care.

    Device Description

    Saige-Q is a software workflow device that processes Digital Breast Tomosynthesis (DBT) and Full-Field Digital Mammography (FFDM) screening mammograms using artificial intelligence to act as a prioritization tool for interpreting radiologists. By automatically indicating whether a given mammogram is suspicious for malignancy. Saige-Q can help the user prioritize or triage cases in their worklist (or queue) that may benefit from prioritized review.

    Saige-Q takes as input a set of x-ray mammogram DICOM files from a single screening mammography study (FFDM or DBT). The software first checks that the study is appropriate for Saige-Q analysis and then extracts, processes and analyses the DICOM images using an artificial intelligence algorithm. As a result of the analysis, the software generates a Saige-Q code indicating the software's suspicion of the presence of findings suggestive of breast cancer. For mammograms given a Saige-Q code of "Suspicious," the software also generates a compressed preview image, which is for informational purposes only and is not intended for diagnostic use.

    The Saige-Q code can be viewed by radiologists on a picture archiving and communication system (PACS), Electronic Patient Record (EPR), and/or Radiology Information System (RIS) worklist and can be used to reorder the worklist. As a software-only device, Saige-Q can be hosted on a compatible host server connected to the necessary clinical IT systems such that DICOM studies can be received and the resulting outputs returned where they can be incorporated into the radiology worklist.

    The Saige-Q codes can be used for triage or prioritization. For example, "Suspicious" studies could be given prioritized review. With a worklist that supports sorting, batches of mammograms could also be sorted based on the Saige-Q code.

    AI/ML Overview

    Here's a breakdown of the acceptance criteria and the study proving the device meets them, based on the provided text:

    Acceptance Criteria and Device Performance

    1. Table of Acceptance Criteria and Reported Device Performance

    Acceptance CriterionSaige-Q FFDM Performance (Reported Value)Saige-Q DBT Performance (Reported Value)BCSC Data (Baseline/Target)Predicate Device (cmTriage)
    Overall AUC0.966 (95% CI: [0.957, 0.975])0.985 (95% CI: [0.979, 0.990])>0.95 (QFM product code requirement for effective triage)Meets or exceeds predicate performance
    Specificity at 86.9% Sensitivity92.2% (95% CI: [90.2%, 93.8%])98.3% (95% CI: [97.3%, 99.0%])>80% CI-
    Sensitivity at 88.9% Specificity91.2% (95% CI: [88.4%, 93.4%])95.7% (95% CI: [93.6%, 97.2%])>80% CI-
    Median Processing Time15.5 seconds196.8 secondsWithin clinical operational expectations-
    Performance by Lesion Type (Soft Tissue Densities) - AUC0.964 (95% CI: [0.954, 0.974])0.983 (95% CI: [0.977, 0.990])Similar performance across subcategories-
    Performance by Lesion Type (Calcifications) - AUC0.973 (95% CI: [0.958, 0.988])0.989 (95% CI: [0.983, 0.996])Similar performance across subcategories-
    Performance by Breast Density (Dense) - AUC0.959 (95% CI: [0.945, 0.973])0.980 (95% CI: [0.971, 0.988])Similar performance across subcategories-
    Performance by Breast Density (Non-Dense) - AUC0.972 (95% CI: [0.961, 0.984])0.988 (95% CI: [0.981, 0.996])Similar performance across subcategories-

    2. Sample Size Used for the Test Set and Data Provenance

    • FFDM Study Test Set:
      • Malignant Exams: 501
      • Normal Exams: 832
      • Total: 1333
    • DBT Study Test Set:
      • Malignant Exams: 517
      • Normal Exams: 1011
      • Total: 1528
    • Data Provenance:
      • Country of Origin: United States (across two states)
      • Retrospective or Prospective: Retrospective
      • Sites: Data was collected from eight clinical sites for FFDM and six clinical sites for DBT. DeepHealth had never collected data from these sites previous to this study for either training or testing, ensuring an independent test set.

    3. Number of Experts Used to Establish the Ground Truth for the Test Set and Qualifications

    • Number of Experts: Two independent expert radiologists.
    • Qualifications of Experts: The document does not explicitly state the qualifications (e.g., years of experience) of the expert radiologists.

    4. Adjudication Method for the Test Set

    • Adjudication Method: 2+1 (Two independent expert radiologists reviewed each case. If discordance was observed between the two initial readers, an adjudicator was used to establish the final reference standard).

    5. Multi-Reader Multi-Case (MRMC) Comparative Effectiveness Study

    • Was an MRMC study done? No, the document describes retrospective, blinded, multi-center studies to evaluate the standalone performance of Saige-Q. It does not mention a comparative effectiveness study involving human readers with and without AI assistance.
    • Effect Size of Human Improvement with AI vs. Without AI Assistance: Not applicable, as no MRMC study was conducted to assess human reader improvement with AI assistance.

    6. Standalone (Algorithm Only) Performance Study

    • Was a standalone study done? Yes, the document explicitly states: "DeepHealth conducted two retrospective, blinded, multi-center studies to evaluate the standalone performance of Saige-Q..."

    7. Type of Ground Truth Used

    • Ground Truth Type:
      • Malignant Exams: Confirmed using pathology reports from biopsied lesions.
      • Normal Exams: Confirmed with a negative clinical interpretation (BI-RADS 1 or 2) followed by another negative clinical interpretation at least two years later.
      • Expert Consensus: Each case in the test set was reviewed by two independent expert radiologists (and an adjudicator if discordance was observed) to establish the reference standard for each case, building upon the pathology/clinical follow-up.

    8. Sample Size for the Training Set

    • The document states that the AI algorithm was trained on "large numbers of mammograms where cancer status is known." However, it does not provide a specific sample size for the training set.

    9. How the Ground Truth for the Training Set Was Established

    • The document implies the ground truth for the training set was established based on "cancer status is known" for the mammograms used for training. While not explicitly detailed, this would typically involve a combination of:
      • Pathology reports for confirmed cancers.
      • Long-term clinical follow-up for confirmed benign cases.
        It's also mentioned that the AI algorithm uses "deep neural networks that have been trained on large numbers of mammograms where cancer status is known," suggesting similar rigorous ground truth establishment as for the test set, but no specific methodology for the training set's ground truth is provided.
    Ask a Question

    Ask a specific question about this device

    Page 1 of 1