Search Filters

Search Results

Found 1 results

510(k) Data Aggregation

    K Number
    K252496

    Validate with FDA (Live)

    Manufacturer
    Date Cleared
    2026-01-29

    (174 days)

    Product Code
    Regulation Number
    892.2050
    Age Range
    20 - 80
    Reference & Predicate Devices
    Predicate For
    N/A
    Why did this record match?
    Reference Devices :

    K221405

    AI/MLSaMDIVD (In Vitro Diagnostic)TherapeuticPediatricDiagnosticis PCCP AuthorizedThirdpartyExpeditedreview
    Intended Use

    Neurophet AQUA AD Plus is intended for automatic labeling, visualization, and volumetric quantification of segmentable brain structures and lesions, as well as SUVR quantification from a set of MR and PET images. Volumetric measurements may be compared to reference percentile data.

    Device Description

    Neurophet AQUA AD Plus is a software device intended for the automatic labeling of brain structures, visualization, and volumetric quantification of segmented brain regions and lesions, as well as standardized uptake value ratio (SUVR) quantification using MR and PET images. The volumetric outcomes are compared to normative reference data to support the evaluation of neurodegeneration and cognitive impairment.

    The device is designed to assist physicians in clinical evaluation by streamlining the clinical workflow from patient registration through image analysis, analysis result archiving, and report generation using software-based functionalities. The device provides percentile-based results by comparing an individual's imaging-derived quantitative analysis results to reference populations. Percentile-based results are provided for reference only and are not intended to serve as a standalone basis for diagnostic decision-making. Clinical interpretation must be performed by qualified healthcare professionals.

    AI/ML Overview

    Here's a breakdown of the acceptance criteria and study details for the Neurophet AQUA AD Plus, based on the provided FDA 510(k) Clearance Letter:


    Acceptance Criteria and Device Performance for Neurophet AQUA AD Plus

    The Neurophet AQUA AD Plus employs multiple AI modules for automated segmentation and quantitative analysis of brain structures and lesions using MR and PET images. The device's performance was validated against predefined acceptance criteria for each module.

    1. Table of Acceptance Criteria and Reported Device Performance

    AI ModulePerformance MetricAcceptance CriteriaReported Device Performance
    T1-SegEngine (T1-weighted structural MRI segmentation)Accuracy (Dice Similarity Coefficient - DSC)95% CI of DSC: [0.750, 0.850] for major cortical brain structures 95% CI of DSC: [0.800, 0.900] for major subcortical brain structuresCortical Regions: Mean DSC: 0.83 ± 0.04 (95% CI: 0.82–0.84) Subcortical Regions: Mean DSC: 0.87 ± 0.03 (95% CI: 0.86–0.88)
    Reproducibility (Average Volume Difference Percentage - AVDP)Equivalence range: 1.0–5.0% for both subcortical and cortical regionsSubcortical Regions: Mean AVDP: 2.50 ± 0.93% (95% CI: 2.26–2.74) Cortical Regions: Mean AVDP: 1.79 ± 0.74% (95% CI: 1.60–1.98)
    FLAIR-SegEngine (T2-FLAIR hyperintensity segmentation)Accuracy (Dice Similarity Coefficient - DSC)Mean DSC ≥ 0.80Mean DSC: 0.90 ± 0.04 (95% CI: 0.89–0.91)
    Reproducibility (Mean AVDP and Absolute Lesion Volume Difference)Absolute difference < 0.25 cc Mean AVDP < 2.5%Mean AVDP: 0.99 ± 0.66% Mean absolute lesion volume difference: 0.08 ± 0.06 cc
    PET-Engine (SUVR and Centiloid quantification)SUVR Accuracy (Intraclass Correlation Coefficient - ICC)ICC ≥ 0.60 across Alzheimer's-relevant regions (compared to FDA-cleared reference product K221405)ICC ≥ 0.993 across seven Alzheimer's-relevant regions
    Centiloid Classification (Kappa value for amyloid positivity)κ ≥ 0.70 (indicating substantial agreement with consensus expert visual reads)Kappa values met or exceeded criterion (specific values not provided, but noted as meeting/exceeding)
    ED-SegEngine (edema-like T2-FLAIR hyperintensity segmentation)Accuracy (Dice Similarity Coefficient - DSC)DSC ≥ 0.70Mean DSC: 0.91 ± 0.09 (95% CI: 0.89–0.93)
    HEM-SegEngine (GRE/SWI hypointense lesion segmentation)Accuracy (F1-score / DSC)F1-score ≥ 0.60Median F1-score (DSC): 0.860 (95% CI: 0.824–0.902)

    2. Sample Sizes and Data Provenance for the Test Set

    • T1-SegEngine (Accuracy): 60 independent T1-weighted MRI cases. Data provenance not explicitly stated, but implicitly from public repositories (e.g., ADNI, AIBL, PPMI) and institutional clinical sites as mentioned for training data, and distinct from training.
    • T1-SegEngine (Reproducibility): 60 subjects with paired T1-weighted scans (120 scans total). Data provenance not explicitly stated.
    • FLAIR-SegEngine (Accuracy): 136 independent T2-FLAIR cases. Data provenance not explicitly stated, but distinct from training data.
    • FLAIR-SegEngine (Reproducibility): Paired T2-FLAIR scans (number not specified). Data provenance not explicitly stated.
    • PET-Engine (SUVR accuracy): 30 paired MRI–PET datasets. Data provenance not explicitly stated, but implicitly from multi-center studies including varied tracers and sites.
    • PET-Engine (Centiloid classification): 176 paired T1-weighted MRI and amyloid PET scans from ADNI and AIBL. These are public repositories, likely involving diverse geographical data (e.g., USA, Australia). Data is retrospective.
    • ED-SegEngine (Accuracy): 100 T2-FLAIR scans collected from U.S. and U.K. clinical sites. Data is retrospective.
    • HEM-SegEngine (Accuracy): 106 GRE/SWI scans from U.S. clinical sites. Data is retrospective.

    For all modules, validation datasets were fully independent from training datasets at the subject level, drawn from distinct sites and/or repositories where applicable.
    The validation cohorts covered adult subjects across a broad age range (approximately 40–80+ years), with both females and males represented.
    Racial/ethnic composition included White, Asian, Black, and African American subjects, depending on the underlying public and institutional datasets.
    Clinical subgroups included clinically normal, mild cognitive impairment, and Alzheimer's disease for structural, FLAIR, and PET modules, and cerebrovascular/amyloid‑related pathologies for ED‑ and HEM‑SegEngines.

    3. Number of Experts and Qualifications for Ground Truth

    For structural and lesion segmentation modules (T1-, FLAIR-, ED-, HEM-SegEngines):

    • Number of Experts: Not explicitly stated as a specific number, but "subspecialty-trained neuroradiologists" were used.
    • Qualifications: "Subspecialty-trained neuroradiologists." Specific years of experience are not mentioned.

    For Centiloid classification in the PET-Engine:

    • Number of Experts: "Consensus expert visual reads." The exact number isn't specified, but implies multiple experts.
    • Qualifications: "Experts" trained in established amyloid PET reading criteria. Specific qualifications beyond "expert" and training in criteria are not detailed.

    4. Adjudication Method for the Test Set

    For structural and lesion segmentation modules (T1-, FLAIR-, ED-, HEM-SegEngines):

    • "Consensus/adjudication procedures and internal quality control to ensure consistency" were used for establishing reference segmentations. The specific 2+1, 3+1, or other detailed method is not provided.

    For Centiloid classification in the PET-Engine:

    • "Consensus expert visual interpretation" was used. The specific method details are not provided.

    5. Multi-Reader Multi-Case (MRMC) Comparative Effectiveness Study

    The provided text does not indicate that an MRMC comparative effectiveness study was done to compare human readers with AI assistance versus without AI assistance. The performance studies primarily focus on the standalone (algorithm-only) performance of the device against expert-derived ground truth or a cleared reference product.

    6. Standalone (Algorithm-Only) Performance Study

    Yes, a standalone (algorithm only without human-in-the-loop performance) study was done for all AI modules. The text explicitly states: "Standalone performance tests were conducted for each module using validation datasets that were completely independent from those used for model development and training." The results presented in the table above reflect this standalone performance.

    7. Type of Ground Truth Used

    • Expert Consensus:
      • For structural and lesion segmentation modules (T1-, FLAIR-, ED-, HEM-SegEngines), reference segmentations were generated by "subspecialty-trained neuroradiologists using predefined anatomical and lesion‑labeling criteria, with consensus/adjudication procedures."
      • For Centiloid classification in the PET-Engine, reference labels were derived from "consensus expert visual interpretation using established amyloid PET reading criteria."
    • Comparison to Cleared Reference Product:
      • For SUVR quantification in the PET-Engine, reference values were obtained from an "FDA‑cleared reference product (K221405)" (Neurophet SCALE PET).

    8. Sample Size for the Training Set

    The exact sample size for the training set is not explicitly stated as a single number. However, the document mentions:

    • "The AI-based modules (T1‑SegEngine, FLAIR‑SegEngine, PET‑Engine, ED‑SegEngine, HEM‑SegEngine) were trained using multi-center MRI and PET datasets collected from public repositories (e.g., ADNI, AIBL, PPMI) and institutional clinical sites."
    • "Training data covered:
      • Adult subjects across a broad age range (approximately 20–80+ years), with both sexes represented and including multiple racial/ethnic groups (e.g., White, Asian, Black).
      • A spectrum of clinical conditions relevant to the intended use, including clinically normal, mild cognitive impairment, and Alzheimer's disease, as well as patients with cerebrovascular and amyloid‑related pathologies for lesion-segmentation modules.
      • MRI acquired on major vendor platforms (GE, Siemens, Philips) at 1.5T and 3T... and amyloid PET acquired on multiple PET systems with commonly used tracers (Amyvid, Neuraceq, Vizamyl)."

    This indicates a large and diverse training set, although a precise count of subjects or images isn't provided.

    9. How the Ground Truth for the Training Set Was Established

    The document implies that the training data included "manual labels" as it states: "No images or manual labels from the training datasets were reused in the validation datasets." However, it does not explicitly detail the process by which these "manual labels" or ground truth for the training set were established (e.g., number of experts, qualifications, adjudication method for training data). It's reasonable to infer that similar expert-driven processes were likely used for training ground truth as for validation, but this is not explicitly confirmed in the provided text.

    Ask a Question

    Ask a specific question about this device

    Page 1 of 1