K Number
K231130
Device Name
TumorSight Viz
Manufacturer
Date Cleared
2023-12-26

(250 days)

Product Code
Regulation Number
892.2050
Panel
RA
Reference & Predicate Devices
AI/MLSaMDIVD (In Vitro Diagnostic)TherapeuticDiagnosticis PCCP AuthorizedThirdpartyExpeditedreview
Intended Use

TumorSight Viz is intended to be used in the visualization and analysis of breast magnetic resonance imaging (MRI) studies for patients with biopsy proven early-stage or locally advanced breast cancer. TumorSight Viz supports evaluation of dynamic MR data acquired from breast studies during contrast administration. TumorSight Viz performs processing functions (such as image registration, subtractions, measurements, 3D renderings, and reformats).

TumorSight Viz also includes user-configurable features for visualizing and analyzing findings in breast MRI studies. Patient management decisions should not be made based solely on the results of TumorSight Viz.

Device Description

TumorSight Viz is an image processing system designed to assist in the visualization and analysis of breast DCE-MRI studies.

TumorSight reads DICOM magnetic resonance images. TumorSight processes and displays the results on the TumorSight web application.

Available features support:

  • Visualization (standard image viewing tools, MIPs, and reformats)
  • Analysis (registration, subtractions, kinetic curves, parametric image maps, segmentation and 3D volume rendering)
  • Communication and storage (DICOM import, retrieval, and study storage)

The TumorSight system consists of proprietary software developed by SimBioSys, Inc. hosted on a cloud-based platform and accessed on an off-the-shelf computer.

AI/ML Overview

Here's a summary of the acceptance criteria and study details for TumorSight Viz, based on the provided text:

1. Acceptance Criteria and Reported Device Performance

The acceptance criteria are implied by demonstrating that the device's performance (Mean Absolute Error and Dice Coefficients) is comparable to inter-radiologist variability and the predicate device, CADstream, and that "all tests met the acceptance criteria".

Measurement DescriptionUnitsAcceptance Criteria (Implied)Validation Testing (Mean Abs. Error ± Std. Dev.)
Tumor Volume (n=157)cubic centimeters (cc)Comparable to inter-radiologist variability6.48 ± 12.67
Tumor-to-breast volume ratio (n=157)%Comparable to inter-radiologist variability0.56 ± 0.93
Tumor longest dimension (n=163)centimeters (cm)Comparable to inter-radiologist variability1.48 ± 1.46
Tumor-to-nipple distance (n=161)centimeters (cm)Comparable to inter-radiologist variability1.00 ± 1.03
Tumor-to-skin distance (n=163)centimeters (cm)Comparable to inter-radiologist variability0.63 ± 0.60
Tumor-to-chest distance (n=163)centimeters (cm)Comparable to inter-radiologist variability0.94 ± 1.34
Tumor center of mass (n=157)centimeters (cm)Comparable to inter-radiologist variability0.735 ± 1.26
Performance MeasurementMetricAcceptance Criteria (Implied)Validation Testing (Mean ± Std. Dev.)
Tumor segmentation (n=157)Volume DiceSufficient for indicating location, volume, surface agreement0.676 ± 0.289
Surface DiceSufficient for indicating location, volume, surface agreement0.873 ± 0.264

Additionally, for the direct comparison with the CADstream predicate device and ground truth:

Performance MeasurementMetricTumorSight Viz / Ground Truth (Mean Abs. Error ± Std. Dev.)CADStream / Ground Truth (Mean Abs. Error ± Std. Dev.)Inter-radiologist Variability (Mean Abs. Error ± Std. Dev.)
Longest Dimension (n=136)Abs. Distance Error1.40 cm ± 1.43 cm1.11 cm ± 1.52 cm1.17 cm ± 1.38 cm
Tumor to Skin (n=136)Abs. Distance Error0.61 cm ± 0.46 cm0.49 cm ± 0.56 cm0.49 cm ± 0.54 cm
Tumor to Chest (n=136)Abs. Distance Error0.77 cm ± 0.90 cm1.37 cm ± 1.01 cm0.79 cm ± 1.01 cm
Tumor to Nipple (n=134)Abs. Distance Error0.98 cm ± 1.06 cm0.80 cm ± 0.86 cm0.82 cm ± 0.98 cm
Tumor Volume (n=134)Abs. Distance Error6.69 cc ± 13.53 cc8.09 cc ± 17.42 ccN/A (not provided for inter-radiologist variability)

The document states: "The mean absolute error and variability between the automated measurements (Validation Testing) and ground truth for tumor volume (measured in cc) and landmark distances (measured in cm) was similar to the variability between device-to-radiologist measurements and inter-radiologist variability. This demonstrates that the error in measurements is consistent to the variability between expert readers." It also notes: "We found that all tests met the acceptance criteria, demonstrating adequate performance for our intended use." And for the comparison to the predicate: "The differences in error between the mean absolute errors (MAE) for the predicate and subject device are clinically acceptable because they are on the order of one to two voxels for the mean voxel size in the dataset. These differences are clinically insignificant."

2. Sample Size Used for the Test Set and Data Provenance

  • Sample Size for Validation (Test Set): 161 patients, corresponding to 163 samples (accounting for bilateral disease).
  • Data Provenance: Obtained from six (6) clinical sites in the U.S. All patients had pathologically confirmed invasive, early stage or locally advanced breast cancer. The data was collected to ensure adequate coverage of MRI manufacturer and field strength and similarity with the broader U.S. population for patient age, breast cancer subtype, T stage, histologic subtype, and race/ethnicity. This data is retrospective.

3. Number of Experts Used to Establish the Ground Truth for the Test Set and Qualifications of Those Experts

  • Number of Experts: Seven (7) U.S. Board Certified radiologists.
  • Qualifications of Experts: U.S. Board Certified radiologists. Specific experience level (e.g., years of experience) is not explicitly stated beyond "expert readers."

4. Adjudication Method for the Test Set

  • Adjudication Method: For each case, two radiologists independently measured various characteristics. If the two radiologists did not agree on whether the candidate segmentation was appropriate, a third radiologist provided an additional opinion and established a ground truth by majority consensus (2+1 adjudication).

5. If a Multi-Reader Multi-Case (MRMC) Comparative Effectiveness Study was Done, If so, what was the effect size of how much human readers improve with AI vs without AI assistance

  • The document describes a performance comparison between TumorSight Viz, CADstream (predicate), and ground truth, as well as inter-radiologist variability. However, it does not describe an MRMC comparative effectiveness study directly measuring how much human readers improve with AI vs. without AI assistance. The study focuses on the standalone performance of TumorSight Viz and its comparability to a predicate device and human variability.

6. If a Standalone (i.e. algorithm only without human-in-the-loop performance) was Done

  • Yes, a standalone study was done. The reported performance metrics (Mean Absolute Error, Dice Coefficients) for TumorSight Viz against a radiologist-established ground truth represent the standalone performance of the algorithm. The document explicitly states: "The measurements generated from the device result directly from the segmentation methodology and are an inferred reflection of the performance of the deep learning algorithm."

7. The type of ground truth used (expert consensus, pathology, outcomes data, etc.)

  • Type of Ground Truth: Pathologically confirmed breast cancer cases (for patient inclusion) combined with expert consensus of U.S. Board Certified radiologists for specific image measurements and segmentation appropriateness. The ground truth was established by two radiologists measuring characteristics, with a third radiologist adjudicating disagreements by majority consensus.

8. The sample size for the training set

  • Training Set Sample Size: 390 samples (from 736 patients mentioned for training and tuning).
  • Tuning Set Sample Size: 376 samples (from 736 patients mentioned for training and tuning).

9. How the ground truth for the training set was established

  • The document states that 736 patients (766 samples) were used for "training and tuning the device." It explicitly mentions that for the validation set, "Seven (7) U.S. Board Certified radiologists reviewed 163 validation samples to establish the ground truth for the dataset..."
  • The method for establishing ground truth for the training set is not explicitly detailed in the provided text. It is generally implied that such ground truth would also be established by experts, but the specifics are not given for the training portion.

§ 892.2050 Medical image management and processing system.

(a)
Identification. A medical image management and processing system is a device that provides one or more capabilities relating to the review and digital processing of medical images for the purposes of interpretation by a trained practitioner of disease detection, diagnosis, or patient management. The software components may provide advanced or complex image processing functions for image manipulation, enhancement, or quantification that are intended for use in the interpretation and analysis of medical images. Advanced image manipulation functions may include image segmentation, multimodality image registration, or 3D visualization. Complex quantitative functions may include semi-automated measurements or time-series measurements.(b)
Classification. Class II (special controls; voluntary standards—Digital Imaging and Communications in Medicine (DICOM) Std., Joint Photographic Experts Group (JPEG) Std., Society of Motion Picture and Television Engineers (SMPTE) Test Pattern).