Search Results
Found 105 results
510(k) Data Aggregation
(128 days)
GE Medical Systems, LLC
Ask a specific question about this device
(126 days)
GE Medical Systems, LLC
The system is intended to produce cross-sectional images of the body by computer reconstruction of x-ray transmission projection data from the same axial plane taken at different angles. The system may acquire data using Axial, Cine, Helical, Cardiac, and Gated CT scan techniques from patients of all ages. These images may be obtained either with or without contrast. This device may include signal analysis and display equipment, patient and equipment supports, components and accessories.
This device may include data and image processing to produce images in a variety of trans-axial and reformatted planes. Further, the images can be post processed to produce additional imaging planes or analysis results.
The system is indicated for head, whole body, cardiac, and vascular X-ray Computed Tomography applications.
The device output is a valuable medical tool for the diagnosis of disease, trauma, or abnormality and for planning, guiding, and monitoring therapy.
If the spectral imaging option is included on the system, the system can acquire CT images using different kV levels of the same anatomical region of a patient in a single rotation from a single source. The differences in the energy dependence of the attenuation coefficient of the different materials provide information about the chemical composition of body materials. This approach enables images to be generated at energies selected from the available spectrum to visualize and analyze information about anatomical and pathological structures.
GSI provides information of the chemical composition of renal calculi by calculation and graphical display of the spectrum of effective atomic number. GSI Kidney stone characterization provides additional information to aid in the characterization of uric acid versus non-uric acid stones. It is intended to be used as an adjunct to current standard methods for evaluating stone etiology and composition.
The CT system is indicated for low dose CT for lung cancer screening. The screening must be performed within the established inclusion criteria of programs/ protocols that have been approved and published by either a governmental body or professional medical society.
This proposed device Revolution Vibe is a general purpose, premium multi-slice CT Scanning system consisting of a gantry, table, system cabinet, scanner desktop, power distribution unit, and associated accessories. It has been optimized for cardiac performance while still delivering exceptional imaging quality across the entire body.
Revolution Vibe is a modified dual energy CT system based on its predicate device Revolution Apex Elite (K213715). Compared to the predicate, the most notable change in Revolution Vibe is the modified detector design together with corresponding software changes which is optimized for cardiac imaging providing capability to image the whole heart in one single rotation same as the predicate.
Revolution Vibe offers an accessible whole heart coverage, full cardiac capability CT scanner which can deliver outstanding routine head and body imaging capabilities. The detector of Revolution Vibe uses the same GEHC's Gemstone scintillator with 256 x 0.625 mm row providing up to 16 cm of coverage in Z direction within 32 cm scan field of view, and 64 x 0.625 mm row providing up to 4 cm of coverage in Z direction within 50 cm scan field of view. The available gantry rotation speeds are 0.23, 0.28, 0.35, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, and 1.0 seconds per rotation.
Revolution Vibe inherits virtually all of the key technologies from the predicate such as: high tube current (mA) output, 80 cm bore size with Whisper Drive, Deep Learning Image Reconstruction for noise reduction (DLIR K183202/K213999, GSI DLIR K201745), ASIR-V iterative recon, enhanced Extended Field of View (EFOV) reconstruction MaxFOV 2 (K203617), fast rotation speed as fast as 0.23 second/rot (K213715), and spectral imaging capability enabled by ultrafast kilovoltage(kv) switching (K163213), as well as ECG-less cardiac (K233750). It also includes the Auto ROI enabled by AI which is integrated within the existing SmartPrep workflow for predicting Baseline and monitoring ROI automatically. As such, the Revolution Vibe carries over virtually all features and functionalities of the predicate device Revolution Apex Elite (K213715).
This CT system can be used for low dose lung cancer screening in high risk populations*.
The provided FDA 510(k) clearance letter and summary for the Revolution Vibe CT system does not include detailed acceptance criteria or a comprehensive study report to fully characterize the device's performance against specific metrics. The information focuses more on the equivalence to a predicate device and general safety/effectiveness.
However, based on the text, we can infer some aspects related to the Auto ROI feature, which is the only part of the device described with specific performance testing details.
Here's an attempt to extract and describe the available information, with clear indications of what is not provided in the document.
Acceptance Criteria and Device Performance for Auto ROI
The document mentions specific performance testing for the "Auto ROI" feature, which utilizes AI. For other aspects of the Revolution Vibe CT system, the submission relies on demonstrating substantial equivalence to the predicate device (Revolution Apex Elite) through engineering design V&V, bench testing, and a clinical reader study focused on overall image utility, rather than specific quantitative performance metrics meeting predefined acceptance criteria for the entire system.
1. Table of Acceptance Criteria and Reported Device Performance (Specific to Auto ROI)
Feature/Metric | Acceptance Criteria (Implicit) | Reported Device Performance |
---|---|---|
Auto ROI Success Rate | "exceeding the pre-established acceptance criteria" | Testing resulted in "success rates exceeding the pre-established acceptance criteria." (Specific numerical value not provided) |
Note: The document does not provide the explicit numerical value for the "pre-established acceptance criteria" or the actual "success rate" achieved for the Auto ROI feature.
2. Sample Size and Data Provenance for the Test Set (Specific to Auto ROI)
- Sample Size: 1341 clinical images
- Data Provenance: "real clinical practice" (Specific country of origin not mentioned). The images were used for "Auto ROI performance" testing, which implies retrospective analysis of existing clinical data.
3. Number of Experts and Qualifications to Establish Ground Truth (Specific to Auto ROI)
- Number of Experts: Not specified for the Auto ROI ground truth establishment.
- Qualifications of Experts: Not specified for the Auto ROI ground truth establishment.
Note: The document mentions 3 readers for the overall clinical reader study (see point 5), but this is for evaluating the diagnostic utility and image quality of the CT system and not explicitly for establishing ground truth for the Auto ROI feature.
4. Adjudication Method for the Test Set (Specific to Auto ROI)
- Adjudication Method: Not specified for the Auto ROI test set.
5. Multi-Reader Multi-Case (MRMC) Comparative Effectiveness Study
-
Was an MRMC study done? Yes, a "clinical reader study of sample clinical data" was carried out. It is described as a "blinded, retrospective clinical reader study."
-
Effect Size of Human Readers Improvement with AI vs. without AI assistance: The document states the purpose of this reader study was to validate that "Revolution Vibe are of diagnostic utility and is safe and effective for its intended use." It does not report an effect size or direct comparison of human readers' performance with and without AI assistance (specifically for the Auto ROI feature within the context of reader performance). The study seemed to evaluate the CT system's overall image quality and clinical utility, possibly implying that the Auto ROI is integrated into this overall evaluation, but a comparative effectiveness study of the AI's impact on human performance is not described.
- Details of MRMC Study:
- Number of Cases: 30 CT cardiac exams
- Number of Readers: 3
- Reader Qualifications: US board-certified in Radiology with more than 5 years' experience in CT cardiac imaging.
- Exams Covered: "wide range of cardiac clinical scenarios."
- Reader Task: "Readers were asked to provide evaluation of image quality and the clinical utility."
- Details of MRMC Study:
6. Standalone (Algorithm Only) Performance
- Was a standalone study done? Yes, for the "Auto ROI" feature, performance was tested "using 1341 clinical images from real clinical practice," and "the tests results in success rates exceeding the pre-established acceptance criteria." This implies an algorithm-only evaluation of the Auto ROI's ability to successfully identify and monitor ROI.
7. Type of Ground Truth Used (Specific to Auto ROI)
- Type of Ground Truth: Not explicitly stated for the Auto ROI. Given the "success rates" metric, it likely involved a comparison against a predefined "true" ROI determined by human experts or a gold standard method. It's plausible that this was established by expert consensus or reference standards.
8. Sample Size for the Training Set
- Sample Size: Not provided in the document.
9. How Ground Truth for the Training Set Was Established
- Ground Truth Establishment: Not provided in the document.
In summary, the provided documentation focuses on demonstrating substantial equivalence of the Revolution Vibe CT system to its predicate, Revolution Apex Elite, rather than providing detailed, quantitative performance metrics against specific acceptance criteria for all features. The "Auto ROI" feature is the only component where specific performance testing (standalone) is briefly mentioned, but key details like numerical acceptance criteria, actual success rates, and ground truth methodology for training datasets are not disclosed. The human reader study was for general validation of diagnostic utility, not a comparative effectiveness study of AI assistance.
Ask a specific question about this device
(190 days)
GE Medical Systems, LLC
Sonic DL is a Deep Learning based reconstruction technique that is available for use on GE HealthCare 1.5T, 3.0T, and 7.0T MR systems. Sonic DL reconstructs MR images from highly under-sampled data, and thereby enables highly accelerated acquisitions. Sonic DL is intended for imaging patients of all ages. Sonic DL is not limited by anatomy and can be used for 2D cardiac cine imaging and 3D Cartesian imaging using fast spin echo and gradient echo sequences. Depending on the region of interest, contrast agents may be used.
Sonic DL is a software feature intended for use with GE HealthCare MR systems. It includes a deep learning based reconstruction algorithm that enables highly accelerated acquisitions by reconstructing MR images from highly under-sampled data. Sonic DL is an optional feature that is integrated into the MR system software and activated through purchasable software option keys.
Here's a breakdown of the acceptance criteria and the study details for Sonic DL, based on the provided FDA 510(k) clearance letter:
Acceptance Criteria and Device Performance for Sonic DL
1. Table of Acceptance Criteria and Reported Device Performance
The document doesn't explicitly list specific quantitative "acceptance criteria" against which a single "reported device performance" is measured in a pass/fail manner for all aspects. Instead, it presents various performance metrics and qualitative assessments that demonstrate the device's acceptable performance compared to existing standards and its stated claims. For the sake of clarity, I've summarized the implied acceptance thresholds or comparative findings from the quantitative studies and the qualitative assessments.
Metric/Criterion | Acceptance Criteria (Implied/Comparative) | Reported Device Performance (Sonic DL) |
---|---|---|
Non-Clinical Testing (Sonic DL 3D) | ||
Peak-Signal-to-Noise (PSNR) | Equal to or above 30dB | Equal to or above 30dB at all acceleration factors (up to 12) |
Structural Similarity Index Measure (SSIM) | Equal to or above 0.8 | Equal to or above 0.8 at all acceleration factors (up to 12) |
Resolution | Preservation of resolution grid structure and resolution | Preserved resolution grid structure and resolution |
Medium/High Contrast Detectability | Retained compared to conventional methods | Retained at all accelerations; comparable or better than conventional methods |
Low Contrast Detectability | Non-inferior to more modestly accelerated conventional reconstruction methods at recommended acceleration rates | Maintained at lower acceleration factors; non-inferior at recommended rates (e.g., 8x Sonic DL 3D ~ 4x parallel imaging; 12x Sonic DL 3D ~ 8x parallel imaging) |
Model Stability (Hallucination) | Low risk of hallucination; dataset integrity preserved | Low risk of hallucination; dataset integrity preserved across all cases |
Clinical Testing (Sonic DL 3D) - Quantitative Post Processing | ||
Volumetric Measurements (Brain Tissues) - Relative MAE 95% CI | Less than 5% for most regions (brain tissues) | Less than 5% for most regions |
Volumetric Measurements (HOS) - Relative MAE 95% CI | Less than 3% for Hippocampal Occupancy Score (HOS) | Less than 3% for HOS |
Intra-class Correlation Coefficient (ICC) | Exceeded 0.75 across all comparisons | Exceeded 0.75 across all comparisons |
Clinical Testing (Sonic DL 3D) - Clinical Evaluation Studies (Likert-score) | ||
Diagnostic Quality | Images are of diagnostic quality | Sonic DL 3D images are of diagnostic quality (across all anatomies, field strengths, and acceleration factors investigated) |
Pathology Retention | Pathology seen in comparator images can be accurately retained | Pathology seen in ARC + HyperSense images can be accurately retained |
Decline with Acceleration | Retain diagnostic quality overall despite decline | Scores gradually declined with increasing acceleration factors yet retained diagnostic quality overall |
Clinical Claims | ||
Scan Time Reduction | Substantial reduction in scan time | Yields substantial reduction in scan time |
Diagnostic Image Quality | Preservation of diagnostic image quality | Preserves diagnostic image quality |
Acceleration Factors | Up to 12x | Provides acceleration factors up to 12 |
2. Sample Size Used for the Test Set and Data Provenance
- Quantitative Post Processing Test Set:
- Sample Size: 15 fully-sampled datasets.
- Data Provenance: Retrospective, acquired at GE HealthCare in Waukesha, USA, from 1.5T, 3.0T, and 7.0T scanners.
- Clinical Evaluation Studies (Likert-score based):
- Study 1 (Brain, Spine, Extremities):
- Number of image series evaluated: 120 de-identified cases.
- Number of unique subjects: 54 subjects (48 patients, 6 healthy volunteers).
- Age range: 11-80 years.
- Gender: 26 Male, 28 Female.
- Pathology: Mixture of small, large, focal, diffuse, hyper- and hypo-intense lesions.
- Contrast: Used in a subset as clinically indicated.
- Data Provenance: Retrospective and prospective (implied by "obtained from clinical sites and from healthy volunteers scanned at GE HealthCare facilities"). Data collected from 7 sites (4 in United States, 3 outside of United States).
- Study 2 (Brain):
- Number of additional cases: 120 cases.
- Number of unique subjects: From 30 fully-sampled acquisitions.
- Data Provenance: Retrospective, collected internally at GE HealthCare, 1.5T, 3.0T, and 7.0T field strengths.
- Study 1 (Brain, Spine, Extremities):
3. Number of Experts Used to Establish the Ground Truth for the Test Set and Their Qualifications
- Quantitative Post Processing: The "ground truth" here is the fully sampled data and the quantitative measurements derived from it. No human experts are explicitly mentioned for establishing this computational 'ground truth'.
- Clinical Evaluation Studies:
- Study 1: 3 radiologists. Their specific qualifications (e.g., years of experience, subspecialty) are not provided in the document.
- Study 2: 3 radiologists. Their specific qualifications are not provided in the document.
4. Adjudication Method for the Test Set
The document does not explicitly state a formal adjudication method (like 2+1 or 3+1). For the clinical evaluation studies, it mentions that "three radiologists were asked to evaluate the diagnostic quality of images" and "radiologists were also asked to comment on the presence of any pathology." This suggests individual assessments were either aggregated, or findings were considered concordant if a majority agreed, but a specific arbitration or adjudication process for disagreements is not detailed.
5. Multi Reader Multi Case (MRMC) Comparative Effectiveness Study
- Was an MRMC study done? Yes, the two Likert-score based clinical studies involved multiple readers (3 radiologists) evaluating multiple cases (120 de-identified cases in Study 1 and 120 additional cases in Study 2) for comparative effectiveness against ARC + HyperSense.
- Effect size of human readers improvement with AI vs. without AI assistance: The document states that "Sonic DL 3D images are of diagnostic quality while yielding a substantial reduction in the scan time compared to ARC + HyperSense images." It also noted that "pathology seen in the ARC + HyperSense images can be accurately retained in Sonic DL 3D images." However, it does not quantify the effect size of how much human readers improve with AI assistance (Sonic DL) versus without it. Instead, the studies aim to demonstrate non-inferiority or comparable diagnostic quality despite acceleration. There's no performance gain stated explicitly for the human reader in terms of diagnostic accuracy or confidence when using Sonic DL images compared to conventional images; rather, the benefit is in maintaining diagnostic quality with faster acquisition.
6. Standalone (Algorithm Only) Performance Study
- Was a standalone study done? Yes, extensive non-clinical testing was performed as a standalone assessment of the algorithmic performance. This included evaluations using:
- Digital reference objects (DROs) and MR scans of physical ACR phantoms to measure PSNR, RMSE, SSIM, resolution, and low contrast detectability.
- A task-based study using a convolutional neural network ideal observer (CNN-IO) to quantify low contrast detectability.
- Reconstruction of in vivo datasets with unseen data inserted to assess model stability and hallucination risk.
These studies directly evaluated the algorithm's output metrics and behavior independently of human interpretation in a clinical workflow, making them standalone performance assessments.
7. Type of Ground Truth Used
- Non-Clinical Testing:
- Quantitative Metrics (PSNR, RMSE, SSIM, Resolution, Contrast Detectability): Fully sampled data was used as the reference "ground truth" against which the under-sampled and reconstructed Sonic DL 3D images were compared.
- Model Stability (Hallucination): The "ground truth" was the original in vivo datasets before inserting previously unseen data, allowing for evaluation of whether the algorithm introduced artifacts or hallucinations.
- Quantitative Post Processing (Clinical Testing):
- Fully sampled data sets were used as the reference for comparison of volumetric measurements with Sonic DL 3D and ARC + HyperSense images.
- Clinical Evaluation Studies (Likert-score based):
- The implied "ground truth" was the diagnostic quality and presence/absence of pathology as assessed by the conventional ARC + HyperSense images, which were considered the clinical standard for comparison. The radiologists were essentially comparing Sonic DL images against the standard of care images without a separate, absolute ground truth like pathology for every lesion.
8. Sample Size for the Training Set
The document does not specify the sample size used for training the Sonic DL 3D deep learning model. It only mentions that Sonic DL is a "Deep Learning based reconstruction technique" and includes a "deep learning convolutional neural network."
9. How the Ground Truth for the Training Set Was Established
The document does not describe how the ground truth for the training set was established. It is standard practice for supervised deep learning models like Sonic DL to be trained on pairs of under-sampled and corresponding fully-sampled or high-quality (e.g., conventionally reconstructed) images, where the high-quality image serves as the 'ground truth' for the network to learn to reconstruct from the under-sampled data. However, the specifics of this process (e.g., data types, annotation, expert involvement) are not mentioned in this document.
Ask a specific question about this device
(34 days)
GE Medical Systems, LLC
The SIGNA Prime Elite is a whole body magnetic resonance scanner designed to support high signal-to-noise ratio, and short scan times. It is indicated for use as a diagnostic imaging device to produce axial, sagittal, coronal, and oblique images, spectroscopic images, parametric maps, and/or spectra, dynamic images of the structures and/or functions of the entire body, including, but not limited to, head, neck, heart, abdomen, pelvis, joints, prostate, blood vessels, and musculoskeletal regions of the body.
Depending on the region of interest being imaged, contrast agents may be used. The images produced by SIGNA Prime Elite reflect the spatial distribution or molecular environment of nuclei exhibiting magnetic resonance.
These images and/or spectra when interpreted by a trained physician yield information that may assist in diagnosis.
SIGNA Prime Elite is a whole body magnetic resonance scanner designed to support high resolution, high signal-to-noise ratio, and short scan time. The system uses a combination of time-varying magnet fields (Gradients) and RF transmissions to obtain information regarding the density and position of elements exhibiting magnetic resonance. The system can image in the sagittal, coronal, axial, oblique, and double oblique planes, using various pulse sequences, imaging techniques and reconstruction algorithms. The system features a 1.5T superconducting magnet with 60 cm bore size. The system is designed to conform to NEMA DICOM standards (Digital Imaging and Communications in Medicine).
The document does not provide a table of acceptance criteria and reported device performance. It focuses on demonstrating substantial equivalence to a predicate device rather than presenting specific quantitative performance metrics against pre-defined acceptance criteria.
Here's an analysis of the provided information concerning acceptance criteria and study details:
1. A table of acceptance criteria and the reported device performance
The document does not contain a specific table outlining quantitative acceptance criteria and corresponding reported device performance metrics. It indicates that the SIGNA Prime Elite's image quality performance was compared to the predicate device through bench and clinical testing and found to be "substantially equivalent."
2. Sample size used for the test set and the data provenance (e.g. country of origin of the data, retrospective or prospective)
The document states: "The subject of this premarket submission, the SIGNA Prime Elite, did not require clinical studies to support substantial equivalence. Sample clinical images have been included in this submission."
Therefore, there is no formal test set sample size mentioned for a specific clinical performance study. The "sample clinical images" are used to demonstrate acceptable diagnostic image performance, but details about their sample size, provenance (country of origin), or whether they are retrospective or prospective are not provided.
3. Number of experts used to establish the ground truth for the test set and the qualifications of those experts (e.g. radiologist with 10 years of experience)
Since no formal clinical study with a defined test set and ground truth establishment is described, this information is not provided. The document mentions that images "when interpreted by a trained physician yield information that may assist in diagnosis," but it doesn't specify expert review for a test set.
4. Adjudication method (e.g. 2+1, 3+1, none) for the test set
As there is no formal test set described for a clinical performance study, an adjudication method is not applicable and not mentioned.
5. If a multi reader multi case (MRMC) comparative effectiveness study was done, If so, what was the effect size of how much human readers improve with AI vs without AI assistance
The document states that the SIGNA Prime Elite is a "whole body magnetic resonance scanner." There is no mention of AI assistance or a multi-reader multi-case (MRMC) comparative effectiveness study comparing human readers with and without AI assistance. The submission is for an imaging device, not an AI-powered diagnostic tool.
6. If a standalone (i.e. algorithm only without human-in-the-loop performance) was done
This question is not applicable as the device is a magnetic resonance scanner, not an algorithm that operates standalone without human interpretation.
7. The type of ground truth used (expert consensus, pathology, outcomes data, etc.)
Given that clinical studies were not required and only "sample clinical images" were included to demonstrate acceptable diagnostic image performance, there is no mention of a formally established ground truth type (e.g., expert consensus, pathology, outcomes data) for a test set. The images are expected to be interpreted by a "trained physician."
8. The sample size for the training set
The document describes the SIGNA Prime Elite as a new 1.5T MR system. It is a hardware device with associated software, not a machine learning model. Therefore, the concept of a "training set" in the context of machine learning is not applicable and not mentioned.
9. How the ground truth for the training set was established
As the device is an MR scanner and not an AI/ML model requiring a training set, this information is not applicable and not provided.
Ask a specific question about this device
(167 days)
GE Medical Systems, LLC
SIGNA MAGNUS system is a head-only magnetic resonance scanner designed to support high resolution, high signal-tonoise ratio, diffusion-weighted imaging, and short scan times. SIGNA MAGNUS is indicated for use as a diagnostic imaging device to produce axial, sagittal, coronal, and oblique images, parametric maps, and/or spectra, dynamic images of the structures and/or functions of the head, neck, TMJ, and limited cervical spine on patients 6 years of age and older. Depending on the region of interest being imaged, contrast agents may be used.
The images produced by SIGNA MAGNUS reflect the spatial distribution or molecular environment of nuclei exhibiting magnetic resonance. These images and/or spectra, when interpreted by a trained physician, yield information that may assist in diagnosis.
SIGNA MAGNUS is a 3.0T high-performance magnetic resonance imaging system designed to support imaging of the head, neck, TMJ and limited cervical spine. The system supports scanning in axial, coronal, sagittal, oblique, and double oblique planes using a variety of pulse sequences, imaging techniques, acceleration methods, and reconstruction algorithms. The system can be delivered as a new system installation, or as an upgrade to existing compatible whole-body 3.0T MR
Key aspects of the system design:
• An asymmetrically designed, head-only gradient coil that achieves up to 300 mT/m peak gradient amplitude and 750 T/m/s peak slew rate.
• A graduated patient bore size, starting at 74 cm at the entry down to 37 cm at isocenter.
• Uses the same magnet as a conventional whole-body 3.0T system, with integral active shielding and a zero boil-off cryostat.
• Can be installed as a new system or upgraded from an existing compatible whole-body 3.0T MR system.
• A dockable mobile patient table.
• Oscillating Diffusion Encoding (ODEN) - a spectral diffusion technique that uses a sinusoidal diffusion gradient waveform.
The provided text is a 510(k) Summary for the GE Healthcare SIGNA MAGNUS, a magnetic resonance diagnostic device. The summary indicates that this device did not require clinical studies to support substantial equivalence. Therefore, there are no acceptance criteria, performance metrics, sample sizes for test/training sets, expert qualifications, or other details related to clinical performance studies to report for this device, as these were not performed or deemed necessary by the manufacturer for this submission.
The document states:
"The subject of this premarket submission, SIGNA MAGNUS did not require clinical studies to support substantial equivalence. Sample clinical images have been included in this submission. The sample clinical images demonstrate acceptable diagnostic image performance of SIGNA MAGNUS in accordance with the FDA Guidance 'Submission of Premarket Notifications fo October 10, 2023. The image quality of SIGNA MAGNUS is substantially equivalent to that of the predicate device."
This means that the substantial equivalence was primarily based on non-clinical tests and a comparison of technological characteristics and indications for use with a predicate device (SIGNA Premier (K193282)). The non-clinical tests focused on safety and performance in compliance with various voluntary standards (e.g., ANSI AAMI ES60601-1, IEC 60601-1-2, IEC 60601-2-33, IEC 62304, IEC 60601-1-6, IEC 62366-1, IEC62464-1, ISO 10993-1, NEMA MS, NEMA PS3 DICOM).
Due to the absence of a clinical study, the requested information cannot be extracted from the provided text.
Ask a specific question about this device
(132 days)
GE Medical Systems, LLC.
ECG-less Cardiac streamlines patient preparation by enabling an alternative acquisition of cardiac CT images for general cardiac assessment without the need of a patient-attached ECG monitor. ECG-less Cardiac is for adults only.
ECG-less Cardiac is a software device that is an additional, optional cardiac scan mode that can be used on the Revolution Apex Elite, Revolution Apex, and Revolution CT with Apex edition systems. There is no change to the predicate device hardware to support the subject device. Currently, the available cardiac scan modes on the Revolution CT Family are Cardiac Axial and Cardiac Helical, which makes use of an ECG signal to physiologically trigger the cardiac acquisitions and/or to retrospectively gate the reconstruction.
ECG-less Cardiac is a third cardiac scan mode that introduces the ability to acquire cardiac images without the need of a patient-attached ECG monitor. Hence, an ECG signal from the patient is not utilized for this scan mode. The ECG-less Cardiac workflow leverages the full-heart coverage capability of 160 mm configurations, fast gantry speeds (0.28 and 0.23 s/rot), and existing cardiac software options of SmartPhase and SnapShot Freeze 2 (K183161) to acquire images that are suitable for coronary and cardiac functional assessment.
The ECG-less cardiac feature allows the user to acquire a cardiac CT scan without the need to complete the steps associated with utilizing an ECG monitor, such as attaching ECG electrodes to the patient, checking electrode impedance, and confirming an ECG trace is displayed on the operator console, thus optimizing the workflow.
ECG-less Cardiac may be best utilized in examinations where excluding the ECG connection would streamline the patient examination, including and unloading of the patient. This may result in an improved workflow for certain clinical presentations. ECG-less Cardiac may also increase access to cardiac assessment for patients that are difficult to receive an ECG signal from. Circumstances where the subject device is expected to increase cardiac access includes scenarios where trauma patient has a diagnostic ECG attached and/or other instrumentation, such that there is added difficulty of attaching ECG leads for a gated scan, and situations where it is challenging to get an ECG signal from a patient such as a patient's t-wave triggering the scan or R-peak being difficult to detect.
Here's a summary of the acceptance criteria and the study proving the device meets them, based on the provided text:
1. Table of Acceptance Criteria and Reported Device Performance
The acceptance criteria are implicitly derived from the study's conclusions, focusing on diagnostic utility and image quality. No specific quantitative thresholds for acceptance are explicitly stated in the document beyond "interpretable without a significant motion artifact penalty" and "of diagnostic utility."
Acceptance Criteria (Inferred) | Reported Device Performance |
---|---|
Diagnostic Utility | ECG-less Cardiac acquisitions were consistently rated as interpretable and of diagnostic utility by board-certified radiologists who specialize in cardiac imaging. |
Image Quality (Motion Artifact) | Images generated from ECG-less Cardiac acquisitions were consistently rated as interpretable without a significant motion artifact penalty. |
Equivalence to ECG-gated "ground truth" | Engineering bench testing showed that ECG-less Cardiac scan acquisitions can produce images that are equivalent to an ECG-gated "ground truth" nominal phase location. |
Safety & Effectiveness | The device is deemed safe and effective for its intended use based on non-clinical testing and the clinical reader study. |
2. Sample Size Used for the Test Set and the Data Provenance
- Sample Size for Test Set: The document does not explicitly state the exact number of cases or images included in the reader study (test set). It refers to "a reader study of sample clinical data" and "prospectively collected clinical data."
- Data Provenance: The data was prospectively collected clinical data from patients undergoing a routine cardiac exam. The country of origin is not specified.
3. Number of Experts Used to Establish the Ground Truth for the Test Set and Their Qualifications
- Number of Experts: Three experts were used.
- Qualifications of Experts: They were board-certified radiologists who specialize in cardiac imaging. The document does not specify their years of experience.
4. Adjudication Method for the Test Set
The adjudication method is not explicitly stated. The document mentions that each image was "read by three board certified radiologists who specialize in cardiac imaging who provided an assessment of image quality." This suggests independent readings, but it does not detail a consensus or adjudication process (e.g., 2+1, 3+1).
5. If a Multi-Reader Multi-Case (MRMC) Comparative Effectiveness Study Was Done, and the Effect Size of How Much Human Readers Improve with AI vs. Without AI Assistance
- No, an MRMC comparative effectiveness study involving human readers improving with AI vs. without AI assistance was not conducted or reported.
- The study was a reader study where experts assessed images generated by the ECG-less Cardiac system. The primary goal was to validate the diagnostic utility and image quality of the ECG-less Cardiac acquisitions themselves, not to assess human reader performance with or without an AI assist feature.
6. If a Standalone (i.e., algorithm only without human-in-the-loop performance) Was Done
Yes, a form of standalone performance was assessed in the engineering bench testing. This testing "assessed how simulated ECG-less Cardiac scan conditions performed against an ECG-gated 'ground truth' nominal phase location." This component evaluated the algorithm's ability to generate images comparable to traditional ECG-gated acquisitions without human interpretation being the primary focus.
7. The Type of Ground Truth Used
- For the engineering bench testing, the ground truth was an ECG-gated "ground truth" nominal phase location. This implies a comparison to a known, established reference standard for cardiac imaging synchronization.
- For the clinical reader study, the ground truth was effectively the expert consensus/assessment of the three board-certified radiologists regarding the interpretability, motion artifact, and diagnostic utility of the ECG-less images. There is no mention of pathology or outcomes data being used as ground truth for this part of the study.
8. The Sample Size for the Training Set
The document does not provide any information regarding the sample size used for the training set of the ECG-less Cardiac software.
9. How the Ground Truth for the Training Set Was Established
The document does not provide any information on how the ground truth for the training set was established. It only discusses the testing (validation) phase.
Ask a specific question about this device
(188 days)
GE Medical Systems,LLC (GE Healthcare)
Sonic DL is a Deep Learning based image reconstruction technique that is available for use on GE Healthcare 1.5T and 3.0T MR systems. Sonic DL reconstructs MR images from highly under-sampled data, and thereby enables highly accelerated acquisitions. Sonic DL is intended for cardiac imaging, and for patients of all ages.
Sonic DL is a new software feature intended for use with GE Healthcare MR systems. It consists of a deep learning based reconstruction algorithm that is applied to data from MR cardiac cine exams obtained using a highly accelerated acquisition technique.
Sonic DL is an optional feature that is integrated into the MR system software and activated through a purchasable software option key.
Here's a breakdown of the acceptance criteria and the study details for the Sonic DL device, based on the provided document:
Sonic DL Acceptance Criteria and Study Details
1. Table of Acceptance Criteria and Reported Device Performance
The document describes the performance of Sonic DL in comparison to conventional ASSET Cine images. While explicit numerical acceptance criteria for regulatory clearance are not stated, the studies aim to demonstrate non-inferiority or superiority in certain aspects. The implicit acceptance criteria are:
- Diagnostic Quality: Sonic DL images must be rated as being of diagnostic quality.
- Functional Measurement Agreement: Functional cardiac measurements (e.g., LV volumes, EF, CO) from Sonic DL images must agree closely with those from conventional ASSET Cine images, ideally within typical inter-reader variability.
- Reduced Scan Time: Sonic DL must provide significantly shorter scan times.
- Preserved Image Quality: Image quality must be preserved despite higher acceleration.
- Single Heartbeat Imaging (Functional): Enable functional imaging in a single heartbeat.
- Rapid Free-Breathing Functional Imaging: Enable rapid functional imaging without breath-holds.
Implicit Acceptance Criterion | Reported Device Performance |
---|---|
Diagnostic Quality | "on average the Sonic DL images were rated as being of diagnostic quality" (second reader study). |
Functional Measurement Agreement | "the inter-method variability (coefficient of variability comparing functional measurements taken with Sonic DL images versus measurements using the conventional ASSET Cine images) was smaller than the inter-observer intra-method variability for the conventional ASSET Cine images for all parameters, indicating that Sonic DL is suitable for performing functional cardiac measurements" (first reader study). |
"Functional measurements using Sonic DL 1 R-R free breathing images from 10 subjects were compared to functional measurements using the conventional ASSET Cine breath hold images, and showed close agreement" (additional clinical testing for 1 R-R free breathing). | |
Reduced Scan Time | "providing a significant reduction in scan time compared to the conventional ASSET Cine images" (second reader study). |
"the Sonic DL feature provided significantly shorter scan times than the conventional Cine imaging" (overall conclusion). | |
Preserved Image Quality | "capable of reconstructing Cine images from highly under sampled data that are similar to the fully sampled Cine images in terms of image quality and temporal sharpness" (nonclinical testing). |
"the image quality of 13 Sonic DL 1 R-R free breathing cases was evaluated by a U.S. board certified radiologist, and scored higher than the corresponding conventional free breathing Cine images from the same subjects" (additional clinical testing for 1 R-R free breathing). | |
Single Heartbeat Functional Imaging | "Sonic DL is capable of achieving a 12 times acceleration factor and obtaining free-breathing images in a single heartbeat (1 R-R)" (additional clinical testing). |
Rapid Free-Breathing Functional Imaging | "Sonic DL is capable of... obtaining free-breathing images in a single heartbeat (1 R-R)" (additional clinical testing). |
2. Sample Size Used for the Test Set and Data Provenance
The document describes two primary reader evaluation studies and additional clinical testing.
- First Reader Study (Functional Measurements):
- Sample Size: 107 image series from 57 unique subjects (46 patients, 11 healthy volunteers).
- Data Provenance: Data from 7 sites: 2 GE Healthcare facilities and 5 external clinical collaborators. This indicates data from multiple sources, likely a mix of prospective and retrospective collection. The geographic origin of these sites is not explicitly stated but implies a multi-center study potentially from different countries where GE Healthcare operates or collaborates.
- Second Reader Study (Image Quality Assessment):
- Sample Size: 127 image sets, which included a subset of the subjects from the first study.
- Data Provenance: Same as the first reader study (clinical sites and healthy volunteers at GE Healthcare facilities).
- Additional Clinical Testing (1 R-R Free Breathing):
- Functional Measurements: 10 subjects.
- Image Quality Evaluation: 13 subjects.
- Data Provenance: In vivo cardiac cine images from 19 healthy volunteers. This implies prospective collection or a subset of prospectively collected healthy volunteer data.
3. Number of Experts Used to Establish the Ground Truth for the Test Set and Qualifications
- First Reader Study (Functional Measurements): Three radiologists. Qualifications are not explicitly stated, but their role in making quantitative measurements implies expertise in cardiac MRI.
- Second Reader Study (Image Quality Assessment): Three radiologists. Qualifications are not explicitly stated, but their role in blinded image quality assessments implies expertise in cardiac MRI interpretation.
- Additional Clinical Testing (1 R-R Free Breathing Image Quality): One U.S. board certified radiologist.
4. Adjudication Method for the Test Set
The document does not explicitly state an adjudication method (like 2+1, 3+1, or none) for either the functional measurements or the image quality assessments. For the first study, it mentions "inter-method variability" and "inter-observer intra-method variability," suggesting that the readings from the three radiologists were compared against each other and against the conventional method, but not necessarily adjudicated to establish a single "ground truth" per case. For the second study, "blinded image quality assessments" were performed, and ratings were averaged, but no adjudication process is described.
5. If a Multi-Reader Multi-Case (MRMC) Comparative Effectiveness Study was done, and the effect size
A clear MRMC comparative effectiveness study, in the sense of measuring human reader improvement with AI vs. without AI assistance, is not explicitly described.
The studies compare the performance of Sonic DL images (algorithm output) against conventional images, with human readers evaluating both.
- The first reader study compares quantitative measurements from Sonic DL images to conventional images, indicating suitability for performing functional cardiac measurements by showing smaller inter-method variability than inter-observer intra-method variability for conventional images. This suggests Sonic DL is at least as reliable as the variability between conventional human measurements.
- The second reader study involves blinded image quality assessments of both conventional and Sonic DL images, confirming that Sonic DL images were rated as diagnostic quality.
- The additional clinical testing for 1 R-R free breathing shows that Sonic DL images were "scored higher than the corresponding conventional free breathing Cine images" by a U.S. board-certified radiologist.
These are comparisons of the image quality and output from the AI system versus conventional imaging, interpreted by readers, rather than measuring human reader performance assisted by the AI system.
Therefore, the effect size of how much human readers improve with AI vs. without AI assistance is not provided because the studies were designed to evaluate the image output quality and measurement agreement of the AI-reconstructed images themselves, not to assess an AI-assisted workflow for human readers.
6. If a Standalone (i.e., algorithm only without human-in-the-loop performance) was done
Yes, standalone performance was assessed for image quality metrics.
- Nonclinical Testing: "Model accuracy metrics such as Peak-Signal-to-Noise (PSNR), Root-Mean-Square Error (RMSE), Structural Similarity Index Measure (SSIM), and Mean Absolute Error (MAE) were used to compare simulated Sonic DL images with different levels of acceleration and numbers of phases to the fully sampled images." This is a standalone evaluation of the algorithm's output quality against a reference.
- In Vivo Testing: "model accuracy and temporal sharpness evaluations were conducted using in vivo cardiac cine images obtained from 19 health volunteers." This is also a standalone technical evaluation of the algorithm's output on real data.
7. The Type of Ground Truth Used
- Nonclinical Testing (Simulated Data): The ground truth was the "fully sampled images" generated from an MRXCAT phantom and a digital phantom.
- Clinical Testing (Reader Studies):
- Functional Measurements: The "ground truth" for comparison was the measurements taken from the "conventional ASSET Cine images." The variability of these conventional measurements across readers also served as a baseline for comparison. This is a form of clinical surrogate ground truth (comparing to an established accepted method).
- Image Quality Assessments: The "ground truth" was the expert consensus/opinion of the radiologists during their blinded assessments of diagnostic quality.
- Additional Clinical Testing (1 R-R Free Breathing): Functional measurements were compared to "conventional ASSET Cine breath hold images" (clinical surrogate ground truth). Image quality was based on the scoring by a "U.S. board certified radiologist" (expert opinion).
No pathology or outcomes data were used as ground truth. The ground truth in the clinical setting was primarily based on established imaging techniques (conventional MR) and expert radiologist assessments.
8. The Sample Size for the Training Set
The document does not explicitly state the sample size for the training set used for the deep learning model. It only describes the data used for testing the device.
9. How the Ground Truth for the Training Set Was Established
Since the training set size is not provided, the method for establishing its ground truth is also not described in the provided text. Typically, for deep learning reconstruction, the "ground truth" for training often involves fully sampled or high-quality reference images corresponding to the undersampled input data.
Ask a specific question about this device
(185 days)
GE Medical Systems, LLC
Critical Care Suite with Pneumothorax Detection AI Algorithm is a computer-aided triage, notification, and diagnostic device that analyzes frontal chest X-ray images for the presence of a pneumothorax. Critical Care Suite identifies and highlights images with a pneumothorax to enable case prioritization or triage and assist as a concurrent reading aide during interpretation of radiographs.
Intended users include qualified independently licensed healthcare professionals (HCPs) trained to independently assess the presence of pneumothoraxes in radiographic images and radiologists.
Critical Care Suite should not be used in-lieu of full patient evaluation or solely relied upon to make or confirm a diagnosis. It is not intended to replace the review of the X-ray image by a qualified physician. Critical Care Suite is indicated for adults and Transitional Adolescents (18 to
Critical Care Suite is a suite of Al algorithms for the automated image analysis of frontal chest X-rays acquired on a digital x-ray system for the presence of critical findings. Critical Care Suite with Pneumothorax Detection Al Algorithm is indicated for adults and transitional adolescents (18 to
Here's a summary of the acceptance criteria and study details for the GE Medical Systems, LLC Critical Care Suite with Pneumothorax Detection AI Algorithm, based on the provided document:
1. Table of Acceptance Criteria and Reported Device Performance
The document primarily focuses on reporting the device's performance against its own established criteria rather than explicitly listing pre-defined "acceptance criteria" tables. However, we can infer the acceptance criteria from the reported performance goals.
Metric | Acceptance Criteria (Implied from Performance) | Reported Device Performance (Standalone) | Reported Device Performance (MRMC with AI Assistance vs. Non-Aided) |
---|---|---|---|
Pneumothorax Detection (Standalone Algorithm) | Detect pneumothorax in frontal chest X-ray images, with high diagnostic accuracy. | AUC of 96.1% (94.9%, 97.2%) | Not Applicable |
Sensitivity (Overall) | High sensitivity for overall pneumothorax detection. | 84.3% (80.6%, 88.0%) | Not Applicable |
Specificity (Overall) | High specificity for overall pneumothorax detection. | 93.2% (90.8%, 95.6%) | Not Applicable |
Sensitivity (Large Pneumothorax) | High sensitivity for large pneumothoraxes. | 96.3% (93.1%, 99.2%) | Not Applicable |
Sensitivity (Small Pneumothorax) | High sensitivity for small pneumothoraxes. | 75.0% (69.2%, 80.8%) | Not Applicable |
Pneumothorax Localization (Standalone Algorithm) | Localize suspected pneumothoraxes effectively. | Partially localized 98.1% (96.6%, 99.6%) of actual pneumothorax within an image (apical, lateral, inferior regions). | Not Applicable |
Full agreement between regions. | 67.8% (62.7%, 73.0%) | Not Applicable | |
Overlap with true pneumothorax area. | DICE Similarity Coefficient of 0.705 (0.683, 0.724) | Not Applicable | |
Reader Performance Improvement (MRMC Study) | Improve reader performance for pneumothorax detection. | Mean AUC improved by 14.5% (7.0%, 22.0%; p=.002) from 76.8% (non-aided) to 91.3% (aided). | 14.5% improvement in mean AUC |
Reader Sensitivity Improvement | Increase reader sensitivity. | Reader sensitivity increased by 16.3% (13.1%, 19.5%; p |
Ask a specific question about this device
(113 days)
GE Medical Systems, LLC
Auto Segmentation generates a Radiotherapy Structure Set (RTSS) DICOM with segmented organs at risk which can be used by dosimetrists, medical physicists, and radiation oncologists as initial contours to accelerate workflow for radiation therapy planning. It is the responsibility of the user to verify the processed output contours and user-defined labels for each organ at risk and correct the contours/labels as needed. Auto Segmentation may be used with images acquired on CT scanners, in adult patients.
Auto Segmentation is a post-processing software designed to automatically generate contours of organ(s) at risk (OARs) from Computed Tomography (CT) images in the form of a DICOM Radiotherapy Structure Set (RTSS) series. The application is intended as a workflow tool for initial segmentation of OARs to streamline the process of organ at risk delineation. The Auto Segmentation is intended to be used by radiotherapy (RT) practitioners after review and editing, if necessary, and confirming the accuracy of the contours for use in radiation therapy planning.
Auto Segmentation uses deep learning algorithms to generate organ at risk contours for the head and neck, thorax, abdomen and pelvis regions from CT images across 40 organ subregion(s). The automatically generated organ at risk contours are networked to predefined DICOM destination(s), such as review workstations supporting RTSS format, for review and editing, as needed.
The organ at risk contours generated with the Auto Segmentation are designed to improve the contouring workflow by automatically creating contours for review by the intended users. The application is compatible with CT DICOM images with single energy acquisition modes and may be used with both GE and non-GE CT scanner acquired images (contrast), in adult patients.
Here's an analysis of the acceptance criteria and study detailed in the provided document for the GE Medical Systems Auto Segmentation device:
1. Table of Acceptance Criteria and Reported Device Performance
OAR | Auto Segmentation (subject device) Dice Mean | Lower CI95 | Acceptance Criteria Type | Acceptance Criteria Dice Mean |
---|---|---|---|---|
Adrenal Left | 78.68% | 76.63% | Estimated | 68.0% |
Adrenal Right | 72.48% | 69.78% | Estimated | 68.0% |
Bladder | 81.50% | 78.33% | Deep learning | 80.0% |
Body | 99.50% | 99.38% | Atlas-based | 98.1% |
Brainstem | 87.69% | 87.15% | Deep learning | 88.4% |
Chiasma | 43.81% | 41.03% | Atlas-based | 11.7% |
Esophagus | 81.69% | 80.38% | Atlas-based | 45.8% |
Eye Left | 91.32% | 89.77% | Deep learning | 90.1% |
Eye Right | 90.25% | 88.23% | Deep learning | 89.9% |
Femur Left | 97.65% | 97.18% | Atlas-based | 71.6% |
Femur Right | 97.92% | 97.78% | Atlas-based | 70.8% |
Kidney Left | 92.53% | 90.30% | Deep learning | 86.8% |
Kidney Right | 94.82% | 93.48% | Deep learning | 85.6% |
Lacrimal Gland Left | 59.79% | 57.65% | Deep learning | 50.0% |
Lacrimal Gland Right | 58.09% | 55.81% | Deep learning | 50.0% |
Lens Left | 76.86% | 74.80% | Deep learning | 73.3% |
Lens Right | 79.09% | 77.40% | Deep learning | 75.6% |
Liver | 94.28% | 92.27% | Deep learning | 91.1% |
Lung Left | 97.70% | 97.38% | Deep learning | 97.4% |
Lung Right | 97.99% | 97.81% | Deep learning | 97.8% |
Mandible | 92.70% | 92.36% | Deep learning | 94.0% |
Optic Nerve Left | 79.22% | 77.99% | Deep learning | 71.1% |
Optic Nerve Right | 80.20% | 78.94% | Deep learning | 71.2% |
Oral Cavity | 87.43% | 86.20% | Deep learning | 91.0% |
Pancreas | 80.34% | 78.50% | Estimated | 73.0% |
Parotid Left | 84.35% | 83.27% | Deep learning | 65.0% |
Parotid Right | 85.55% | 84.48% | Deep learning | 65.0% |
Proximal Bronchial Tree (PBtree) | 84.94% | 83.71% | Atlas-based | 54.8% |
Inferior PCM (Pharyngeal Constrictor Muscle) | 70.51% | 68.72% | Estimated | 68.0% |
Middle PCM | 67.09% | 65.21% | Estimated | 68.0% |
Superior PCM | 59.57% | 57.85% | Estimated | 50.0% |
Pericardium | 93.58% | 92.00% | Atlas-based | 84.4% |
Pituitary | 75.62% | 74.12% | Deep learning | 78.0% |
Prostate | 79.67% | 77.60% | Atlas-based | 52.1% |
Spinal Cord | 88.55% | 87.43% | Deep learning | 87.0% |
Submandibular Left | 86.85% | 85.95% | Deep learning | 77.0% |
Submandibular Right | 85.70% | 84.79% | Deep learning | 78.0% |
Thyroid | 85.37% | 84.27% | Deep learning | 83.0% |
Trachea | 91.02% | 90.47% | Atlas-based | 69.2% |
Whole Brain | 98.53% | 98.46% | Estimated | 93.0% |
Note: The reported device performance (Dice Mean and Lower CI95) for almost all organs meets or exceeds the specified acceptance criteria. The only exception where the device's Dice Mean is slightly below the acceptance criteria is for Mandible (92.70% vs 94.0%) and Oral Cavity (87.43% vs 91.0%) and Pituitary (75.62% vs 78.0%), however there is no further discussion or justification provided in the text for these specific instances. The document does state that "The evaluation of the Dice mean for the Auto Segmentation algorithms demonstrates that the algorithm performance is in line with the performance of the predicate, as well as state of the art, recently cleared similar automated contouring devices."
2. Sample Size Used for the Test Set and Data Provenance
- Sample Size for Test Set: 302 retrospective CT radiation therapy planning exams (generating 2552 contours).
- Data Provenance: Multiple clinical sites in North America, Asia, and Europe. The demographic distribution includes adults (18-89 years old) of various genders and ethnicities from 9 global sources (USA, EU, Asia). The data was acquired using a variety of CT scanners and scanner protocols from different manufacturers.
- Retrospective/Prospective: Retrospective.
3. Number of Experts Used to Establish Ground Truth for the Test Set and Their Qualifications
- Number of Experts: Three (3).
- Qualifications of Experts: Independent, qualified radiotherapy practitioners.
- Comment: The document states that the ground truth annotations were established following RTOG and DAHANCA clinical guidelines.
4. Adjudication Method for the Test Set
- The document implies a consensus-based approach guided by clinical guidelines, as "ground truth annotations were established (...) manually by three independent, qualified radiotherapy practitioners," but it does not specify an explicit adjudication method like "2+1" or "3+1" for resolving disagreements between the three experts. The phrase "established following RTOG and DAHANCA clinical guidelines" suggests that these guidelines were used to define the correct contours.
5. Multi-Reader Multi-Case (MRMC) Comparative Effectiveness Study
- The document describes a qualitative preference study that involved three qualified radiotherapy practitioners reviewing the contours generated by the Auto Segmentation application. They assessed the adequacy of the generated contours for radiotherapy planning using a Likert scale.
- However, this was NOT a comparative effectiveness study of human readers with and without AI assistance. It was a study to determine the adequacy of the AI-generated contours themselves for initial use. Therefore, no effect size of human readers improving with AI vs. without AI assistance can be reported from this document.
6. Standalone Performance Study (Algorithm Only)
- Yes, a standalone performance study was conducted. The "Performance testing to evaluate the device's performance in segmenting organs-at-risk was performed using a database of 302 retrospective CT radiation therapy planning exams." The Dice Similarity Coefficient (DSC) was used as the primary metric to compare the Auto Segmentation generated contours to ground truth contours. The reported Dice Mean values and their 95% confidence intervals are direct metrics of the algorithm's standalone performance.
7. Type of Ground Truth Used
- Expert Consensus/Manual Annotation: Ground truth annotations were "established following RTOG and DAHANCA clinical guidelines manually by three independent, qualified radiotherapy practitioners."
8. Sample Size for the Training Set
- 911 different CT exams.
9. How the Ground Truth for the Training Set Was Established
- The document states that "The Auto Segmentation algorithms were developed and trained using a dataset of 911 different CT exams from several clinical sites from multiple countries. The original development and training data was used for radiotherapy planning..."
- It does not explicitly detail the process for establishing ground truth for the training set, but given the context of the test set ground truth and the overall development, it is highly probable it involved manual annotation by experts for radiotherapy planning purposes.
Ask a specific question about this device
(107 days)
GE Medical Systems, LLC
Spectral Bone Marrow is an automated image processing software application, utilizing deep learning technology for bone segmentation, to facilitate optimized visualization of bone marrow in spectral body and extremity CT images. Spectral Bone Marrow's output can be used during the review of traumatic and non-traumatic bone pathologies.
Spectral Bone Marrow is a deep-learning based software analysis package designed for the visualization of bone marrow based on GE HealthCare's spectral CT acquisitions data. Spectral Bone Marrow assists clinicians by providing an automatically generated fused material density image of the segmented bone region over a base monochromatic image optimized for the visualization of bone marrow during the review of traumatic or non-traumatic bone pathologies. The software creates a fully automated post processing workflow for creating these images and improving reader efficiency.
The Spectral Bone Marrow application involves generating a bone mask with a deep learning bone segmentation algorithm and uses this segmented region to define bone regions of water minus hydroxyapatite (Water(HAP)) material density images, which are subsequently colored. The application outputs the colored Water(HAP) material density images overlayed on monochromatic CT images or Virtual Unenhanced (VUE) images.
Additionally, the Spectral Bone Marrow application has the optional ability to automatically set the window width and window level of the color overlay images to provide for optimal visualization of bone marrow. The software provides multiplanar export of the fused images. Spectral Bone Marrow is hosted on GE's Edison Health Link (EHL) computational platform.
Here's a breakdown of the acceptance criteria and study details for the Spectral Bone Marrow device, based on the provided FDA 510(k) summary:
Acceptance Criteria and Device Performance
The provided document does not explicitly present a table of quantitative acceptance criteria for device performance. However, it states that the engineering bench testing showed that the algorithm is capable of accurately segmenting bones and is safe and effective for its intended use. Furthermore, the clinical assessment validated that the Spectral Bone Marrow software provides additional diagnostic value for the evaluation of bone marrow and increased overall reader efficiency.
This implies that the acceptance criteria revolved around demonstrating accurate bone segmentation, diagnostic value, and improved reader efficiency. The specific metrics and thresholds for "accuracy," "additional diagnostic value," and "increased efficiency" are not quantified in this summary.
Inferred Acceptance Criteria:
Criterion | Reported Device Performance |
---|---|
Bone Segmentation Accuracy | The engineering bench testing demonstrated that the algorithm is "capable of accurately segmenting bones." |
Diagnostic Value | The clinical assessment "validated that the Spectral Bone Marrow software provides additional diagnostic value for the evaluation of bone marrow." The radiologists provided an assessment of "image quality related to diagnostic use according to a Likert scale." |
Reader Efficiency Enhancement | The clinical assessment "validated...increased overall reader efficiency." Readers were asked to "rate their efficiency when using the algorithm compared to using without." |
Safety and Effectiveness (Overall) | The engineering bench testing showed the algorithm is "safe and effective for its intended use." The submission concludes that the device is "substantially equivalent and hence as safe and as effective as the legally marketed predicate device." "GE's quality system's design, verification, and risk management processes did not identify any unexpected results, or adverse effects stemming from the changes to the predicate." The substantial equivalence is based on the software documentation for a "Moderate" level of concern. |
Study Details
1. Sample Size Used for the Test Set and Data Provenance
- Test Set Sample Size: 146 retrospective Spectral CT series.
- Data Provenance: Retrospective, with the country of origin not explicitly stated, but the mention of "US board certified radiologists" suggests the data could be from the US or at least interpreted by US experts.
2. Number of Experts Used to Establish Ground Truth for the Test Set and Qualifications
- Number of Experts: Three.
- Qualifications: US board certified radiologists with expertise in both the evaluation of bone marrow and dual energy imaging review.
3. Adjudication Method for the Test Set
The document does not explicitly state a specific adjudication method like "2+1" or "3+1." It mentions that the ground truth for the 146 retrospective Spectral CT series was generated by three US board certified radiologists. This implies an expert consensus approach, but the specific decision-making process (e.g., majority vote, agreement by all three) is not detailed.
4. Multi-Reader Multi-Case (MRMC) Comparative Effectiveness Study
- Was an MRMC study done? Yes, a form of clinical assessment involving multiple readers was conducted. The clinical testing involved "three board certified radiologists" assessing a "representative set of clinical sample images" processed by the software. They evaluated "image quality related to diagnostic use according to a Likert scale" and rated "their efficiency when using the algorithm compared to using without." This setup resembles an MRMC study, focusing on reader performance with and without the AI.
- Effect size of human readers improve with AI vs without AI assistance: The document states that the assessment validated that the Spectral Bone Marrow software provides additional diagnostic value for the evaluation of bone marrow and increased overall reader efficiency. However, it does not provide quantitative effect sizes (e.g., specific percentage improvement in diagnostic accuracy or time saved). It merely confirms an improvement was observed and validated.
5. Standalone (Algorithm Only Without Human-in-the-Loop Performance) Study
Yes, a standalone study (referred to as "engineering bench testing") was done to evaluate the bone segmentation algorithm itself. This testing used a database of 146 retrospective Spectral CT series. The "result of the engineering bench testing showed that the algorithm is capable of accurately segment bones and is safe and effective for its intended use."
6. Type of Ground Truth Used
- Expert Consensus: The ground truth for the engineering bench testing (bone segmentation evaluation) was "generated by three US board certified radiologists."
- Clinical Assessment by Experts: For the clinical evaluation of diagnostic value and reader efficiency, the "assessment used retrospectively collected clinical cases" and each image was "read by each board certified radiologist who provided an assessment of image quality related to diagnostic use according to a Likert scale." This also relies on expert interpretation and assessment as the ground truth for these subjective measures.
7. Sample Size for the Training Set
The document does not specify the sample size for the training set used to develop the deep learning bone segmentation algorithm. It only refers to the training of a "deep learning bone segmentation algorithm."
8. How the Ground Truth for the Training Set Was Established
The document does not explicitly state how the ground truth for the training set was established. It only mentions that the bone segmentation algorithm uses deep learning technology. Typically, ground truth for training deep learning models involves manual annotation by experts, but this detail is not provided in the summary.
Ask a specific question about this device
Page 1 of 11