US11576557B2 - Method for supporting a user, computer program product, data medium and imaging system - Google Patents
Method for supporting a user, computer program product, data medium and imaging system Download PDFInfo
- Publication number
- US11576557B2 US11576557B2 US16/572,414 US201916572414A US11576557B2 US 11576557 B2 US11576557 B2 US 11576557B2 US 201916572414 A US201916572414 A US 201916572414A US 11576557 B2 US11576557 B2 US 11576557B2
- Authority
- US
- United States
- Prior art keywords
- image
- endoscope
- registration
- instrument
- uncertainty
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
- 238000003384 imaging method Methods 0.000 title claims abstract description 67
- 238000000034 method Methods 0.000 title claims abstract description 57
- 238000004590 computer program Methods 0.000 title abstract description 19
- 230000000694 effects Effects 0.000 claims description 36
- 238000012800 visualization Methods 0.000 claims description 27
- 230000001419 dependent effect Effects 0.000 claims description 17
- 230000008859 change Effects 0.000 claims description 2
- 238000005562 fading Methods 0.000 claims description 2
- 238000002679 ablation Methods 0.000 description 22
- 238000012545 processing Methods 0.000 description 21
- 239000000523 sample Substances 0.000 description 14
- 210000001519 tissue Anatomy 0.000 description 14
- 206010028980 Neoplasm Diseases 0.000 description 10
- 238000001574 biopsy Methods 0.000 description 10
- 230000006870 function Effects 0.000 description 9
- 230000003190 augmentative effect Effects 0.000 description 7
- 230000003902 lesion Effects 0.000 description 7
- 230000009466 transformation Effects 0.000 description 6
- 238000002594 fluoroscopy Methods 0.000 description 5
- 238000013188 needle biopsy Methods 0.000 description 5
- 230000003287 optical effect Effects 0.000 description 5
- 238000004422 calculation algorithm Methods 0.000 description 4
- 238000002591 computed tomography Methods 0.000 description 4
- 238000011161 development Methods 0.000 description 4
- 230000018109 developmental process Effects 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 210000004185 liver Anatomy 0.000 description 3
- 210000000056 organ Anatomy 0.000 description 3
- 230000005855 radiation Effects 0.000 description 3
- 241001522301 Apogonichthyoides nigripinnis Species 0.000 description 2
- 230000009471 action Effects 0.000 description 2
- 230000002411 adverse Effects 0.000 description 2
- 210000003484 anatomy Anatomy 0.000 description 2
- 239000003550 marker Substances 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 238000005259 measurement Methods 0.000 description 2
- 210000002307 prostate Anatomy 0.000 description 2
- 238000002604 ultrasonography Methods 0.000 description 2
- 238000000342 Monte Carlo simulation Methods 0.000 description 1
- 230000004913 activation Effects 0.000 description 1
- 230000006978 adaptation Effects 0.000 description 1
- 239000000470 constituent Substances 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 238000010859 live-cell imaging Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 238000002271 resection Methods 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
- 238000007619 statistical method Methods 0.000 description 1
- 230000001960 triggered effect Effects 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/12—Arrangements for detecting or locating foreign bodies
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/02—Arrangements for diagnosis sequentially in different planes; Stereoscopic radiation diagnosis
- A61B6/03—Computed tomography [CT]
- A61B6/032—Transmission computed tomography [CT]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00002—Operational features of endoscopes
- A61B1/00004—Operational features of endoscopes characterised by electronic signal processing
- A61B1/00009—Operational features of endoscopes characterised by electronic signal processing of image signals during a use of endoscope
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00002—Operational features of endoscopes
- A61B1/00004—Operational features of endoscopes characterised by electronic signal processing
- A61B1/00009—Operational features of endoscopes characterised by electronic signal processing of image signals during a use of endoscope
- A61B1/000095—Operational features of endoscopes characterised by electronic signal processing of image signals during a use of endoscope for image enhancement
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00002—Operational features of endoscopes
- A61B1/00043—Operational features of endoscopes provided with output arrangements
- A61B1/00045—Display arrangement
- A61B1/0005—Display arrangement combining images e.g. side-by-side, superimposed or tiled
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00163—Optical arrangements
- A61B1/00193—Optical arrangements adapted for stereoscopic vision
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/04—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
- A61B1/05—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances characterised by the image sensor, e.g. camera, being in the distal end portion
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B10/00—Other methods or instruments for diagnosis, e.g. instruments for taking a cell sample, for biopsy, for vaccination diagnosis; Sex determination; Ovulation-period determination; Throat striking implements
- A61B10/02—Instruments for taking cell samples or for biopsy
- A61B10/0233—Pointed or sharp biopsy instruments
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B10/00—Other methods or instruments for diagnosis, e.g. instruments for taking a cell sample, for biopsy, for vaccination diagnosis; Sex determination; Ovulation-period determination; Throat striking implements
- A61B10/02—Instruments for taking cell samples or for biopsy
- A61B10/04—Endoscopic instruments
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/05—Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves
- A61B5/055—Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves involving electronic [EMR] or nuclear [NMR] magnetic resonance, e.g. magnetic resonance imaging
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/44—Constructional features of apparatus for radiation diagnosis
- A61B6/4429—Constructional features of apparatus for radiation diagnosis related to the mounting of source units and detector units
- A61B6/4435—Constructional features of apparatus for radiation diagnosis related to the mounting of source units and detector units the source unit and the detector unit being coupled by a rigid structure
- A61B6/4441—Constructional features of apparatus for radiation diagnosis related to the mounting of source units and detector units the source unit and the detector unit being coupled by a rigid structure the rigid structure being a C-arm or U-arm
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/52—Devices using data or image processing specially adapted for radiation diagnosis
- A61B6/5211—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data
- A61B6/5223—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data generating planar views from image data, e.g. extracting a coronal view from a 3D image
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/52—Devices using data or image processing specially adapted for radiation diagnosis
- A61B6/5211—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data
- A61B6/5229—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image
- A61B6/5235—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image combining images from the same or different ionising radiation imaging techniques, e.g. PET and CT
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/52—Devices using data or image processing specially adapted for radiation diagnosis
- A61B6/5211—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data
- A61B6/5229—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image
- A61B6/5247—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image combining images from an ionising-radiation diagnostic technique and a non-ionising radiation diagnostic technique, e.g. X-ray and ultrasound
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/30—Determination of transform parameters for the alignment of images, i.e. image registration
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/313—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor for introducing through surgical openings, e.g. laparoscopes
- A61B1/3132—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor for introducing through surgical openings, e.g. laparoscopes for laparoscopy
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B10/00—Other methods or instruments for diagnosis, e.g. instruments for taking a cell sample, for biopsy, for vaccination diagnosis; Sex determination; Ovulation-period determination; Throat striking implements
- A61B10/02—Instruments for taking cell samples or for biopsy
- A61B10/04—Endoscopic instruments
- A61B2010/045—Needles
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
- A61B2034/107—Visualisation of planned trajectories or target regions
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/30—Surgical robots
- A61B2034/301—Surgical robots for introducing or steering flexible instruments inserted into the body, e.g. catheters or endoscopes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/48—Diagnostic techniques
- A61B6/486—Diagnostic techniques involving generating temporal series of image data
- A61B6/487—Diagnostic techniques involving generating temporal series of image data involving fluoroscopy
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10068—Endoscopic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10116—X-ray image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
Definitions
- the present embodiments relate to supporting a user.
- a technique that is nowadays widely used in this case combines data from different sources, where the corresponding data (e.g., images or data sets) are registered with each other so that the images or data sets may be overlaid in a positionally accurate manner within a shared coordinate system.
- the corresponding data e.g., images or data sets
- maximal precision of the registration is desirable. Errors, inaccuracies, or uncertainties in the registration may have disadvantageous consequences, accordingly.
- the present embodiments may obviate one or more of the drawbacks or limitations in the related art.
- a user may be supported in situations where there is an erroneous or uncertain registration.
- a method is used to support a user, for example, when examining an examination object or target object.
- a three-dimensional (3D) data set depicting the target object is first provided.
- the provision may in this case include or signify, for example, providing or making available a data store on which the 3D data set is stored. Equally, the provision may signify or include acquiring or recording (e.g., measuring) the 3D data set.
- the 3D data set may be recorded or have been recorded, for example, by a first imaging modality (e.g., by a computed tomograph, such as a computed tomography (CT) X-ray device, a magnetic resonance (MR) installation, or similar).
- a computed tomograph such as a computed tomography (CT) X-ray device, a magnetic resonance (MR) installation, or similar.
- CT computed tomography
- MR magnetic resonance
- the target object may be, for example, a patient, part of a patient, a specific organ, or a tissue sample, but also ultimately, any object or material that may be depicted by a corresponding imaging modality. For example, it is normal practice in industrial application fields for technical devices to be depicted and examined by X-radiation.
- the 3D data set may be a set (e.g., a collection of measured values or data points) or a 3D image, for example.
- At least one two-dimensional (2D) image of the target object is also acquired.
- This acquisition may include or signify, for example, a retrieval of the 2D image (e.g., corresponding data) from a data store that is provided, or equally a measurement or recording of the 2D image.
- the recording of the 2D image may be effected automatically, for example, or possibly triggered by activation of a foot switch or similar.
- the 2D image may be recorded or have been recorded by a second imaging modality.
- the second imaging modality may correspond to the first imaging modality in this case, or differ therefrom (e.g., variously).
- the first imaging modality may be a 3D X-ray device
- the second imaging modality may be a 2D X-ray device or the 3D X-ray device operated in a 2D mode.
- the 2D image and the 3D data set are therefore combined together or overlaid on each other in a positionally accurate manner (e.g., arranged in a shared coordinate system such that respective mutually corresponding parts or regions of the depicted target object overlap each other in the 2D image and the 3D data set in a corresponding overlay).
- the 2D image in this case may be arranged in three spatial dimensions or directions in space, which are also referred to as x, y, and z in the following.
- the uncertainty or inaccuracy of the registration is therefore greater in the specified spatial direction than in the two other spatial directions.
- Known methods for 2D/3D registration are often not equally precise in all three spatial directions. For example, it is frequently observed that the 2D/3D registration takes place with a relatively high degree of precision or accuracy in two orthogonal spatial directions (e.g., x and y), but may only be performed relatively inaccurately (e.g., with significantly less precision or accuracy in the third spatial direction, such as in the z-direction). This may occur or may be the case, for example, if a recording direction of the 2D image (e.g., a corresponding radiation direction or projection direction of the X-ray device) extends along the z-direction or z-axis.
- a recording direction of the 2D image e.g., a corresponding radiation direction or projection direction of the X-ray device
- the recording direction of the 2D image may therefore be specified as the spatial direction in which the 2D/3D registration has greatest uncertainty.
- the uncertainty or inaccuracy of the 2D/3D registration is therefore ultimately due to a relatively low or inadequate depth resolution of the 2D image.
- the spatial direction of greatest uncertainty of the 2D/3D registration (e.g., of greatest registration uncertainty) may be effected in this case by, for example, recognizing or tracking a pose of the X-ray device that is used to record the 2D image and of the target object.
- corresponding data it is then possible directly or, for example, by a corresponding object recognition algorithm or image processing algorithm, possibly with reference to further sensor data and/or user inputs, to automatically specify the recording direction for the 2D image (e.g., an angle of a corresponding axis in the shared coordinate system or in a coordinate system of the 3D data set).
- a corresponding object recognition algorithm or image processing algorithm possibly with reference to further sensor data and/or user inputs, to automatically specify the recording direction for the 2D image (e.g., an angle of a corresponding axis in the shared coordinate system or in a coordinate system of the 3D data set).
- a signal for aligning or for supporting the alignment of an instrument that is provided for the purpose of examining the target object is also generated and output as a function of the specified spatial direction of greatest uncertainty in order to support the user.
- the user may therefore be, for example, an operator of the X-ray device and/or the instrument (e.g., a doctor or a technician).
- the alignment of the instrument or the support of the alignment may signify that the instrument itself may be moved or repositioned in this case or for this purpose.
- the alignment may equally signify that another object (e.g., an imaging device, a patient couch, or similar) may be moved or repositioned, thereby producing a new or adapted relative alignment of the instrument (e.g., the alignment thereof relative to the object that is actually moved or repositioned).
- another object e.g., an imaging device, a patient couch, or similar
- the “alignment” may be a direct or relative alignment or, if applicable, the corresponding support of the alignment likewise.
- the instrument may be, for example, a camera, a sensor, a needle, a probe, or one of a plurality of similar or further instruments or devices. A number of examples concerning this are explained in greater detail below.
- the signal may be, for example, an acoustic, optical, and/or haptic indication, a graphical representation, a data signal or data set, a control signal, and/or similar. This is likewise explained in greater detail below.
- data or information is ultimately therefore provided to the user, and/or a measure or a process is initiated.
- the user is advantageously able, for example, to evaluate a respective situation (e.g., the target object) more accurately and/or more reliably, and/or the user is at least informed of the uncertainty of the 2D/3D registration and therefore, if applicable, a corresponding alignment, positioning, and/or effect of the instrument in the specified spatial direction.
- a corresponding registration uncertainty is often simply accepted as unavoidable.
- a multiplicity of further images, depictions, or data of the target object is recorded instead of the 2D image in order to obtain sufficient data for a more accurate registration in all spatial directions.
- this may disadvantageously result in a higher dosage or exposure of the target object as well as a time delay and/or logistical difficulties or collision problems in a workflow during the depiction or examination of the target object.
- a correspondingly more sophisticated and hence more expensive imaging system e.g., a corresponding X-ray device
- an endoscope e.g., a camera or an image recording device, such as a laparoscope
- a pose e.g., a position and orientation
- the two coordinate systems at least have a predetermined spatial location relationship to each other. This may be provided, for example, by corresponding measurement or calibration in advance.
- At least one overlay image is then generated from at least one endoscope image of the target object, as recorded by the endoscope, and the 3D data set.
- a stereo endoscope is used, for example, it is possible correspondingly to use two endoscope images (e.g., the two partial stereo images) and correspondingly to generate two overlay images, for example.
- the 2D/3D registration is used in order to generate the overlay image in a positionally accurate manner (e.g., in order to overlay the endoscope image and the 3D data set with positional accuracy).
- a respectively current pose of the target object may therefore be specified and related to the 3D data set or to a coordinate system of the 3D data set.
- the endoscope image and the 3D data set may be overlaid on each other in this case.
- the endoscope image may equally be combined (e.g., overlaid) with a representation or structure that is generated or derived from the 3D data set (e.g., a virtual model or a segmentation grid or a segmented representation).
- the endoscope image may therefore be, for example, a camera image or a video that depicts the target object in a manner that is optically realistic (e.g., true to reality).
- the overlay image may be conceived or generated as augmented reality, accordingly.
- An indication to the user is then output as the signal or as part of the signal, indicating that and how the user may change the alignment of the endo scope such that an acquisition direction of the endoscope extends at least substantially along the specified spatial direction of greatest registration uncertainty, such that a visualization error caused by the uncertainty of the 2D/3D registration is reduced in the overlay image.
- the acquisition direction of the endoscope corresponds in this case to an optical axis of the endoscope, and to that direction in or along which a central ray of a light pencil that is acquired by the endoscope is, for example, acquired.
- the acquisition direction of the endoscope may therefore be, for example, orthogonal to an image plane of the endoscope image.
- the alignment of the endoscope or the acquisition direction of the endoscope along the specified spatial direction may signify in this case an at least substantially parallel or antiparallel alignment.
- the acquisition direction of the endoscope may then be aligned along the recording direction of this one 2D image.
- the alignment of the endoscope or the acquisition direction thereof, at least substantially along the specified spatial direction may signify a corresponding alignment up to errors or deviations of, for example, 10° in this case.
- the specified spatial direction of greatest uncertainty of the 2D/3D registration then coincides at least substantially with an observation direction or depth direction of the overlay image. Although the uncertainty may still exist, experience shows that the uncertainty is perceived as less disruptive by an observer and may therefore be tolerated as a result of this alignment.
- a graphical aid may therefore be displayed by the signal or as the signal, indicating a current angular deviation between the acquisition direction of the endoscope and the specified spatial direction.
- An aid may be displayed or output by signal, as the signal, or as part of the signal in order to assist the user during the alignment of the endoscope. This may be realized by corresponding arrows, color coding, and/or a bulls-eye view, for example.
- Corresponding displays or representations may be displayed (e.g., output) on a corresponding display device (e.g., a monitor or a head-mounted display (HMD), such as a display that is attached to a head of the respective observer or user).
- a corresponding display device e.g., a monitor or a head-mounted display (HMD), such as a display that is attached to a head of the respective observer or user.
- HMD head-mounted display
- the acquisition direction of the endoscope need not necessarily correspond to a central or longitudinal axis of the endoscope itself (e.g., the housing or body thereof).
- a corresponding angular deviation may, however, be predetermined as a parameter or parameter value.
- a robot-guided endoscope is used as the instrument.
- a pose of the endoscope in a coordinate system of the 3D image and/or the 2D image is automatically tracked, at least up to a predetermined transformation of coordinates.
- an overlay image is generated from an endo scope image of the target object, as recorded by the endoscope, and the 3D data set.
- a control signal for the robot is then generated as the signal or as part of the signal, and is used by the robot to automatically align the endoscope such that the acquisition direction thereof extends along the specified spatial direction in order to reduce a visualization error caused by the uncertainty of the 2D/3D registration in the overlay image.
- the robot-guided endoscope is therefore automatically aligned (e.g., tracked) in order to give optimized visualization or image quality of the overlay image.
- a current pose of the target object may additionally be monitored and tracked, for example, and automatically taken into consideration when aligning the endoscope (e.g., when generating the control signal correspondingly).
- the endoscope may be aligned in a particularly rapid, precise, and reliable manner, accordingly.
- An optimal visualization quality of the overlay image may thus be provided in a particularly accurate and reliable manner. It is advantageous that no attentiveness or effort is required on the part of the respective user in this case, and therefore, the examination of the target object may be performed more quickly if required and with fewer distractions or less pressure for the respective user.
- an endoscope is likewise used as the instrument and the pose thereof in a coordinate system of the 3D image and/or the 2D image is tracked, at least up to a predetermined transformation of coordinates.
- an overlay image e.g., AR image
- a control signal for an imaging modality that is used to record the 2D image is then generated as the signal or as part of the signal.
- the imaging modality e.g., the X-ray device
- the imaging modality is automatically aligned by this control signal such that the imaging or recording direction thereof automatically extends along an acquisition direction of the endoscope, such that a visualization error, caused by the uncertainty of the 2D/3D registration, is reduced in the overlay image.
- an angulation of the second imaging modality may be automatically adapted or corrected as a function of the registration uncertainty and the current pose (e.g., orientation) of the endoscope and the acquisition direction thereof.
- the second imaging modality is therefore automatically moved or adjusted as an aid to the respective user such that the recording direction thereof (e.g., a corresponding X-ray projection direction) extends parallel or antiparallel to a viewing direction (e.g., the acquisition direction of the endoscope). It is thereby likewise possible to minimize the visualization error in the overlay image in an accurate, rapid, and reliable manner and to reduce the pressure on the user.
- an endoscope is likewise used as the instrument, and the pose thereof in a coordinate system of the 3D image and/or the 2D image is tracked, at least up to a predetermined transformation of coordinates.
- an overlay image is generated from an endoscope image of the target object, as recorded by the endoscope, and the 3D data set.
- a visualization of the overlay image is then adapted in a positionally dependent manner according to a degree of the uncertainty of the 2D/3D registration (e.g., by blurring, dimming, stretching, or distorting, and/or fading out a part of the overlay image).
- an uncertainty of the overlay or the overlay image may, for example, be dynamically dependent on a relative alignment of the endoscope or the acquisition direction of the endoscope in relation to the specified spatial direction of greatest uncertainty of the 2D/3D registration.
- the overlay image may be adapted accordingly (e.g., graphically processed) in order to signal or display to the respective observer or user the degree of the uncertainty or inaccuracy (e.g., of a visualization error).
- the blurring may be proportional to sine( ⁇ ), for example.
- arrows or other symbols may likewise display the angular deviation and/or the specified spatial direction.
- only the 3D data set or portions of the overlay image that are based on the 3D data set are adapted correspondingly.
- the endoscope image therefore remains, for example, unchanged as part of the overlay image, and only these superimposed overlay structures overlaid thereon are adapted or modified.
- a threshold value may be predetermined for the deviation of the angular deviation (e.g., the angle ⁇ between the acquisition direction of the endoscope and the specified spatial direction). If this threshold value or threshold value angle is reached or exceeded, the overlay structures (e.g., portions of the overlay image that are based on the 3D data set) may be faded out completely. In this way, the respective observer or user may be made aware of the uncertainty of the representation in a particularly intuitive and clear manner (e.g., in a manner that is particularly easy to understand and therefore particularly reliable).
- the visualization of the overlay image is only adapted in the specified spatial direction of greatest uncertainty of the 2D/3D registration.
- the blurring of the overlay structures only takes place in or along this spatial direction (e.g., in z-direction).
- the overlay structures may be stretched (e.g., represented in a stretched or distorted manner, such as in only the z-direction) in order to represent or display the uncertainty in the overlay image (e.g., a resulting visualization error or a resulting visualization uncertainty). It is thereby possible in a particularly simple and particularly accurate manner to display and convey to the respective observer or user the problem that the 2D/3D registration is particularly uncertain or inaccurate in the specified spatial direction. In this case, it is possible, for example, to dispense with additional markings, superimpositions, or symbols, thereby reducing or avoiding distractions or demands on the user or observer.
- a stereo endoscope is used as the instrument.
- a pose of the stereo endoscope in a coordinate system of the 3D image and/or the 2D image is then automatically tracked, at least up to a predetermined transformation of coordinates.
- an overlay image is generated from a stereoscopic endoscope image of the target object, as recorded by the stereo endoscope, and the 3D data set.
- a spatial effect in the overlay image e.g., an impression of depth in the specified spatial direction
- an improved adaptation may be achieved between the endoscope image or the stereoscopic depth location thereof and the 3D data set or portions of the overlay image generated therefrom. It is thereby possible to convey an improved impression (e.g., a more realistic impression) to the observer or user.
- the uncertainty is, for example, direction-dependent.
- the path may be not only a series of way points or positions in this case, but may also include or specify respective orientation information (e.g., orientations or poses of the instrument that are to be adopted or adjusted at the points of the path or along the path).
- respective orientation information e.g., orientations or poses of the instrument that are to be adopted or adjusted at the points of the path or along the path.
- the predetermined target may be, include, or indicate a desired position or pose of the instrument, a specific point of the target object, a sample collection from a specific tissue, tissue section, or part of the target object, for example, an action on a specific tissue, a specific tissue section or a specific part of the target object, and/or similar.
- a structure of the target object that is derived or predetermined from the 3D data set and/or from the 2D image and which may restrict or limit possible paths and/or alignments of the instrument may be taken into consideration.
- the path may be specified, for example, such that the instrument moves or is able to move along the path as little as possible in the specified spatial direction of greatest uncertainty and/or with minimal deviation between the alignment of the instrument and the specified spatial direction. It is thereby possible overall to improve or maximize a positioning accuracy of the instrument in relation to the target object or the predetermined target, and accordingly, the predetermined target may be reached with greater accuracy or reliability and, if applicable, less adverse effect on the target object.
- the instrument typically has a limited range of effect, for example.
- an additional way point or position of effect (e.g., deployment position) of the instrument may then be planned or provided as part of the path (e.g., in the specified spatial direction), such that the predetermined target (e.g., a lesion of the target object) is included in or covered by the range of effect of the instrument despite the registration uncertainty.
- the effect of the instrument may be or include, for example, depiction of the predetermined target, optionally with a predetermined minimal image sharpness, or also manipulation of the target object or the predetermined target (e.g., collection of a sample and/or an ablation).
- the instrument may be a biopsy needle for the purpose of sample collection, by which a plurality of samples may be collected along the path (e.g., in a movement direction of the biopsy needle but not in sideways directions relative thereto).
- the path is planned or specified such that the instrument, while being guided along the specified path in the region of the predetermined target according to the application case (e.g., according to the shape of the predetermined target region), moves along or, for example, perpendicular to the specified spatial direction of greatest uncertainty of the 2D/3D registration.
- the application case e.g., according to the shape of the predetermined target region
- an ablation zone in the case of a needle ablation may not be circular or spherical but elliptically elongated along an axis of an ablation needle that is used correspondingly.
- tumor tissue is to be removed, the present embodiments make it possible to provide with greater probability or reliability that the tumor tissue is actually situated completely within a volume or range of effect of the ablation (e.g., within the ablation zone), and that surrounding tissue is not affected or is affected as little as possible in this case.
- the instrument itself is represented in the overlay image (e.g., as an image or a model).
- the representation of the instrument and/or a range of influence or range of effect of the instrument in the overlay image is then broadened in a spatially anisotropic manner as a function of the respective local uncertainty of the 2D/3D registration.
- the representation of the instrument and/or a range of effect of the instrument may therefore be distorted in accordance with the uncertainty of the 2D/3D registration and thus in accordance with an uncertainty of the visualization or representation of the overlay image.
- the type, degree, and direction of the uncertainty or of a corresponding visualization error may thus be displayed to the user in a manner that is particularly simple, intuitively understandable, and not unnecessarily distracting.
- the instrument actually has, for example, a cylindrical or rod-shaped form or shape, this may be broadened (e.g., elliptically) in the overlay image according to the current alignment of the instrument in relation to the specified spatial direction.
- the representation of the range of effect may be adapted correspondingly.
- the anisotropic broadening of the representation may equally signify that the representation of the instrument in the overlay image is surrounded by a corresponding broadening zone or uncertainty zone (e.g., a correspondingly marked spatial region).
- a path for the instrument may then be planned or traced automatically or by the respective user.
- the path guides the instrument such that the entire broadened representation of the instrument or a range of effect of the instrument passes through the predetermined target (e.g., a lesion or similar) and therefore, acquires the target completely. It is thus possible to provide that despite the registration uncertainty, all possible instrument paths within a region indicated or displayed by the broadened representation actually pass through or encounter the predetermined target (e.g., the lesion or a tumor or similar).
- the range of effect of the instrument is smaller than the broadened representation, it is possible to plan or trace, for example, a plurality of positions of effect or deployment positions accordingly (e.g., a plurality of sample collection points or a plurality of adjacent paths through the predetermined target or in the region of the predetermined target), such that an overlap of the predetermined target with the entire broadened representation and/or with the range of effect of the instrument is provided.
- a plurality of positions of effect or deployment positions accordingly (e.g., a plurality of sample collection points or a plurality of adjacent paths through the predetermined target or in the region of the predetermined target), such that an overlap of the predetermined target with the entire broadened representation and/or with the range of effect of the instrument is provided.
- the instrument is, for example, a needle and if the uncertainty of the 2D/3D registration is relatively low in the x- and y-directions but is relatively high in the z-direction, the needle would be represented, for example, as a line during alignment in or along the z-direction, and would be represented as increasingly broad as a result of increasing deviation from this alignment.
- An alignment of the needle in or along the z-direction signifies in this case that a longitudinal extension direction or main extension direction of the needle runs in the z-direction.
- At least one deployment point of the instrument is specified automatically and a corresponding indication (e.g., corresponding position data), a corresponding recommendation, or a corresponding identification code is output as the signal.
- the deployment point in this context is therefore a position (e.g., with an associated orientation of the instrument), at which the instrument is to develop an intended effect and/or is to be positioned in order to develop an effect.
- the deployment point may indicate, for example, a point or a part of the target object at which or from which a sample is to be collected by the biopsy needle or at which material or tissue is to be ablated by the ablation needle. It is therefore possible, as a function of the direction-dependent registration uncertainty, taking into consideration the direction of effect and the anisotropy of the range of effect of the instrument, to specify and output a recommendation for at least one deployment point (e.g., a position of effect or application position) for the instrument. In this way, it is possible automatically to provide in a reliable manner that the effect of the instrument actually reaches a predetermined target in spite of the registration uncertainty. Without the corresponding indication, it may otherwise be difficult for a user, due to the registration uncertainty and possible resulting visualization error, to estimate whether and if necessary with what probability the predetermined target is actually reached in the case of a respectively current positioning and alignment of the instrument.
- a respectively current alignment of the instrument is automatically acquired (e.g., tracked).
- a probability that a predetermined target is reached by the instrument in the acquired alignment is automatically calculated.
- the calculated probability is then output as the signal or as part of the signal. It is therefore possible to calculate, for example, by a statistical analysis such as a Monte-Carlo simulation or similar, a probability with which a specific path or a combination of paths of the instrument will result in the predetermined target (e.g., a lesion or a tumor) actually being reached or a sample collection from a specific tissue region actually being achieved.
- This probability is then displayed to the respective user, who may adapt the path or the alignment of the instrument accordingly and/or perform an improved (e.g., more accurate) registration until a probability of reaching the target, the probability being sufficient in the respective particular case, is reached or established.
- the success of a respective application may thus be improved or verifiable with statistical objectivity (e.g., predicted in a clear manner). If an examination of a collected sample gives a negative result, for example, this may be classified or evaluated with reference to the probability that the sample was collected from the predetermined target tissue.
- the probability is established by the path that was used for the instrument at the time of the sample collection. If it is verifiably provided as a result of using the present embodiments that the probability of reaching the target is or was 100% or close to 100%, it is optionally possible to forgo a further sample collection.
- an uncertainty, inaccuracy, or tolerance when guiding or positioning and/or tracking the instrument may also be added as a statistically independent variable to the uncertainty of the 2D/3D registration (e.g., combined therewith).
- the probability of reaching the predetermined target may be calculated even more accurately and reliably.
- some or all of these uncertainties or probabilities may likewise be taken into consideration when the representation of the instrument is adapted as described above (e.g., by further broadening the representation correspondingly). It may thus be conveyed to the respective user, for example, how far and/or with what probability an actual path or an actual position of the instrument may deviate from a planned path or a planned position.
- the spatial direction of greatest uncertainty of the 2D/3D registration may be specified or updated afresh in each case, the 2D/3D registration can be performed or updated afresh in each case, and/or a correspondingly updated signal may be generated and output.
- a further aspect relates to a computer program product including a computer program.
- the computer program in this case encodes or represents a method according to the present embodiments (e.g., the method acts of at least one embodiment variant of the method).
- the computer program product or the computer program is designed and configured to be loaded into a storage device of a data processing device (e.g., an imaging system) and to execute the method when the computer program is executed by the data processing device.
- the data processing device may include a corresponding processor device (e.g., at least a microchip) or a microcontroller that is connected to the storage device.
- the computer program product may therefore include program means for executing the method.
- the program means may include further components that are not explicitly cited here (e.g., corresponding control commands, register references, and/or similar).
- a further aspect relates to a data medium (e.g., a non-transitory computer-readable storage medium) or data store that is, for example, electronic and/or electronically readable and/or computer readable (e.g., for an imaging system).
- a data medium e.g., a non-transitory computer-readable storage medium
- data store that is, for example, electronic and/or electronically readable and/or computer readable (e.g., for an imaging system).
- Stored on the data medium in this case is a program code that may be executed, for example, automatically by a data processing device and includes a computer program that encodes or represents at least one embodiment variant of the method (e.g., corresponding method acts) in order to effect an execution of the method (e.g., the corresponding method acts) when the computer program or program code is executed by the data processing device.
- At least one embodiment variant of the computer program product or the computer program included therein may therefore be stored on the data medium.
- a further aspect relates to an imaging system.
- the imaging system has an acquisition device for acquiring a 3D data set that depicts a target object, and at least one 2D image of the target object.
- the 3D data set may have been recorded at an earlier time point (e.g., by another system).
- the acquisition of the 3D data set may signify, for example, the retrieval thereof from an electronic or computer readable data store or similar.
- the acquisition device itself may be designed and configured to record (e.g., measure) the 3D data set.
- the imaging system also includes a data processing device.
- the data processing device includes a processor device and a data medium that is connected thereto.
- the imaging system also includes an output device for outputting an automatically generated signal.
- the output device may be or include, for example, an interface of the data processing device, a program module of the computer program or program code, and/or a display device such as a screen or an HMD.
- the imaging system of one or more of the present embodiments is therefore designed and configured, for example, to perform or execute the method of one or more of the present embodiments. Accordingly, the imaging system may therefore have some or all of the properties and/or components or parts cited in connection with the other aspects of the present embodiments (e.g., in connection with the method, the computer program product, and the data medium).
- the imaging system may be configured in this case as a compact system or device (e.g., a system or device that is integrated or arranged in a housing). Equally, at least some components or parts of the imaging system (e.g., therefore, the acquisition device and the data processing device) may be arranged at different locations, these being spatially distributed or spatially separate from each other.
- the data processing device of the imaging system e.g., as a computer center or part of a computer center
- the data processing device may be arranged “on-premise” in this case.
- the data processing device and the acquisition device may be arranged on the same operational site (e.g., on the same company premises, factory premises, or hospital premises).
- the data processing device may, however, be, for example, a remotely located server device (e.g., a cloud server or remote server). This may then be arranged in a computer center and connected to the acquisition device of the imaging system via, for example, a wide area network (WAN), such as, for example, the Internet.
- WAN wide area network
- the method of one or more of the present embodiments may be executed, for example, during or in preparation for an interventional or surgical action, it does not itself however include or even assume any surgical steps. Rather, the method of one or more of the present embodiments may be regarded as a method for operating an imaging system or a data processing device, where received data is processed, and output data is generated and output in the form of the signal. This may be completely separate and independent from an actual interventional or surgical application. The method may not serve to automatically effect a diagnosis.
- FIG. 1 shows a schematic representation of one embodiment of an imaging system
- FIG. 2 shows an exemplary schematic program model of a method for supporting a user of the imaging system of FIG. 1 .
- the exemplary embodiments explained below are variants of the invention.
- the components of the embodiment variants, as described represent in each case individual features of the invention, being regarded as independent of each other.
- the features respectively develop the invention independently of each other likewise and are therefore also to be considered as constituents of the invention individually or in a combination other than that shown or described.
- the embodiment variants as described may also be supplemented by further features of the invention described above.
- FIG. 1 shows a schematic representation of one embodiment of an imaging system 1 configured as, for example, a C-arm X-ray device.
- the imaging system 1 in this case has a radiation source 3 and a detector 2 for detecting X-radiation emitted by the radiation source 3 .
- a target object may be depicted along or in a recording direction 4 in this case.
- a patient 6 situated on a patient support 5 is depicted as a target object at least locally by the imaging system 1 .
- the imaging system 1 also has a data processing device 7 for processing sensor data or image data provided by the detector 2 and, if applicable, further data received or acquired via an interface 8 of the data processing device 7 .
- the data processing device 7 also has a storage device 9 and a processor device 10 that is connected to the storage device 9 and the interface 8 .
- Stored on or in the storage device 9 in this case is a computer program (e.g., program code; instructions) that encodes or represents a method for supporting a user.
- This computer program or this program code may be executed by the processor device 10 in order to execute the method.
- FIG. 2 shows an exemplary program model 11 in the form of an extract, with schematically indicated program modules 18 to 24 for such a method for supporting a user of the imaging system 1 .
- the imaging system 1 in this case further includes a tracking system 12 that is configured as, for example, a stereo camera and/or for electromagnetic tracking, for example.
- An instrument such as a laparoscope 13 represented in FIG. 1 may be tracked by the tracking system 12 .
- this has a corresponding marker 14 in this case.
- the laparoscope 13 it is possible optically to depict the patient 6 in a viewing or acquisition direction 15 of the laparoscope 13 .
- a plurality of patient markers 16 Arranged in the patient 6 in this case are a plurality of patient markers 16 that may likewise be acquired by the tracking system 12 in order to specify and track a pose of the patient 6 .
- the imaging system 1 in this case further includes a display device 17 that is configured as a screen or monitor by way of example in FIG. 1 .
- FIG. 1 A use of the imaging system 1 for supporting the user and the program model 11 illustrated in FIG. 2 are explained in greater detail below with reference to FIG. 1 and FIG. 2 .
- a 3D data set (e.g., a preoperative or pre-interventional 3D volume scan of the patient 6 ) is acquired for this purpose by the program module 18 in this case.
- At least one 2D image of the patient 6 is then acquired by the program module 18 and the imaging system 1 as part of the intervention.
- An endoscope image and/or endoscope video of the patient 6 is acquired by the program module 18 and is recorded by the laparoscope 13 in this case.
- the objective is to initially generate an overlay image (e.g., an AR image) from these different sets of data.
- overlaid structures or superimpositions are in this case to be registered with the endoscope image or endoscope video (e.g., a laparoscope image or laparoscope video) as precisely as possible in relation to a real anatomy of the patient 6 .
- the endoscope image or endoscope video e.g., a laparoscope image or laparoscope video
- registration that is as precise as possible in relation to the real anatomy of the patient 6 is necessary, for example, for navigated guidance of needles (e.g., based on laser guidance or optical navigation).
- needle paths are often calculated based on a 2D/3D registration or 2D/2D/3D registration of one or two 2D X-ray projections (e.g., 2D X-ray images) with the 3D volume scan or a corresponding 3D data set.
- a further example is needle guidance with the aid of 2D or 3D ultrasound, where an algorithm computes a 2D/3D registration with a 3D ultrasound scan or with a 3D CT scan or 3D MR scan.
- the registrations described are not equally precise in all spatial directions.
- the two reciprocally orthogonal spatial directions are a designated x-direction and a designated y-direction without restricting general applicability, while a third spatial direction (e.g., a z-direction) may only be registered very inaccurately.
- this case occurs, for example, if during an operation or intervention a 2D X-ray image is recorded by the imaging system 1 along the z-direction or z-axis, which then corresponds or may correspond to the recording direction 4 .
- a previously recorded or acquired (e.g., available) 3D data set is then to be registered with this 2D X-ray image by a 2D/3D registration.
- This 2D/3D registration or 2D/2D/3D registration is performed by a program module 19 in this case.
- the 2D/3D registration is then used to overlay the 3D data set, or a representation derived therefrom such as, for example, a segmented mesh, onto the endoscope image by or in the form of an AR representation (e.g., by or in the form of augmented reality). Additionally or alternatively, the 2D/3D registration is used as a basis for planning the path of a needle (e.g., a biopsy needle or ablation needle or other instrument).
- a needle e.g., a biopsy needle or ablation needle or other instrument.
- the program module 20 acquires and tracks the laparoscope 13 with reference to the marker 14 , and the patient 6 with reference to the patient markers 16 (e.g., respective poses of the laparoscope 13 and the patient 6 are acquired and tracked by the program module 20 ).
- This provides that the acquisition direction 15 of the laparoscope 13 is therefore known and is therefore available for data processing.
- the recording direction 4 is also known from corresponding control parameters or operating parameters of the imaging system 1 , and is therefore available for data processing.
- the program module 21 specifies the acquisition direction 15 relative to the recording direction 4 (e.g., an angle ⁇ between the recording direction 4 and the acquisition direction 15 ).
- this inaccuracy or uncertainty along a spatial axis or spatial direction may be very disruptive for an observer or the user, and depending on the viewing direction of the endoscope 13 (e.g., depending on the relative alignment of the acquisition direction 15 ) may have varying degrees of effect or noticeability.
- An imprecise overlay of which the user (e.g., a surgeon) is unaware conceals the danger of an erroneous decision, and in the worst case, may result in, for example, an incision being made at the wrong point of the patient 6 or a specific organ.
- a lesion that has been identified with the aid of the 3D data set e.g., 3D imaging
- an ablation zone for a needle ablation is not circular, but extends further along a needle axis than in the other directions. However, a tumor is to lie completely within a volume of the ablation zone in this case.
- the overlay e.g., the overlay image or AR image
- the spatially anisotropic precision or certainty of the 2D/3D registration was not previously taken into consideration during the planning of needle paths.
- One possible solution is to improve or provide a registration accuracy or registration certainty in all spatial directions by using a plurality of X-ray projection directions (e.g., a plurality of 2D X-ray images from or having different angulations) using a biplanar X-ray system, or using 3D X-ray images also during the respective intervention.
- a plurality of X-ray projection directions e.g., a plurality of 2D X-ray images from or having different angulations
- 3D X-ray images also during the respective intervention.
- the imaging system 1 is configured as a C-arm X-ray device for intraoperative 3D imaging and 2D fluoroscopy.
- Optical live imaging is effected by the laparoscope 13 .
- This has, for example, an angled lens, such that the acquisition direction 15 of the laparoscope 13 is angled relative to a direction of longitudinal extension or a shaft or housing of the laparoscope 13 (e.g., by 30°). It is intended in this case to perform, for example, a laparoscopic examination on the patient 6 lying on the patient support 5 (e.g., an OP table). For example, it is intended to resection a tumor on the liver.
- the user e.g., a surgeon uses the laparoscope 13 for this purpose.
- the laparoscope 13 is tracked with respect to a position and an orientation (e.g., a pose) using the tracking system 12 . It is optionally possible for further surgical instruments not individually represented here (e.g., forceps, a pointer, an electro-cauterizer, and/or similar) to be acquired and tracked by the tracking system 12 or by a further tracking system. Equally, it is also possible, for example, to specify position and location of the laparoscope 13 with reference to X-ray images or fluoroscopy images recorded by the imaging system 1 .
- further surgical instruments not individually represented here (e.g., forceps, a pointer, an electro-cauterizer, and/or similar) to be acquired and tracked by the tracking system 12 or by a further tracking system. Equally, it is also possible, for example, to specify position and location of the laparoscope 13 with reference to X-ray images or fluoroscopy images recorded by the imaging system 1 .
- the tracking system 12 or a coordinate system of the tracking system 12 in which the laparoscope 13 is acquired and tracked is calibrated to a coordinate system of the X-ray images and 3D data set recorded by the imaging system 1 . It is then possible, using a predetermined transformation of coordinates, to calculate how objects with 3D coordinates that were calculated from the X-ray images and/or from the 3D data set may be represented as overlaid in a positionally correct manner in the endoscope image recorded by the laparoscope 13 or a corresponding coordinate system.
- the laparoscope 13 may be a stereo laparoscope in this case.
- the display device 17 may be a stereoscopic display device (e.g., a 3D monitor or a stereo head-mounted display (HMD)).
- HMD stereo head-mounted display
- the laparoscope 13 view is aligned parallel to the X-ray projection device of the imaging system 1 .
- This provides that the acquisition direction 15 in this case is therefore aligned parallel or antiparallel to the recording direction 4 .
- An image plane of the endoscope image is then oriented along the spatial directions x, y.
- the virtual overlay structures derived or generated from the 3D data set may then be superimposed on the endoscope image in a positionally accurate manner (e.g., in a manner that fits exactly) by virtue of the relatively precise registration in the x- and y-directions.
- the described registration inaccuracy or registration uncertainty in the z-direction (e.g., along the recording direction 4 ) is then effectively unnoticeable or not significantly noticeable.
- the laparoscope 13 is a stereo laparoscope, it is additionally possible to generate a 3D depth impression using a vergence and/or parallax of the superimposition between a left-hand part-image and a right-hand part-image of the stereo endoscope image, where the corresponding depth direction in this case is likewise parallel to the X-ray projection direction z (e.g., the recording direction 4 ).
- the 2D/3D registration or the overlay in the overlay image may even be inaccurate here, since the eye of an observer or user does not in reality perceive small deviations or inaccuracies in a depth direction as significantly disruptive, and therefore, the inaccuracy or uncertainty in the depth direction may be tolerated in this case.
- the angle ⁇ between the viewing or acquisition direction 15 of the laparoscope 13 and the X-ray projection direction is other than zero.
- the image plane of the endoscope image is no longer parallel to the spatial directions x, y.
- the greater registration uncertainty or a greater registration error in the z-direction has a direct effect on the 2D overlay of the preoperative images (e.g., proportional to sine( ⁇ )).
- the registration uncertainty is initially specified by the program module 22 with respect to a direction and a level.
- Previously known methods may be used for this purpose (e.g., based on the angle ⁇ ; a relative alignment of the recording direction 4 and the acquisition direction 15 to each other, and predetermined empirical values and/or return values of a registration algorithm that is used for the 2D/3D registration and/or for the generation of the overlay image).
- the overlay image is then generated by the program module 23 .
- a signal for aligning the laparoscope 13 is then generated and output automatically as a function of the specific registration uncertainty using the program module 24 .
- the overlays in the overlay image are graphically adapted or processed in order to signal the degree of the inaccuracy or uncertainty to the user. This may take place, for example, by blurring the overlay structures (e.g., proportional to sine( ⁇ ); by blurring in z-direction only). Further possibilities include, for example, progressive (e.g., spatially gradual) dimming of the overlays or overlay structures, superimposition of arrows and/or other symbols in z-direction, and/or stretching of the overlay structures or overlay contents in z-direction. If the directional deviation is too great (e.g., if the angle ⁇ exceeds a predetermined threshold value), the overlay structures may be faded out completely.
- the overlay structures may be faded out completely.
- the imaging system 1 or a navigation system may therefore provide a graphical aid that displays a respectively current directional deviation between the recording direction 4 and the acquisition direction 15 (e.g., a magnitude of the angle ⁇ ).
- the imaging system 1 or the navigation system further provides an aid to support the alignment of the acquisition direction 15 parallel or antiparallel to the recording direction 4 .
- a bulls-eye representation or navigation may be provided for this purpose.
- the acquisition direction 15 may be accepted as appropriate in each case, and an angulation of the imaging system 1 (e.g., of the C-arm and hence the recording direction 4 ) may be moved using the aid that is provided to the respective user or may be moved automatically such that the recording direction 4 is aligned parallel or antiparallel to the acquisition direction 15 .
- An aid is then available or used for the parallel alignment of the acquisition direction 15 , and/or an aid (e.g., an automatic operation) is then available or used to align the recording direction 4 .
- a registration uncertainty and/or visualization uncertainty may be displayed to the respective user in this case by adapting overlay contents in the overlay image correspondingly.
- the laparoscope 13 serves merely as an example. Any other desired endoscope, camera, operation microscope, or similar may likewise be used.
- a needle in which a needle is used as an instrument, it may be intended, for example, to perform a needle biopsy on the patient 6 because, for example, a suspicious lesion was found with the aid of a preoperative 3D CT or MR scan.
- This may be situated, for example, on a liver or a prostate of the patient 6 .
- Needle guidance e.g., guidance or positioning (movement) of the biopsy needle
- a 2D fluoroscopy recording is performed by the imaging system 1 .
- the preoperative 3D scan is then registered with the current location of the patient 6 .
- an angle between these two angulations may be significantly smaller than an optimal value of 90°, resulting in an anisotropic accuracy (e.g., an anisotropic or direction-dependent uncertainty) of the registration.
- an anisotropic accuracy e.g., an anisotropic or direction-dependent uncertainty
- This inaccuracy or uncertainty with regard to path planning which is dependent on the direction of the biopsy needle and typically anisotropic, may be displayed to the respective user or doctor in the context of path planning so that the user or doctor may take this into consideration when determining the needle path.
- the inaccuracy or uncertainty may be taken into consideration correspondingly in the case of automatic path planning
- a corresponding planning tool or program module e.g., the program module 24
- corresponding values are specified along a needle axis (e.g., a main extension direction of the biopsy needle) or along a the needle path.
- a broadening (e.g., elliptical) of a representation of the needle in the overlay image, the broadening being dependent on the uncertainty of the registration or the needle path or on a corresponding representation, may be displayed live and, for example, intraoperatively during the needle navigation.
- One example is a needle guide that is overlaid as augmented reality onto an endoscope image or laparoscope video that was recorded by the laparoscope 13 .
- a corresponding region of uncertainty is dissolved in the region of the needle, and the lesion is displayed virtually.
- a tumor is to lie completely within an ablation zone in the case of a needle ablation.
- a typically anisotropic uncertainty zone that is dependent on the respective alignment of the needle path or on the needle itself is represented in the overlay image around the tumor and indicates or displays the uncertainty of the registration. It is optionally possible, as in the other examples, for a spatial uncertainty of the instrument guidance and/or the tracking system 12 to be taken into consideration and added correctly.
- the ablation zone around the needle is visualized by the planning module (e.g., as a function of parameters of the respective ablation, such as a power and/or a duration, and if applicable, as a function of or with reference to a model or simulation of a corresponding tissue in the region of the ablation zone).
- parameters or parameter values and a pose (e.g., positioning or placement) of the ablation needle are then selected such that as far as possible the tumor and the whole spatially anisotropic region of uncertainty around the tumor lie within the ablation zone that is represented. It is optionally possible to plan a plurality of adjacent ablation volumes until the whole certainty zone is covered collectively. This may be considered analogous to planning or providing a plurality of needle paths that do not necessarily have to be arranged or run parallel to each other in this case.
- Needle biopsies and other instrumental applications may be more selective.
- the method described also allows 2D/3D or 2D/2D/3D registrations to be used more effectively for needle guidance, whereby a dosage for an additional intraoperative 3D scan may often be avoided.
- Relevant support for ablation procedures is also possible, where, for example, an actual or effective ablation zone may be predicted with greater precision using the planning tool, if applicable.
- a perception that there is a spatial direction that has significantly greater registration uncertainty than the other spatial directions is utilized to achieve improved visualization and, if applicable, patient benefits by virtue of improved and, for example, more precise instrument guidance.
- a respective endoscope and/or, for example, a respective X-ray device may be aligned such that the fewest possible registration uncertainties or resulting visualization errors are noticeable for a respective observer.
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Engineering & Computer Science (AREA)
- Surgery (AREA)
- Medical Informatics (AREA)
- Physics & Mathematics (AREA)
- Heart & Thoracic Surgery (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Biomedical Technology (AREA)
- Public Health (AREA)
- Veterinary Medicine (AREA)
- Molecular Biology (AREA)
- Animal Behavior & Ethology (AREA)
- General Health & Medical Sciences (AREA)
- Pathology (AREA)
- Radiology & Medical Imaging (AREA)
- Biophysics (AREA)
- Optics & Photonics (AREA)
- High Energy & Nuclear Physics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Signal Processing (AREA)
- Robotics (AREA)
- Multimedia (AREA)
- Pulmonology (AREA)
- Apparatus For Radiation Diagnosis (AREA)
- Endoscopes (AREA)
Abstract
Description
Claims (17)
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP18195330 | 2018-09-19 | ||
EP18195330.8 | 2018-09-19 | ||
EP18195330.8A EP3626176B1 (en) | 2018-09-19 | 2018-09-19 | Method for supporting a user, computer program product, data carrier and imaging system |
Publications (2)
Publication Number | Publication Date |
---|---|
US20200085281A1 US20200085281A1 (en) | 2020-03-19 |
US11576557B2 true US11576557B2 (en) | 2023-02-14 |
Family
ID=63642787
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/572,414 Active 2041-07-23 US11576557B2 (en) | 2018-09-19 | 2019-09-16 | Method for supporting a user, computer program product, data medium and imaging system |
Country Status (4)
Country | Link |
---|---|
US (1) | US11576557B2 (en) |
EP (1) | EP3626176B1 (en) |
JP (1) | JP6952740B2 (en) |
CN (1) | CN110916702B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6988001B2 (en) * | 2018-08-30 | 2022-01-05 | オリンパス株式会社 | Recording device, image observation device, observation system, observation system control method, and observation system operation program |
CN111631754B (en) * | 2020-05-26 | 2021-07-09 | 清华大学 | Automatic sampling system of pharynx swab |
Citations (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2008120136A1 (en) | 2007-03-30 | 2008-10-09 | Koninklijke Philips Electronics N.V. | 2d/3d image registration |
JP2009056238A (en) | 2007-09-03 | 2009-03-19 | Olympus Medical Systems Corp | Endoscope apparatus |
US20090088632A1 (en) * | 2007-10-02 | 2009-04-02 | Siemens Corporate Research, Inc. | Method for Dynamic Road Mapping |
US20090088773A1 (en) * | 2007-09-30 | 2009-04-02 | Intuitive Surgical, Inc. | Methods of locating and tracking robotic instruments in robotic surgical systems |
US20120165655A1 (en) * | 2009-06-26 | 2012-06-28 | Dirk Mucha | Method for generating position data of an instrument |
US20120289825A1 (en) * | 2011-05-11 | 2012-11-15 | Broncus, Technologies, Inc. | Fluoroscopy-based surgical device tracking method and system |
US20150030229A1 (en) | 2013-07-24 | 2015-01-29 | Anja Borsdorf | Methods for Updating 2D/3D Registration on Movement and Computing Device |
US20150085981A1 (en) | 2013-09-24 | 2015-03-26 | Siemens Aktiengesellschaft | Method of image registration in a multi-source/single detector radiographic imaging system, and image acquisition apparatus |
US20160260220A1 (en) | 2015-03-05 | 2016-09-08 | Broncus Medical Inc. | Gpu-based system for performing 2d-3d deformable registration of a body organ using multiple 2d fluoroscopic views |
WO2016182550A1 (en) | 2015-05-11 | 2016-11-17 | Siemens Aktiengesellschaft | Method and system for registration of 2d/2.5d laparoscopic and endoscopic image data to 3d volumetric image data |
JP2017507708A (en) | 2014-02-11 | 2017-03-23 | コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. | Spatial visualization of the internal thoracic artery during minimally invasive bypass surgery |
JP2017086819A (en) | 2015-11-17 | 2017-05-25 | 東芝メディカルシステムズ株式会社 | Medical image diagnostic apparatus |
US20170165008A1 (en) | 2015-12-14 | 2017-06-15 | Nuvasive, Inc. | 3D Visualization During Surgery with Reduced Radiation Exposure |
US20170209071A1 (en) | 2014-07-28 | 2017-07-27 | Intuitive Surgical Operations, Inc. | Systems and Methods for Intraoperative Segmentation |
DE102016202512B3 (en) | 2016-02-18 | 2017-08-17 | Friedrich-Alexander-Universität Erlangen-Nürnberg | Method for 2D-3D registration, computing device and computer program |
DE102017216953A1 (en) | 2017-09-25 | 2019-03-28 | Siemens Healthcare Gmbh | Display of a virtual object associated with a real object in an augmented reality display device |
-
2018
- 2018-09-19 EP EP18195330.8A patent/EP3626176B1/en active Active
-
2019
- 2019-07-10 JP JP2019128017A patent/JP6952740B2/en active Active
- 2019-09-16 CN CN201910868835.1A patent/CN110916702B/en active Active
- 2019-09-16 US US16/572,414 patent/US11576557B2/en active Active
Patent Citations (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2008120136A1 (en) | 2007-03-30 | 2008-10-09 | Koninklijke Philips Electronics N.V. | 2d/3d image registration |
JP2009056238A (en) | 2007-09-03 | 2009-03-19 | Olympus Medical Systems Corp | Endoscope apparatus |
US20090088773A1 (en) * | 2007-09-30 | 2009-04-02 | Intuitive Surgical, Inc. | Methods of locating and tracking robotic instruments in robotic surgical systems |
US20090088632A1 (en) * | 2007-10-02 | 2009-04-02 | Siemens Corporate Research, Inc. | Method for Dynamic Road Mapping |
US20120165655A1 (en) * | 2009-06-26 | 2012-06-28 | Dirk Mucha | Method for generating position data of an instrument |
US20120289825A1 (en) * | 2011-05-11 | 2012-11-15 | Broncus, Technologies, Inc. | Fluoroscopy-based surgical device tracking method and system |
US20150030229A1 (en) | 2013-07-24 | 2015-01-29 | Anja Borsdorf | Methods for Updating 2D/3D Registration on Movement and Computing Device |
US20150085981A1 (en) | 2013-09-24 | 2015-03-26 | Siemens Aktiengesellschaft | Method of image registration in a multi-source/single detector radiographic imaging system, and image acquisition apparatus |
JP2017507708A (en) | 2014-02-11 | 2017-03-23 | コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. | Spatial visualization of the internal thoracic artery during minimally invasive bypass surgery |
US20170172663A1 (en) | 2014-02-11 | 2017-06-22 | Koninklijke Philips N.V. | Spatial visualization of internal mammary artery during minimally invasive bypass surgery |
US20170209071A1 (en) | 2014-07-28 | 2017-07-27 | Intuitive Surgical Operations, Inc. | Systems and Methods for Intraoperative Segmentation |
JP2017523836A (en) | 2014-07-28 | 2017-08-24 | インテュイティブ サージカル オペレーションズ, インコーポレイテッド | System and method for segmentation during surgery |
US20160260220A1 (en) | 2015-03-05 | 2016-09-08 | Broncus Medical Inc. | Gpu-based system for performing 2d-3d deformable registration of a body organ using multiple 2d fluoroscopic views |
WO2016182550A1 (en) | 2015-05-11 | 2016-11-17 | Siemens Aktiengesellschaft | Method and system for registration of 2d/2.5d laparoscopic and endoscopic image data to 3d volumetric image data |
US20180150929A1 (en) | 2015-05-11 | 2018-05-31 | Siemens Aktiengesellschaft | Method and system for registration of 2d/2.5d laparoscopic and endoscopic image data to 3d volumetric image data |
JP2017086819A (en) | 2015-11-17 | 2017-05-25 | 東芝メディカルシステムズ株式会社 | Medical image diagnostic apparatus |
US20170165008A1 (en) | 2015-12-14 | 2017-06-15 | Nuvasive, Inc. | 3D Visualization During Surgery with Reduced Radiation Exposure |
DE102016202512B3 (en) | 2016-02-18 | 2017-08-17 | Friedrich-Alexander-Universität Erlangen-Nürnberg | Method for 2D-3D registration, computing device and computer program |
US20170243361A1 (en) | 2016-02-18 | 2017-08-24 | Roman Schaffert | 2D/3D Registration |
DE102017216953A1 (en) | 2017-09-25 | 2019-03-28 | Siemens Healthcare Gmbh | Display of a virtual object associated with a real object in an augmented reality display device |
US20190096084A1 (en) | 2017-09-25 | 2019-03-28 | Siemens Aktiengesellschaft | Display of a virtual object linked to a real object in an augmented reality display device |
Non-Patent Citations (5)
Title |
---|
Communication under Rule 71(3) EPC for European Application No. 18 195 330.8-1122 dated Aug. 27, 2020. |
European Decision to Grant for European Application No. 18195330.8 -1122 / 3626176 dated Dec. 3, 2020. |
European Search Report for European Application No. 18195330.8-1124 dated Mar. 28, 2019. |
European Search Report for European Patent Application No. 18195330.8-1124 dated Mar. 28, 2019. |
Japanese Office Action for Japanese Application No. 2019-128017 dated Apr. 6, 2021, with English translation. |
Also Published As
Publication number | Publication date |
---|---|
JP2020058779A (en) | 2020-04-16 |
EP3626176B1 (en) | 2020-12-30 |
EP3626176A1 (en) | 2020-03-25 |
CN110916702A (en) | 2020-03-27 |
JP6952740B2 (en) | 2021-10-20 |
CN110916702B (en) | 2022-09-02 |
US20200085281A1 (en) | 2020-03-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8165660B2 (en) | System and method for selecting a guidance mode for performing a percutaneous procedure | |
US7734329B2 (en) | Method for pre-interventional planning of a 2D fluoroscopy projection | |
US10130430B2 (en) | No-touch surgical navigation method and system thereof | |
EP2967297B1 (en) | System for dynamic validation, correction of registration for surgical navigation | |
RU2445007C2 (en) | Superposition of coordinate systems | |
US11576746B2 (en) | Light and shadow guided needle positioning system and method | |
CN107106241B (en) | System for navigating to surgical instruments | |
US8374678B2 (en) | Medical apparatus with image acquisition device and position determination device combined in the medical apparatus | |
US10482614B2 (en) | Method and system for registration verification | |
US8886286B2 (en) | Determining and verifying the coordinate transformation between an X-ray system and a surgery navigation system | |
JP7381474B2 (en) | Medical guidance system and method using local planes | |
US20110071380A1 (en) | Manual Instrumented Medical Tool System | |
JP6559532B2 (en) | Real-time simulation of fluoroscopic images | |
US20080234575A1 (en) | Method and device for making correction information available | |
CN111970986A (en) | System and method for performing intraoperative guidance | |
JP2008126075A (en) | System and method for visual verification of ct registration and feedback | |
KR101954868B1 (en) | Navigation system for vascular intervention and method for generaing virtual x-ray image | |
KR20190078853A (en) | Laser projection apparatus and control method thereof, laser guidance system including the apparatus | |
US11576557B2 (en) | Method for supporting a user, computer program product, data medium and imaging system | |
JP4510415B2 (en) | Computer-aided display method for 3D objects | |
US11950950B2 (en) | Zoom detection and fluoroscope movement detection for target overlay | |
CN106175931B (en) | Marking of a fluoroscopic field of view | |
TWI501749B (en) | Instrument guiding method of surgical navigation system | |
US20230190377A1 (en) | Technique Of Determining A Scan Region To Be Imaged By A Medical Image Acquisition Device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
AS | Assignment |
Owner name: SIEMENS HEALTHCARE GMBH, GERMANY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:REGENSBURGER, ALOIS;REEL/FRAME:052284/0600 Effective date: 20191122 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: SIEMENS HEALTHINEERS AG, GERMANY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SIEMENS HEALTHCARE GMBH;REEL/FRAME:066267/0346 Effective date: 20231219 |