CN111000631B - Endoscope simulation method and system based on Unity3D volume rendering - Google Patents

Endoscope simulation method and system based on Unity3D volume rendering Download PDF

Info

Publication number
CN111000631B
CN111000631B CN201911303274.7A CN201911303274A CN111000631B CN 111000631 B CN111000631 B CN 111000631B CN 201911303274 A CN201911303274 A CN 201911303274A CN 111000631 B CN111000631 B CN 111000631B
Authority
CN
China
Prior art keywords
endoscope
model
label
mouse
camera
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911303274.7A
Other languages
Chinese (zh)
Other versions
CN111000631A (en
Inventor
肖建如
吕天予
周振华
马科威
刘铁龙
邵帅
曹佳实
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Jiaao Information Technology Development Co ltd
Shanghai Changzheng Hospital
Original Assignee
Shanghai Jiaao Information Technology Development Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Jiaao Information Technology Development Co ltd filed Critical Shanghai Jiaao Information Technology Development Co ltd
Priority to CN201911303274.7A priority Critical patent/CN111000631B/en
Publication of CN111000631A publication Critical patent/CN111000631A/en
Application granted granted Critical
Publication of CN111000631B publication Critical patent/CN111000631B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/10Computer-aided planning, simulation or modelling of surgical operations
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/10Computer-aided planning, simulation or modelling of surgical operations
    • A61B2034/101Computer-aided simulation of surgical operations
    • A61B2034/105Modelling of the patient, e.g. for ligaments or bones
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/10Computer-aided planning, simulation or modelling of surgical operations
    • A61B2034/107Visualisation of planned trajectories or target regions

Landscapes

  • Health & Medical Sciences (AREA)
  • Surgery (AREA)
  • Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Robotics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Medical Informatics (AREA)
  • Molecular Biology (AREA)
  • Animal Behavior & Ethology (AREA)
  • General Health & Medical Sciences (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Endoscopes (AREA)

Abstract

The invention provides an endoscope simulation method and system based on Unity3D body rendering, which are used for constructing a body model or a face model of a natural duct or an operation incision of a human body according to acquired medical data, observing in the body model or the face model by adopting a camera and simulating the use of an endoscope in the natural duct or the operation incision of the human body. Through the technology of observing the state of an illness by using an endoscope in computer simulation operation, a doctor can directly observe the internal condition of a human body under the condition of not making an incision, the accuracy and the safety of the operation are further improved, and the operation time is shortened.

Description

Endoscope simulation method and system based on Unity3D volume rendering
Technical Field
The invention relates to the technical field of computers, in particular to an endoscope simulation method and system based on Unity3D volume rendering.
Background
A conventional endoscope is disclosed in patent publication No. CN104968254B, and includes: a distal end portion for insertion into a subject; a light guide for transmitting light to the tip portion; an image guide for transmitting image information acquired in the tip portion; a holding frame made of a material having a shear stress larger than that of the distal end portion, the holding frame being adhesively fixed in the distal end portion in a state where one end surface of the holding frame is exposed from the distal end surface of the distal end portion, and holding the light guide and the distal end side of the image guide; and at least one groove provided in an insertion direction of the light guide and the image guide on an outer peripheral surface of the holding frame, and having one end exposed to the distal end surface side so as to be applicable along the at least one groove when a notch is applied to the distal end portion.
Meanwhile, the realization of three-dimensional imaging through technologies such as CT, nuclear magnetic resonance and the like is also a known technology in the field of medical equipment at present.
With the development of endoscope technology, the endoscope technology is applied to more and more operations, and although the endoscope technology can reduce the invasiveness of the operations and increase the accuracy and safety of the operations, an incision is inevitably made on a human body, an endoscope is put into the human body, the condition of a patient is observed, and the endoscope technology has a limit in use time. Therefore, the use of the endoscope requires high expertise and skill of the doctor.
Disclosure of Invention
Aiming at the defects in the prior art, the invention aims to provide an endoscope simulation method and system based on Unity3D volume rendering.
According to the endoscope simulation method based on the Unity3D body rendering, a body model or a face model is constructed for the natural pore canal or the surgical incision of the human body according to the acquired medical data, a camera is used for observing in the body model or the face model, and the use of the endoscope in the natural pore canal or the surgical incision of the human body is simulated.
Preferably, the Unity3D volume rendering-based endoscope simulation method comprises the following steps:
when a user clicks the record for the first time, the position P1 of the current camera, the current rotation angle R1 and the orientation Ry of the y coordinate axis are recorded, and only the position of the current camera is recorded at the subsequent click, wherein the positions are recorded as P2, P3 and P4.
Preferably, the Unity3D volume rendering-based endoscope simulation method comprises the following steps:
automatically playing an observation path:
calculating a rotation angle F1 from P1 to P2, and calculating a quaternion Q1, wherein a forward vector of Q1 is F1, an up vector of Q1 is Ry, the quaternion Q1 is converted into a rotation angle R2, and an angle A1 required for rotating R1 to R2 is recorded;
calculating the distance L1 from P1 to P2;
and calculating whether the camera currently stays at the recorded position for angle rotation or position movement according to the playing progress.
Preferably, the Unity3D volume rendering-based endoscope simulation method comprises the following steps:
inserting a label: after the label is added through clicking, an icon with the label at the position of the mouse moves along with the mouse, when the mouse is clicked in the viewport, ray detection is transmitted into a screen through an input point detected by a current camera, if a body model or a face model is detected through rays, the label is placed on the detected point, and if the model is not detected through the rays or the clicking position of the mouse is not in the viewport, the placement of the label is considered to be cancelled.
According to the endoscope simulation system based on the Unity3D body rendering, a body model or a face model is constructed for a natural duct or an operation incision of a human body according to collected medical data, a camera is used for observing in the body model or the face model, and the use of an endoscope in the natural duct or the operation incision of the human body is simulated.
Preferably, the method comprises the following steps:
when a user clicks the record for the first time, the position P1 of the current camera, the current rotation angle R1 and the orientation Ry of the y coordinate axis are recorded, and only the position of the current camera is recorded at the subsequent click, wherein the positions are recorded as P2, P3 and P4.
Preferably, the method further comprises the following steps:
automatically playing an observation path:
calculating a rotation angle F1 from P1 to P2, and calculating a quaternion Q1, wherein a forward vector of Q1 is F1, an up vector of Q1 is Ry, the quaternion Q1 is converted into a rotation angle R2, and an angle A1 required for rotating R1 to R2 is recorded;
calculating the distance L1 from P1 to P2;
and calculating whether the camera currently stays at the recorded position for angle rotation or position movement according to the playing progress.
Preferably, the method further comprises the following steps:
inserting a label: after the label is added through clicking, an icon with the label at the position of the mouse moves along with the mouse, when the mouse is clicked in the viewport, ray detection is transmitted into a screen through an input point detected by a current camera, if a body model or a face model is detected through rays, the label is placed on the detected point, and if the model is not detected through the rays or the clicking position of the mouse is not in the viewport, the placement of the label is considered to be cancelled.
Compared with the prior art, the invention has the following beneficial effects:
through the technology of observing the state of an illness by using an endoscope in computer simulation operation, a doctor can directly observe the internal condition of a human body under the condition of not making an incision, the accuracy and the safety of the operation are further improved, and the operation time is shortened.
Drawings
Other features, objects and advantages of the invention will become more apparent upon reading of the detailed description of non-limiting embodiments with reference to the following drawings:
FIG. 1 is a schematic diagram of an embodiment of the present invention.
Detailed Description
The present invention will be described in detail with reference to specific examples. The following examples will assist those skilled in the art in further understanding the invention, but are not intended to limit the invention in any way. It should be noted that it would be obvious to those skilled in the art that various changes and modifications can be made without departing from the spirit of the invention. All falling within the scope of the present invention.
According to the endoscope simulation method based on the Unity3D body rendering, a body model or a face model is constructed for the natural pore canal or the surgical incision of the human body according to the acquired medical data, a camera is used for observing in the body model or the face model, and the use of the endoscope in the natural pore canal or the surgical incision of the human body is simulated.
The endoscope simulation method based on Unity3D volume rendering comprises the following steps:
when a user clicks the record for the first time, the position P1 of the current camera, the current rotation angle R1 and the orientation Ry of the y coordinate axis are recorded, and only the position of the current camera is recorded at the subsequent click, wherein the positions are recorded as P2, P3 and P4.
The endoscope simulation method based on Unity3D volume rendering comprises the following steps:
automatically playing an observation path:
calculating a rotation angle F1 from P1 to P2, and calculating a quaternion Q1, wherein a forward vector of Q1 is F1, an up vector of Q1 is Ry, the quaternion Q1 is converted into a rotation angle R2, and an angle A1 required for rotating R1 to R2 is recorded;
calculating the distance L1 from P1 to P2;
and calculating whether the camera currently stays at the recorded position for angle rotation or position movement according to the playing progress.
The endoscope simulation method based on Unity3D volume rendering comprises the following steps:
inserting a label: after the label is added through clicking, an icon with the label at the mouse position moves along with the mouse, when the mouse is clicked in the viewport, ray detection is transmitted into a screen through an input point detected by a current camera, if a body model or a face model is detected through rays, the label is placed on the detected point, and as shown in fig. 1, if the model is not detected through the rays or the mouse click position is not in the viewport, the placement of the label is considered to be cancelled.
On the basis of the Unity3D body rendering-based endoscope simulation method, the invention also provides a Unity3D body rendering-based endoscope simulation system, which constructs a body model or a face model for the natural duct or the surgical incision of the human body according to the acquired medical data, adopts a camera to observe in the body model or the face model, and simulates the use of an endoscope in the natural duct or the surgical incision of the human body.
Those skilled in the art will appreciate that, in addition to implementing the system and its various devices, modules, units provided by the present invention as pure computer readable program code, the system and its various devices, modules, units provided by the present invention can be fully implemented by logically programming method steps in the form of logic gates, switches, application specific integrated circuits, programmable logic controllers, embedded microcontrollers and the like. Therefore, the system and various devices, modules and units thereof provided by the invention can be regarded as a hardware component, and the devices, modules and units included in the system for realizing various functions can also be regarded as structures in the hardware component; means, modules, units for performing the various functions may also be regarded as structures within both software modules and hardware components for performing the method.
The foregoing description of specific embodiments of the present invention has been presented. It is to be understood that the present invention is not limited to the specific embodiments described above, and that various changes or modifications may be made by one skilled in the art within the scope of the appended claims without departing from the spirit of the invention. The embodiments and features of the embodiments of the present application may be combined with each other arbitrarily without conflict.

Claims (1)

1. An endoscope simulation system based on Unity3D body rendering is characterized in that a body model or a face model is constructed for a natural duct or an operation incision of a human body according to acquired medical data, a camera is used for observing in the body model or the face model, and the use of an endoscope in the natural duct or the operation incision of the human body is simulated;
further comprising inserting a label: after the label is added through clicking, an icon with the label at the position of the mouse moves along with the mouse, when the mouse is clicked in a viewport, ray detection is transmitted into a screen through an input point detected by a current camera, if a body model or a face model is detected through rays, the label is placed on the detected point, and if the model is not detected through the rays or the clicking position of the mouse is not in the viewport, the placement of the label is considered to be cancelled;
when a user clicks a record for the first time, recording the position P1 of the current camera, the current rotation angle R1 and the orientation Ry of a y coordinate axis, and recording the position of the current camera only when clicking the record for the subsequent time, wherein the position is recorded as P2, P3 and P4.. the.;
further comprising automatically playing the observation path:
calculating a rotation angle F1 from P1 to P2, and calculating a quaternion Q1, wherein a forward vector of Q1 is F1, an up vector of Q1 is Ry, the quaternion Q1 is converted into a rotation angle R2, and an angle A1 required for rotating R1 to R2 is recorded;
calculating the distance L1 from P1 to P2;
and calculating whether the camera currently stays at the recorded position for angle rotation or position movement according to the playing progress.
CN201911303274.7A 2019-12-17 2019-12-17 Endoscope simulation method and system based on Unity3D volume rendering Active CN111000631B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911303274.7A CN111000631B (en) 2019-12-17 2019-12-17 Endoscope simulation method and system based on Unity3D volume rendering

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911303274.7A CN111000631B (en) 2019-12-17 2019-12-17 Endoscope simulation method and system based on Unity3D volume rendering

Publications (2)

Publication Number Publication Date
CN111000631A CN111000631A (en) 2020-04-14
CN111000631B true CN111000631B (en) 2021-04-06

Family

ID=70116380

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911303274.7A Active CN111000631B (en) 2019-12-17 2019-12-17 Endoscope simulation method and system based on Unity3D volume rendering

Country Status (1)

Country Link
CN (1) CN111000631B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022091210A1 (en) * 2020-10-27 2022-05-05 リバーフィールド株式会社 Surgery assisting device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE19543410A1 (en) * 1994-11-23 1996-06-27 Gen Electric Non=invasive method for investigation of body cavities
CN101849843A (en) * 2009-03-31 2010-10-06 上海交通大学医学院附属新华医院 Navigation method of three-dimensional cardiac ultrasonic virtual endoscope
CN103049934A (en) * 2012-12-13 2013-04-17 航天科工仿真技术有限责任公司 Roam mode realizing method in three-dimensional scene simulation system
CN103279984A (en) * 2013-06-04 2013-09-04 浙江工业大学 Sight glass visual angle tracking method based on image transformation matrix
CN106901834A (en) * 2016-12-29 2017-06-30 陕西联邦义齿有限公司 The preoperative planning of minimally invasive cardiac surgery and operation virtual reality simulation method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE19543410A1 (en) * 1994-11-23 1996-06-27 Gen Electric Non=invasive method for investigation of body cavities
CN101849843A (en) * 2009-03-31 2010-10-06 上海交通大学医学院附属新华医院 Navigation method of three-dimensional cardiac ultrasonic virtual endoscope
CN103049934A (en) * 2012-12-13 2013-04-17 航天科工仿真技术有限责任公司 Roam mode realizing method in three-dimensional scene simulation system
CN103279984A (en) * 2013-06-04 2013-09-04 浙江工业大学 Sight glass visual angle tracking method based on image transformation matrix
CN106901834A (en) * 2016-12-29 2017-06-30 陕西联邦义齿有限公司 The preoperative planning of minimally invasive cardiac surgery and operation virtual reality simulation method

Also Published As

Publication number Publication date
CN111000631A (en) 2020-04-14

Similar Documents

Publication Publication Date Title
US11730545B2 (en) System and method for multi-client deployment of augmented reality instrument tracking
KR102013837B1 (en) Method and apparatus for providing surgical video
CN106251751B (en) A kind of simulated medical surgery analogue system based on VR technology
US10580325B2 (en) System and method for performing a computerized simulation of a medical procedure
US20200105070A1 (en) Overlay and Manipulation of Medical Images in a Virtual Environment
US20140320600A1 (en) System and method for tracking non-visible structure of a body
JP5336370B2 (en) An image context-dependent application related to anatomical structures for efficient diagnosis
CN104883975A (en) Real-time scene-modeling combining 3d ultrasound and 2d x-ray imagery
CN101310249A (en) Systems and methods for editing a model of a physical system for a simulation
CN104274247A (en) Medical surgical navigation method
US11322048B2 (en) Ultrasound-guided medical tool insertion simulators
JP7432534B2 (en) Methods and apparatus for efficiently rendering, managing, recording, and playing back interactive, multi-user virtual reality experiences
CN111613122A (en) Virtual-actual fused vascular interventional operation simulation system
Wright et al. Design and evaluation of an augmented reality simulator using leap motion
US20140343405A1 (en) System and method for tracking non-visible structures of bodies relative to each other
CN111000631B (en) Endoscope simulation method and system based on Unity3D volume rendering
Chen et al. A naked eye 3D display and interaction system for medical education and training
JP6276529B2 (en) Interpretation support apparatus, method and program
Beyar et al. Robotics, imaging, and artificial intelligence in the catheterisation laboratory: Cath lab robotics, imaging, and AI
Zhang et al. [Retracted] Study on Collision Detection and Force Feedback Algorithm in Virtual Surgery
Franz et al. Simplified development of image-guided therapy software with MITK-IGT
US20240029367A1 (en) Extended reality training platform
EP2916763A1 (en) Surgical location monitoring system and method using natural markers
CN116086462A (en) Track data processing method, device, medium and computing equipment
CN113397708B (en) Particle puncture surgical robot navigation system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20230406

Address after: Shanghai Changzheng Hospital, 415 Fengyang Road, Huangpu District, Shanghai 200001

Patentee after: SHANGHAI CHANGZHENG Hospital

Patentee after: SHANGHAI JIAAO INFORMATION TECHNOLOGY DEVELOPMENT Co.,Ltd.

Address before: 201304 1-2, 4, 1628 Li Zheng Road, Pudong New Area academy, Shanghai.

Patentee before: SHANGHAI JIAAO INFORMATION TECHNOLOGY DEVELOPMENT Co.,Ltd.