US20240079125A1 - Noninvasive diagnostic system - Google Patents
Noninvasive diagnostic system Download PDFInfo
- Publication number
- US20240079125A1 US20240079125A1 US18/238,338 US202318238338A US2024079125A1 US 20240079125 A1 US20240079125 A1 US 20240079125A1 US 202318238338 A US202318238338 A US 202318238338A US 2024079125 A1 US2024079125 A1 US 2024079125A1
- Authority
- US
- United States
- Prior art keywords
- bone
- patient
- virtual
- model
- deformable
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 210000000988 bone and bone Anatomy 0.000 claims abstract description 131
- 230000033001 locomotion Effects 0.000 claims abstract description 60
- 238000000034 method Methods 0.000 claims abstract description 44
- 238000002604 ultrasonography Methods 0.000 claims description 42
- 210000000629 knee joint Anatomy 0.000 description 32
- 238000013528 artificial neural network Methods 0.000 description 21
- 230000006378 damage Effects 0.000 description 16
- 208000027418 Wounds and injury Diseases 0.000 description 15
- 208000014674 injury Diseases 0.000 description 15
- 210000003127 knee Anatomy 0.000 description 15
- 238000013459 approach Methods 0.000 description 13
- 210000002303 tibia Anatomy 0.000 description 12
- 230000000694 effects Effects 0.000 description 10
- 210000000689 upper leg Anatomy 0.000 description 10
- 238000003745 diagnosis Methods 0.000 description 9
- 210000004417 patella Anatomy 0.000 description 9
- 238000001514 detection method Methods 0.000 description 8
- 210000003041 ligament Anatomy 0.000 description 8
- 239000000523 sample Substances 0.000 description 7
- 230000005856 abnormality Effects 0.000 description 6
- 238000002594 fluoroscopy Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 5
- 238000005259 measurement Methods 0.000 description 5
- 210000004872 soft tissue Anatomy 0.000 description 5
- 238000013519 translation Methods 0.000 description 5
- 239000013598 vector Substances 0.000 description 5
- 230000001133 acceleration Effects 0.000 description 4
- 238000004458 analytical method Methods 0.000 description 4
- 210000001264 anterior cruciate ligament Anatomy 0.000 description 4
- 230000008901 benefit Effects 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- 210000001519 tissue Anatomy 0.000 description 4
- 238000012549 training Methods 0.000 description 4
- 206010060820 Joint injury Diseases 0.000 description 3
- 230000002159 abnormal effect Effects 0.000 description 3
- 230000007812 deficiency Effects 0.000 description 3
- 230000007850 degeneration Effects 0.000 description 3
- 238000013461 design Methods 0.000 description 3
- 210000002683 foot Anatomy 0.000 description 3
- 230000005484 gravity Effects 0.000 description 3
- 230000004044 response Effects 0.000 description 3
- 239000010409 thin film Substances 0.000 description 3
- 210000000845 cartilage Anatomy 0.000 description 2
- 210000004439 collateral ligament Anatomy 0.000 description 2
- 230000003750 conditioning effect Effects 0.000 description 2
- 239000007943 implant Substances 0.000 description 2
- 230000005499 meniscus Effects 0.000 description 2
- 210000003205 muscle Anatomy 0.000 description 2
- 230000004962 physiological condition Effects 0.000 description 2
- 210000002967 posterior cruciate ligament Anatomy 0.000 description 2
- 238000001356 surgical procedure Methods 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 238000012800 visualization Methods 0.000 description 2
- 208000025674 Anterior Cruciate Ligament injury Diseases 0.000 description 1
- 241000282412 Homo Species 0.000 description 1
- 208000012659 Joint disease Diseases 0.000 description 1
- 241001465754 Metazoa Species 0.000 description 1
- 206010061363 Skeletal injury Diseases 0.000 description 1
- 208000026137 Soft tissue injury Diseases 0.000 description 1
- 210000003423 ankle Anatomy 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000001143 conditioned effect Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 239000013078 crystal Substances 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 230000002950 deficient Effects 0.000 description 1
- 210000001513 elbow Anatomy 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000005284 excitation Effects 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 210000003811 finger Anatomy 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000012010 growth Effects 0.000 description 1
- 210000001624 hip Anatomy 0.000 description 1
- 238000007654 immersion Methods 0.000 description 1
- 238000012623 in vivo measurement Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000005865 ionizing radiation Effects 0.000 description 1
- 230000009191 jumping Effects 0.000 description 1
- 210000002414 leg Anatomy 0.000 description 1
- 238000002595 magnetic resonance imaging Methods 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000003349 osteoarthritic effect Effects 0.000 description 1
- 230000001575 pathological effect Effects 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 230000002035 prolonged effect Effects 0.000 description 1
- 230000001902 propagating effect Effects 0.000 description 1
- 230000005855 radiation Effects 0.000 description 1
- 230000002040 relaxant effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 238000000926 separation method Methods 0.000 description 1
- 210000002832 shoulder Anatomy 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 208000024891 symptom Diseases 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 210000001738 temporomandibular joint Anatomy 0.000 description 1
- 230000000451 tissue damage Effects 0.000 description 1
- 231100000827 tissue damage Toxicity 0.000 description 1
- 208000037816 tissue injury Diseases 0.000 description 1
- 210000003371 toe Anatomy 0.000 description 1
- 238000012876 topography Methods 0.000 description 1
- 210000000707 wrist Anatomy 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H40/00—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
- G16H40/60—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
- G16H40/63—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for local operation
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0002—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
- A61B5/0004—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network characterised by the type of physiological signal transmitted
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0002—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
- A61B5/0015—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network characterised by features of the telemetry system
- A61B5/002—Monitoring the patient using a local or closed circuit, e.g. in a room or building
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0059—Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/1036—Measuring load distribution, e.g. podologic studies
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/1036—Measuring load distribution, e.g. podologic studies
- A61B5/1038—Measuring plantar pressure during gait
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/1113—Local tracking of patients, e.g. in a hospital or private home
- A61B5/1114—Tracking parts of the body
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/1121—Determining geometric values, e.g. centre of rotation or angular range of movement
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/1126—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb using a particular sensing technique
- A61B5/1127—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb using a particular sensing technique using markers
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/45—For evaluating or diagnosing the musculoskeletal system or teeth
- A61B5/4528—Joints
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/45—For evaluating or diagnosing the musculoskeletal system or teeth
- A61B5/4533—Ligaments
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/45—For evaluating or diagnosing the musculoskeletal system or teeth
- A61B5/4538—Evaluating a particular part of the muscoloskeletal system or a particular medical condition
- A61B5/4585—Evaluating the knee
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/68—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
- A61B5/6801—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
- A61B5/6802—Sensor mounted on worn items
- A61B5/6804—Garments; Clothes
- A61B5/6807—Footwear
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/68—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
- A61B5/6801—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
- A61B5/6813—Specially adapted to be attached to a specific body part
- A61B5/6828—Leg
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/68—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
- A61B5/6801—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
- A61B5/683—Means for maintaining contact with the body
- A61B5/6831—Straps, bands or harnesses
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/68—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
- A61B5/6801—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
- A61B5/683—Means for maintaining contact with the body
- A61B5/6832—Means for maintaining contact with the body using adhesives
- A61B5/6833—Adhesive patches
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7264—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
- A61B5/7267—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems involving training the classification device
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/74—Details of notification to user or communication with user or patient ; user input means
- A61B5/742—Details of notification to user or communication with user or patient ; user input means using visual displays
- A61B5/743—Displaying an image simultaneously with additional graphical information, e.g. symbols, charts, function plots
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/74—Details of notification to user or communication with user or patient ; user input means
- A61B5/742—Details of notification to user or communication with user or patient ; user input means using visual displays
- A61B5/7435—Displaying user selection data, e.g. icons in a graphical user interface
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/08—Detecting organic movements or changes, e.g. tumours, cysts, swellings
- A61B8/0858—Detecting organic movements or changes, e.g. tumours, cysts, swellings involving measuring tissue layers, e.g. skin, interfaces
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/08—Detecting organic movements or changes, e.g. tumours, cysts, swellings
- A61B8/0875—Detecting organic movements or changes, e.g. tumours, cysts, swellings for diagnosis of bone
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/42—Details of probe positioning or probe attachment to the patient
- A61B8/4209—Details of probe positioning or probe attachment to the patient by using holders, e.g. positioning frames
- A61B8/4227—Details of probe positioning or probe attachment to the patient by using holders, e.g. positioning frames characterised by straps, belts, cuffs or braces
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/42—Details of probe positioning or probe attachment to the patient
- A61B8/4245—Details of probe positioning or probe attachment to the patient involving determining the position of the probe, e.g. with respect to an external reference frame or to the patient
- A61B8/4254—Details of probe positioning or probe attachment to the patient involving determining the position of the probe, e.g. with respect to an external reference frame or to the patient using sensors mounted on the probe
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/42—Details of probe positioning or probe attachment to the patient
- A61B8/4245—Details of probe positioning or probe attachment to the patient involving determining the position of the probe, e.g. with respect to an external reference frame or to the patient
- A61B8/4263—Details of probe positioning or probe attachment to the patient involving determining the position of the probe, e.g. with respect to an external reference frame or to the patient using sensors not mounted on the probe, e.g. mounted on an external reference frame
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/52—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/5215—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data
- A61B8/5223—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data for extracting a diagnostic or physiological parameter from medical diagnostic data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/50—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for simulation or modelling of medical disorders
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
- A61B2034/101—Computer-aided simulation of surgical operations
- A61B2034/105—Modelling of the patient, e.g. for ligaments or bones
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
- A61B2034/2046—Tracking techniques
- A61B2034/2048—Tracking techniques using an accelerometer or inertia sensor
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
- A61B2034/2046—Tracking techniques
- A61B2034/2051—Electromagnetic tracking systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2562/00—Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
- A61B2562/02—Details of sensors specially adapted for in-vivo measurements
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2562/00—Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
- A61B2562/02—Details of sensors specially adapted for in-vivo measurements
- A61B2562/0204—Acoustic sensors
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2562/00—Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
- A61B2562/02—Details of sensors specially adapted for in-vivo measurements
- A61B2562/0219—Inertial sensors, e.g. accelerometers, gyroscopes, tilt switches
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2562/00—Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
- A61B2562/02—Details of sensors specially adapted for in-vivo measurements
- A61B2562/0247—Pressure sensors
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2562/00—Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
- A61B2562/04—Arrangements of multiple sensors of the same type
- A61B2562/046—Arrangements of multiple sensors of the same type in a matrix array
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0002—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/112—Gait analysis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
- A61B5/316—Modalities, i.e. specific diagnostic methods
- A61B5/389—Electromyography [EMG]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/42—Details of probe positioning or probe attachment to the patient
- A61B8/4209—Details of probe positioning or probe attachment to the patient by using holders, e.g. positioning frames
- A61B8/4236—Details of probe positioning or probe attachment to the patient by using holders, e.g. positioning frames characterised by adhesive patches
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/44—Constructional features of the ultrasonic, sonic or infrasonic diagnostic device
- A61B8/4444—Constructional features of the ultrasonic, sonic or infrasonic diagnostic device related to the probe
- A61B8/4472—Wireless probes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/44—Constructional features of the ultrasonic, sonic or infrasonic diagnostic device
- A61B8/4477—Constructional features of the ultrasonic, sonic or infrasonic diagnostic device using several separate ultrasound transducers or probes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/46—Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
- A61B8/461—Displaying means of special interest
- A61B8/466—Displaying means of special interest adapted to display 3D data
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/46—Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
- A61B8/467—Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient characterised by special input means
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/48—Diagnostic techniques
- A61B8/483—Diagnostic techniques involving the acquisition of a 3D volume of data
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/52—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/5215—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data
- A61B8/5238—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data for combining image data of patient, e.g. merging several images from different acquisition modes into one image
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H20/00—ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance
- G16H20/30—ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance relating to physical therapies or activities, e.g. physiotherapy, acupressure or exercising
Definitions
- the present invention relates to diagnosis of bodily abnormalities, and more particularly, to devices and methods for evaluating the physiological condition of bodily tissue to discern whether abnormalities exist and the extent of any abnormalities. While the exemplary embodiments disclosed herein are utilized and discussed with respect to a human knee joint, it is to be understood that other joints and bodily tissues may be likewise diagnosed.
- the knee joint is functionally controlled by a mechanical system governed by three unique types of forces: (1) active forces resulting in motion, such as those resulting from muscle flexing or relaxing; (2) constraining forces that constrain motion, such as those resulting from ligaments being in tension; and (3) compressive forces that resist motion, such as those acting upon bones.
- active forces resulting in motion such as those resulting from muscle flexing or relaxing
- constraining forces that constrain motion such as those resulting from ligaments being in tension
- compressive forces that resist motion such as those acting upon bones.
- cartilage and meniscus also produce a dampening effect dissipating the compressive forces propagating to other joints.
- Knee joint motions are stabilized primarily by four ligaments, which restrict and regulate the relative motion between the femur, tibia, and patella.
- These ligaments are the anterior cruciate ligament (ACL), the posterior cruciate ligament (PCL), the medial collateral ligament (MCL), and the lateral collateral ligament (LCL), as shown in FIGS. 1 and 2 .
- An injury to any of these ligaments or other soft-tissue structures can cause detectable changes in knee kinematics and the creation of detectable patterns of vibration representative of the type of knee joint injury and the severity of the injury.
- These visual and auditory changes are produced by the bones while moving in a distorted kinematic pattern, and they differ significantly from the look and vibration of a properly balanced knee joint moving through a range of motion.
- the present invention in one embodiment, provides a patient specific 3D model of a patient's joint, including bone and soft tissue. This model is then registered to the patient's actual bone so that as the joint is taken through a range of motion it can be visualized on a computer screen. A physician can then use the computer generated image to make a diagnosis or compare the motion of the actual bone to a database of clinically relevant information on desirable or undesirable joint motion.
- the exemplary embodiments of the present invention include a diagnostic system for mammalian bodies to determine the type of injury and extent of injury using kinematic data and/or vibration data.
- an exemplary method and embodiment are directed to a knee joint diagnostic system for automatically determining the type of injury and the extent to which ligaments, muscles, bones, meniscus, and cartilage may be affected by an injury through analyzing the kinematics of the knee joint, while also analyzing the pattern and special distribution of the vibration produced knee joint movement.
- An exemplary process flow diagram for this exemplary method is shown in FIG. 3 .
- patient-specific 3D models of the distal femur, proximal tibia, and the patella are constructed using pulse echo A-mode ultrasound based 3D model reconstruction technology.
- patient-specific kinematic data is obtained for the motions of the femur, tibia, and patella using pulse A-mode ultrasound.
- patient specific vibration data is obtained while the knee joint is taken through a range of motion and loaded in real-world conditions.
- the vibration data and kinematic data are taken at the same time using the single data acquisition device.
- the data is displayed in real-time on a split screen monitor.
- a recording device and memory may be utilized to record the data in a time synched manner.
- the patient may be given an actuator that is operative to note the general time frame within which the patient felt a particular pain or severe pain to allow a correlation between pain felt by the patient and the kinematics and vibration occurring at roughly the same time.
- Patient-specific data is analyzed by a trained neural network in order to provide an automated output as to the existence of an injury, the type of injury, and the severity of the injury.
- This neural network may be accessible via the internet or may reside on a physician's local computer.
- patient-specific data may be analyzed by a physician to make the diagnosis directly without the aid of the neural network.
- a physician may diagnose a bodily injury without requiring experimental surgery or requiring exposure of the patient to radiation from still X-rays or fluoroscopy.
- the data taken regarding each patient is continuous through a range of motion, in contrast to X-rays and fluoroscopy which take images at distinct points with significant range of motion gaps.
- data taken in accordance with the exemplary method and devices disclosed herein also contrasts data taken by a magnetic resonance imaging machine, not only because the data taken is continuous along the range of motion, but also because the bodily portion evaluated is acting under loaded conditions in a dynamic environment.
- Another object of the present invention is to provide a method of tracking the motion of a patient's actual bone through space and showing the same on a computer screen.
- Yet another object of the present invention is to provide a method of tracking at least two bones relative to one another as three dimensional models on a computer screen as the actual bones are taken through a range of motion.
- FIG. 1 is a posterior view of a human knee joint in a fixed position
- FIG. 2 is a posterior view of a human knee joint in an extended position
- FIG. 3 is an exemplary process flow diagram using exemplary methods within the scope of the present invention.
- FIG. 4 is a schematic diagram of the modules of an exemplary diagnostic system
- FIG. 5 is a screen shot of a software user interface for bone modeling
- FIG. 6 is an anterior view of the bones of a human knee joint in an extended position
- FIG. 7 is a pictorial representation of a human leg having an exemplary brace attached to a distal segment of the femur, and exemplary brace attached to a proximal segment of the tibia, a sensor mounted proximate the patella, and a foot pressure sensing shoe;
- FIG. 8 is a is an illustration of a CT slice of the transcutaneous detection of a bone's surface using pulse echo A-mode ultrasound;
- FIG. 9 is a schematic of an exemplary inertia-based localizer circuit
- FIG. 10 is a schematic of an exemplary brace circuit architecture
- FIG. 11 is a circuit schematic of an exemplary high voltage amplifier
- FIG. 12 is a circuit layout for the exemplary high voltage amplifier of FIG. 11 ;
- FIG. 13 is a block diagram for an exemplary high voltage multiplexer
- FIG. 14 is a block diagram for an exemplary receiving circuit
- FIG. 15 is a pictorial representation of an exemplary kinematics tracking brace
- FIG. 16 is a pictorial representation of an alternative exemplary kinematics tracking brace
- FIG. 17 is a pictorial representation of a further alternative kinematics exemplary tracking brace
- FIG. 18 is a pictorial representation of a vibration detection module
- FIGS. 19 A, 19 B, and 19 C are pictorial representations of exemplary kinematics data, vibration signal, and force data respectively;
- FIG. 20 is a graphical representation showing average ACLD medial and lateral condyle contact positions during a deep knee bend activity
- FIGS. 21 A, 21 B, and 21 C are a series of views showing contact path tracking in accordance with the exemplary embodiments
- FIG. 22 is a is a schematic of the overall classification system flow chart
- FIG. 23 is a schematic representation of an exemplary neural network classifier
- FIG. 24 is an exemplary process flow for training an exemplary neural network
- FIG. 25 is an exemplary process flow for knee deficiency diagnosis using a trained neural network.
- the exemplary embodiments of the present invention are described and illustrated below to encompass diagnosis of bodily abnormalities and, more particularly, devices and methods for evaluating the physiological condition of bodily tissue to discern whether abnormalities exist and the next of any abnormalities.
- the preferred embodiments discussed below are exemplary in nature and may be reconfigured without departing from the scope and spirit of the present invention.
- the exemplary embodiments as discussed below may include optional steps, methods and features that one of ordinary skill should recognize as not being a requisite to fall within the scope of the present invention.
- the embodiments disclosed herein are described with respect to diagnosing a knee joint injury. Nevertheless, the embodiments may be utilized to diagnose other joints and bodily tissue injuries, as the knee joint is merely exemplary to facilitate an understanding of the embodiments disclosed.
- a first exemplary diagnostic system includes four modules: (1) a pulse echo A-mode ultrasound based 3D model reconstruction (PEAUMR) module for constructing 3D patient specific models of the knee joint bones; (2) a joint kinematics tracking (JKT) module for tracking kinematics of the knee joint using the patient-specific 3D model of the knee joint from the PEAUMR module; (3) a vibration detection (VD) module for capturing sounds emanating from the knee joint while in motion; and (4) an intelligent diagnosis (ID) module for identifying pathological cases of the knee joint using kinematic data and associated vibration data gathered during the joint motion.
- PEAUMR pulse echo A-mode ultrasound based 3D model reconstruction
- JKT joint kinematics tracking
- VD vibration detection
- ID intelligent diagnosis
- the system described above is usable with or without the use of the vibration detection module.
- the database could contain mathematical descriptions of healthy or clinically undesirable joint motion.
- the PEAUMR module constructs a 3D model of a subject's (e.g., a patient) bones by transcutaneously acquiring a set of 3D data points that in total are representative of the shape of the bone's surface using a tracked pulse echo A-mode ultrasound probe.
- the probe consists of a single ultrasound transducer attached to a global localizer.
- the global localizer may be optical, inertial, electromagnetic or ultra wide band radio frequency.
- the probe is battery-powered and connected wirelessly to a computer in order to record the set points and construct a unique or patient-specific bone model using an atlas-based deformable model technique.
- the computer includes software that interprets data from the tracked pulse echo A-mode ultrasound probe and is operative to construct the 3D models of the patient's bones, which will look very similar to the model shown in FIG. 6 .
- the patient-specific bone is reconstructed using the set of points collected from the bone's surface transcuateously by the tracked ultrasound probe. These points are then used by the atlas-based deformable model software to reconstruct the 3D model of the patient's bone.
- the software includes a plurality of bone models of the femur, tibia, and patella that are classified, for example, based upon ethnicity, gender, skeletal bone to be modeled, and the side of the body the bone is located. Each of these classifications is accounted for by the dropdown menus of the software so that the model initially chosen by the software most closely approximates the bone of the patient.
- the ultrasound transducer probe is repositioned on the exterior of the skin and data points are generated and applied to the model bone (in this case a distal femur), numerically recorded and viewable in a data window, and ultimately utilized by the software to conform the bone model to the patient's actual bone shape.
- model bone in this case a distal femur
- data points are generated and applied to the model bone (in this case a distal femur), numerically recorded and viewable in a data window, and ultimately utilized by the software to conform the bone model to the patient's actual bone shape.
- a higher number of data points imposed on the model will generally result in a more accurate patient model.
- the JKT module tracks the kinematics of the knee joint using the patient-specific 3D bone models from the PEAUMR module.
- motion tracking of the patient's knee joint bones is performed using one or more bone motion tracking braces.
- the bone motion tracking brace includes pulse echo A-mode ultrasound transducers to transcutaneously localize points on the bones surface.
- the pulse echo A-mode ultrasound transducers may or may not be identical to the pulse echo A-mode ultrasound transducers used by the PEAUMR module.
- Commercially available transducers for use with the exemplary embodiments include, without limitation, the Olympus immersion unfocused 3.5 MHz transducer.
- the force sensing shoe detects the ground reactive pressures simultaneous with knee joint kinematic data acquisition.
- Each ultrasound transducer is tracked using an accelerometer or a sensor-specific localizer (or any other appropriate inertial sensor).
- the resulting localized bone points generated from the outputs of the ultrasound transducers are used in combination with the patient specific 3D bone models to discern bone movement while the knee joint is taken through a range of motion.
- three braces and a foot force sensing shoe are used to track knee joint kinematics and dynamic forces: (a) a first brace is positioned proximate the distal portion of the femur; (b) a second brace is positioned proximate the distal end of the tibia; and, (c) a third brace is positioned proximate the patella region.
- an exemplary bone motion tacking brace includes a plurality of pulse echo A-mode ultrasound transducers for transcutaneous detection of the bone's surface and inertia-based localizers to track the motion of the ultrasound transducers, which in turn, track the bones motion.
- Each brace is wirelessly connected to a computer operative to perform computations and visualization in real-time showing movements of the patient-specific 3D bone models paralleling movements of the patient's actual knee joint in a time synchronized manner.
- Each exemplary brace include a rigid or semi-rigid body having a plurality (two or more) of complementary metal oxide semiconductor (CMOS) inertia-based sensors attached thereto.
- CMOS complementary metal oxide semiconductor
- F r is a summation of all the generalized active forces in the system
- F r* is a summation of all the generalized inertia forces in the system.
- the homogenous transformation between the localizer's reference coordinate frame and the world coordinate frame is calculated using the positions of multiple inertia sensors.
- the previous equations describe the dynamic motion and positioning of a point in 3D Euclidean space. Additionally information is needed to describe a 3D body orientation and motion.
- the orientation of the transducer can be described by using a gravity based accelerometer (example: ADXL-330, analog device) by extracting the tilting information from each pair of orthogonal axis.
- the yaw, pitch and row can be calculated as shown in the following:
- pitch is ⁇ , which is x-axis relative to the ground
- roll is ⁇ , which is y-axis relative to the ground
- row is ⁇ , which is z-axis relative to the ground.
- the accelerometer is based using gravity
- the orientation does not require information from the previous state once the sensor is calibrated.
- the static calibration requires the resultant sum of accelerations from the 3 axis equal to 1 g.
- an orientation sensor that gives us yaw, pitch, and row information of the body are also commercially available (example: IDG-300, Invensense).
- the orientation of the transducer can then be resolved by using direction cosine matrix transformation:
- an accelerometer based localizer is used to track each pulse echo A-mode ultrasound transducer mounted to the brace.
- the localizer comprises a plurality of nodes, with each node comprising a CMOS accelerometer and a temperature sensor for thermal drift comparison. Each node is integrated to minimize noise and distortion.
- the outputs of the accelerometers regarding the X, Y, and Z coordinates and temperature sensor are directed to a multiplexer that multiplexes the signals. Multiplexed outputs are amplified by an amplifier and then directed to an analog-to-digital converter. The digital conversion of the signal can be performed within or outside the CMOS sensors chip. Outputted digital signals are directed to a wireless transmitter by way of a parallel input/serial output device.
- each of the three exemplary design alternatives for the brace has a similar electronic architecture.
- An exemplary electronic architecture includes a high voltage amplifier circuit feeding a voltage multiplexer circuit to excite each ultrasound transducer and thereby acting as an analog switch.
- the echo signals from each transducer are multiplexed pursuant to a logic control directing the opening of the switches in the multiplexer circuit at precise intervals.
- An exemplary logic control is the MSP430 available from Texas Instruments.
- the output from the multiplexer circuit is amplified by an amplifier circuit, signal conditioned using a signal conditioning circuit, and digitized using an analog-to-digital converter. Electric power to the foregoing components is supplied by way of a battery, which also supplies power to a wireless transmitter module.
- the wireless transmitter module utilizes the universal asynchronous receiver/transmitter (UART) protocol.
- the module includes a wireless transmitter circuit receiving the output of the first in-first out (FIFO) buffer of the analog-to-digital converter by way of a serial interface. An output from the wireless transmitter circuit is conveyed using a serial link coupled to an antenna. Signals conveyed through the antenna are broadcast for reception by a wireless receiver coupled to a controller computer.
- UART universal asynchronous receiver/transmitter
- an exemplary high voltage circuit is utilized to trigger and generate the excitation energy for the piezoelectric crystal in the transducer.
- Exemplary high voltage circuits for use in this embodiment include, without limitation, the pulsar integrated circuit (HV379) available from Supertex.
- an exemplary high voltage multiplexer is utilized to trigger and excite multiple piezoelectric transducers without increasing the number of high voltage circuits mentioned with regard to FIG. 11 .
- Exemplary high voltage multiplexers for use in this embodiment include, without limitation, the high voltage multiplexer (HV2221) available from Supertex.
- HV2221 available from Supertex.
- the advantage of using a high voltage multiplexer is the ability to use CMOS level control circuitry, thereby making the control logic compatible with virtually any microcontroller or field programmable gate array commercially available.
- an exemplary receiving circuit which comprises the multiplexer circuit, the amplifier circuit, the signal conditioning circuit, and the analog-to-digital converter, is utilized to receive the echo signals from each transducer.
- Exemplary receiving circuits for use in this embodiment include, without limitation, the AD9271 8-channel ultrasound receiving integrated circuit available from Analog Devices.
- a first exemplary bone tracking brace includes a plurality of transducers mounted thereto. Each transducer is responsible for determining the location of a point on the surface of the bone for each motion tracking frame. Problems of locating and tracking the bone using ultrasound data are reduced as the motion of the bone relative to the skin is small compared to the gross joint motion.
- the first approach commonly referred to herein as the ITT (Individual transducer tracking) approach, involves each transducer in the brace having an inertia-based localizer to individually track each transducer.
- the transducers are held together by flexible length straps.
- a second approach commonly referred to herein as the ITML (Inter-transducers Mechanical Links) approach, involves the transducers being connected to each other by movable mechanical links.
- Each mechanical link includes length and angle sensors that allow for detection of the movement of the transducers relative to one another and the relative translational motions of the links. Every two links are connected by a pivot pin that allows rotation and translation of the links relative to each other.
- An angle sensor is mounted to at least one link proximate the pivot pin to allow for detection of the angle between the links.
- the ITML approach features less localizers than the individual transducer tracking design.
- a third approach commonly referred to herein as the RT (Rotating Transducer) approach, involves using a single ultrasound transducer that is mounted to a carriage.
- the carriage traverses along a track located on the inner circumference of the brace.
- the carriage may be moved along the tack by a string loop that is wrapped around the drive shaft of a motor.
- the transducer reaches the motor, the rotation direction of the motor is changed and the transducer moves in the opposite direction.
- An inertia-based localizer is mounted to the transducer to track its motion. As the transducer rotates within the inner circumference of the brace, it collects data as to the outer circumferential topography of the bone surface.
- the RT approach includes the advantage of lower cost than the stationary transducer designs and higher accuracy due to the greater number of localized bone surface points for each tracking step, while maintaining a mechanical flexibility.
- a third module of the exemplary diagnostic system includes thin film accelerometers that detect the vibration produced by motion of the knee joint.
- Thin film accelerometers are used in lieu of sound sensors, because of better performance and less noise susceptibility.
- the thin film accelerometers may be the same ones used for the localizer, as well as having the same circuitry for driving the accelerometers.
- the accelerometers are attached to the patients and communicatively connected to the kinematic tracking braces so the outputs from the accelerometers can be amplified, digitized, and sent wirelessly to the controller computer.
- X-ray video fluoroscopy and in-vivo measurements of dynamic knee kinematics are important for understanding the effects of joint injuries, diseases, and evaluating the outcome of surgical procedures.
- six degrees of freedom (DOF) are determined between the femur and tibia, femur and patella, and tibia and patella that involve the position and orientation of each with respect to the other.
- DOF degrees of freedom
- the accuracy of this approach is within one degree of rotation and one mm of translation (except for translation parallel to the viewing direction).
- this approach is highly accurate, it constrains the patient to remain within the small working volume of the fluoroscope unit and subjects the patient to ionizing radiation for a prolonged period of time.
- an exemplary system accurately measures joint motion during dynamic activities using a portable brace, such as those previously discussed herein.
- a portable brace having sensors mounted thereto, X-ray fluoroscopy may be omitted.
- Implementation of joint movement visualization includes using the exemplary 3C model reconstruction using pulse-echo A-mode ultrasound system to measure vibrations produced to accurately localize the exact vibration center and causes for its occurrence.
- the interpretation of the vibration and kinematic data is a complicated task involving an in-depth understanding of data acquisition, training data sets and signal analysis, as well as the mechanical system characteristics. Vibrations generated through the implant components, bones, and/or soft tissues interaction result from a forced vibration induced by driving force leading to a dynamic response.
- the driving force can be associated with the impact following knee ligament instability, bone properties, and conditions.
- a normal, intact knee will have a distinct pattern of motion, coupled with distinct vibrational characteristics. Once degeneration or damage occurs to the knee joint, both the kinematic patterns and vibrational characteristics become altered. This altering, for each type of injury or degeneration, leads to distinct changes that can be captured using both kinematic and vibration determination.
- the intelligent diagnosis module which is a software module, is operative to diagnose ligament, other soft tissue, and bone injuries.
- the diagnostic module is a two stage device that includes a first stage involving motion measurement extraction, while a second stage classifies any injury that is detected.
- This first stage includes acquisition of kinematic feature vectors using multiple physiological measurements taken from the patient while the patient moves the joint in question through a range of motion.
- Exemplary measurements include, without limitation, medical condyle anteroposterior motion (MAP) and lateral condyle anteroposterior (LAP), with the latter pertaining to the anterior-posterior A/P distance of the medial and lateral condyle points relative to the tibia geometric center.
- MAP medical condyle anteroposterior motion
- LAP lateral condyle anteroposterior
- FIG. 20 is an exemplary graphical representation showing average ACLD medial and lateral condyle contact positions during a deep knee bend activity.
- the motion features vectors extracted from the kinematic and vibration analyses are output to a multilayer back propagation neural network for determining the injured ligament.
- an exemplary neural network classifier has multiple binary outputs. Each output is either a one or zero, with one corresponding to yes and zero corresponding to no.
- each output represents the response of the neural network to a particular injury type; for example one output will represent the response for anterior cruciate ligament deficiency (ACLD), its state will be one if an ACL injury is detected, and zero otherwise.
- ACLD anterior cruciate ligament deficiency
- the neural network may be significantly more sophisticated or less sophisticated, depending upon the underlying model of the joint in question.
- construction of the exemplary neural network includes formulating a supervised classifier using a training set of the kinematic and vibration data corresponding to normal and injured knee joints.
- the NN is trained with a set of vectors. Each vector consists of data (kinematics, vibrations and forces) collected from one joint. Fluoroscopy data can be used to calculate the kinematics.
- the NN Once the NN is trained, it can be used to classify new cases and categorize the injury type using these kinematics, vibration and forces data.
- the types and classifications desired to be accommodated by the neural network necessarily include training the neural network on these very types and classifications.
- Exemplary types and classifications of injuries to a mammalian knee joint include, without limitation, osteoarthritic conditions, soft tissue damage, and abnormal growths.
- the neural network also needs to be trained as to indicators of normal knee function. In this manner, once the neural network is trained, it has the capability to differentiate between and output diagnosis data concerning normal and abnormal knee conditions.
- the vibration and kinematics features of a person's knee joint are compiled and fed to the trained neural network.
- the trained neural network then diagnoses the condition of the patient's knee joint, identifying and degeneration by type and severity.
- Exemplary embodiments may be adapted to collect data outside of a clinical setting.
- an exemplary embodiment may be worn by a patient for an extended period of time while performing normal activities.
- a patient may wear vibration sensors and/or a kinematics tracking brace during activities that are not reproducible in the office (for example, weight lifting, racquet ball etc.) that elicit the pain or symptom.
- the patient may turn the device on immediately prior to the activity and/or the patient may mark the event when it occurs. This enables analysis of the data just a few seconds before the marked time to see what abnormal sounds or joint kinematic were occurring.
- Data may be stored on a portable hard drive (or any other portable storage device) and then may be downloaded to exemplary systems for analysis.
- the data can be transmitted and stored in a computer wirelessly. It can also be stored with a miniature memory drive if field data is desired. If the occurrence of the pain was more random, exemplary devices allow continuous gathering of data. In embodiments, the patient may mark the event. Devices capable of continuous monitoring may require a larger data storage capacity.
- embodiments may be easily adapted to other joints of a mammalian animal.
- embodiments may be adapted for use on hips, ankles, toes, spines, shoulders, elbows, wrists, fingers, and temporomandibular joints.
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Engineering & Computer Science (AREA)
- Biomedical Technology (AREA)
- Public Health (AREA)
- Medical Informatics (AREA)
- General Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Pathology (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Heart & Thoracic Surgery (AREA)
- Surgery (AREA)
- Animal Behavior & Ethology (AREA)
- Veterinary Medicine (AREA)
- Dentistry (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Physiology (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Orthopedic Medicine & Surgery (AREA)
- Rheumatology (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computer Networks & Wireless Communication (AREA)
- Primary Health Care (AREA)
- Epidemiology (AREA)
- Data Mining & Analysis (AREA)
- Psychiatry (AREA)
- Signal Processing (AREA)
- Geometry (AREA)
- Rehabilitation Therapy (AREA)
- Databases & Information Systems (AREA)
- Physical Education & Sports Medicine (AREA)
- General Business, Economics & Management (AREA)
- Business, Economics & Management (AREA)
- Evolutionary Computation (AREA)
- Theoretical Computer Science (AREA)
- Mathematical Physics (AREA)
- General Engineering & Computer Science (AREA)
Abstract
A method for diagnosing a joint condition includes in one embodiment: creating a 3d model of the patient specific bone; registering the patient's bone with the bone model; tracking the motion of the patient specific bone through a range of motion; selecting a database including empirical mathematical descriptions of the motion of a plurality actual bones through ranges of motion; and comparing the motion of the patient specific bone to the database.
Description
- This Application is a continuation of U.S. patent application Ser. No. 12/364,267, filed Feb. 2, 2009, the disclosure of which is incorporated by reference herein in its entirety.
- The present invention relates to diagnosis of bodily abnormalities, and more particularly, to devices and methods for evaluating the physiological condition of bodily tissue to discern whether abnormalities exist and the extent of any abnormalities. While the exemplary embodiments disclosed herein are utilized and discussed with respect to a human knee joint, it is to be understood that other joints and bodily tissues may be likewise diagnosed.
- In humans, the knee joint is functionally controlled by a mechanical system governed by three unique types of forces: (1) active forces resulting in motion, such as those resulting from muscle flexing or relaxing; (2) constraining forces that constrain motion, such as those resulting from ligaments being in tension; and (3) compressive forces that resist motion, such as those acting upon bones. In addition to the foregoing bodily tissues accounting for these three forces, cartilage and meniscus also produce a dampening effect dissipating the compressive forces propagating to other joints.
- Knee joint motions are stabilized primarily by four ligaments, which restrict and regulate the relative motion between the femur, tibia, and patella. These ligaments are the anterior cruciate ligament (ACL), the posterior cruciate ligament (PCL), the medial collateral ligament (MCL), and the lateral collateral ligament (LCL), as shown in
FIGS. 1 and 2 . An injury to any of these ligaments or other soft-tissue structures can cause detectable changes in knee kinematics and the creation of detectable patterns of vibration representative of the type of knee joint injury and the severity of the injury. These visual and auditory changes are produced by the bones while moving in a distorted kinematic pattern, and they differ significantly from the look and vibration of a properly balanced knee joint moving through a range of motion. - Many research studies have been conducted to assess knee vibration and correlate it with clinical data regarding various joint problems using microphones with or without stethoscope equipment. However, it has been shown that microphones cannot reliably detect joint frequencies, especially those experiencing strong interference from noise, and the signal clearance can substantially influenced by skin friction. It has been hypothesized that the failure associated with the interpretation of sound emissions and possible reasons for occurrence is directly attributable to the complicity of the sound signal, the unknown noise factors, and unknown sound center. It is desirable, therefore, to provide a diagnostic tool that compares patient specific data with kinematic data by providing visual feedback to clinicians.
- The present invention, in one embodiment, provides a patient specific 3D model of a patient's joint, including bone and soft tissue. This model is then registered to the patient's actual bone so that as the joint is taken through a range of motion it can be visualized on a computer screen. A physician can then use the computer generated image to make a diagnosis or compare the motion of the actual bone to a database of clinically relevant information on desirable or undesirable joint motion.
- The exemplary embodiments of the present invention include a diagnostic system for mammalian bodies to determine the type of injury and extent of injury using kinematic data and/or vibration data. In particular, an exemplary method and embodiment are directed to a knee joint diagnostic system for automatically determining the type of injury and the extent to which ligaments, muscles, bones, meniscus, and cartilage may be affected by an injury through analyzing the kinematics of the knee joint, while also analyzing the pattern and special distribution of the vibration produced knee joint movement. An exemplary process flow diagram for this exemplary method is shown in
FIG. 3 . - To evaluate knee kinematics, patient-specific 3D models of the distal femur, proximal tibia, and the patella are constructed using pulse echo A-mode ultrasound based 3D model reconstruction technology. In addition, patient-specific kinematic data is obtained for the motions of the femur, tibia, and patella using pulse A-mode ultrasound. Finally, patient specific vibration data is obtained while the knee joint is taken through a range of motion and loaded in real-world conditions. In exemplary form, the vibration data and kinematic data are taken at the same time using the single data acquisition device. In a further exemplary embodiment, if the data is acquired in a physician's office, the data is displayed in real-time on a split screen monitor. If, however, the data is acquired outside of the doctor's office, a recording device and memory may be utilized to record the data in a time synched manner. In a yet a further exemplary embodiment, the patient may be given an actuator that is operative to note the general time frame within which the patient felt a particular pain or severe pain to allow a correlation between pain felt by the patient and the kinematics and vibration occurring at roughly the same time.
- Patient-specific data is analyzed by a trained neural network in order to provide an automated output as to the existence of an injury, the type of injury, and the severity of the injury. This neural network may be accessible via the internet or may reside on a physician's local computer. In addition, or in the alternative, patient-specific data may be analyzed by a physician to make the diagnosis directly without the aid of the neural network.
- Using the exemplary methods and devices as disclosed herein, a physician may diagnose a bodily injury without requiring experimental surgery or requiring exposure of the patient to radiation from still X-rays or fluoroscopy. In addition, the data taken regarding each patient is continuous through a range of motion, in contrast to X-rays and fluoroscopy which take images at distinct points with significant range of motion gaps. In addition, data taken in accordance with the exemplary method and devices disclosed herein also contrasts data taken by a magnetic resonance imaging machine, not only because the data taken is continuous along the range of motion, but also because the bodily portion evaluated is acting under loaded conditions in a dynamic environment.
- It is an object of the present invention to provide a method of creating a three dimensional model of a patient's bone using tracked pulse-echo A-Mode ultrasound and atlas-based deformable models.
- It is another object of the present invention to provide a method of registering a patient's bone with a three dimensional model of the patient's actual bone.
- Another object of the present invention is to provide a method of tracking the motion of a patient's actual bone through space and showing the same on a computer screen.
- Yet another object of the present invention is to provide a method of tracking at least two bones relative to one another as three dimensional models on a computer screen as the actual bones are taken through a range of motion.
- It is also an object of the present invention to provide a method of diagnosis for joint conditions based on a database of kinematic or other information about joint motion.
-
FIG. 1 is a posterior view of a human knee joint in a fixed position; -
FIG. 2 is a posterior view of a human knee joint in an extended position; -
FIG. 3 is an exemplary process flow diagram using exemplary methods within the scope of the present invention; -
FIG. 4 is a schematic diagram of the modules of an exemplary diagnostic system; -
FIG. 5 is a screen shot of a software user interface for bone modeling; -
FIG. 6 is an anterior view of the bones of a human knee joint in an extended position; -
FIG. 7 is a pictorial representation of a human leg having an exemplary brace attached to a distal segment of the femur, and exemplary brace attached to a proximal segment of the tibia, a sensor mounted proximate the patella, and a foot pressure sensing shoe; -
FIG. 8 is a is an illustration of a CT slice of the transcutaneous detection of a bone's surface using pulse echo A-mode ultrasound; -
FIG. 9 is a schematic of an exemplary inertia-based localizer circuit; -
FIG. 10 is a schematic of an exemplary brace circuit architecture; -
FIG. 11 is a circuit schematic of an exemplary high voltage amplifier; -
FIG. 12 is a circuit layout for the exemplary high voltage amplifier ofFIG. 11 ; -
FIG. 13 is a block diagram for an exemplary high voltage multiplexer; -
FIG. 14 is a block diagram for an exemplary receiving circuit; -
FIG. 15 is a pictorial representation of an exemplary kinematics tracking brace; -
FIG. 16 is a pictorial representation of an alternative exemplary kinematics tracking brace; -
FIG. 17 is a pictorial representation of a further alternative kinematics exemplary tracking brace; -
FIG. 18 is a pictorial representation of a vibration detection module; -
FIGS. 19A, 19B, and 19C are pictorial representations of exemplary kinematics data, vibration signal, and force data respectively; -
FIG. 20 is a graphical representation showing average ACLD medial and lateral condyle contact positions during a deep knee bend activity; -
FIGS. 21A, 21B, and 21C are a series of views showing contact path tracking in accordance with the exemplary embodiments; -
FIG. 22 is a is a schematic of the overall classification system flow chart; -
FIG. 23 is a schematic representation of an exemplary neural network classifier; -
FIG. 24 is an exemplary process flow for training an exemplary neural network; and -
FIG. 25 is an exemplary process flow for knee deficiency diagnosis using a trained neural network. - The exemplary embodiments of the present invention are described and illustrated below to encompass diagnosis of bodily abnormalities and, more particularly, devices and methods for evaluating the physiological condition of bodily tissue to discern whether abnormalities exist and the next of any abnormalities. Of course, it will be apparent to those of ordinary skill in the art that the preferred embodiments discussed below are exemplary in nature and may be reconfigured without departing from the scope and spirit of the present invention. However, for clarity and precision, the exemplary embodiments as discussed below may include optional steps, methods and features that one of ordinary skill should recognize as not being a requisite to fall within the scope of the present invention. In exemplary fashion, the embodiments disclosed herein are described with respect to diagnosing a knee joint injury. Nevertheless, the embodiments may be utilized to diagnose other joints and bodily tissue injuries, as the knee joint is merely exemplary to facilitate an understanding of the embodiments disclosed.
- Referencing
FIG. 4 , a first exemplary diagnostic system includes four modules: (1) a pulse echo A-mode ultrasound based 3D model reconstruction (PEAUMR) module for constructing 3D patient specific models of the knee joint bones; (2) a joint kinematics tracking (JKT) module for tracking kinematics of the knee joint using the patient-specific 3D model of the knee joint from the PEAUMR module; (3) a vibration detection (VD) module for capturing sounds emanating from the knee joint while in motion; and (4) an intelligent diagnosis (ID) module for identifying pathological cases of the knee joint using kinematic data and associated vibration data gathered during the joint motion. Each of these four modules is described in further detail in the following sections. The foot sensor interacts in real time with these other modules providing dynamic force data. - It will be understood by those of skill in the art that the system described above is usable with or without the use of the vibration detection module. For example, one may use the present invention by mathematically describing the relative motion of bones in a patient's joint as such motion is tracked in a 3D patient specific bone model and comparing such description with a database of mathematical descriptions of joint motion. The database could contain mathematical descriptions of healthy or clinically undesirable joint motion.
- Referring to
FIG. 5 , the PEAUMR module constructs a 3D model of a subject's (e.g., a patient) bones by transcutaneously acquiring a set of 3D data points that in total are representative of the shape of the bone's surface using a tracked pulse echo A-mode ultrasound probe. The probe consists of a single ultrasound transducer attached to a global localizer. The global localizer may be optical, inertial, electromagnetic or ultra wide band radio frequency. The probe is battery-powered and connected wirelessly to a computer in order to record the set points and construct a unique or patient-specific bone model using an atlas-based deformable model technique. - The computer includes software that interprets data from the tracked pulse echo A-mode ultrasound probe and is operative to construct the 3D models of the patient's bones, which will look very similar to the model shown in
FIG. 6 . The patient-specific bone is reconstructed using the set of points collected from the bone's surface transcuateously by the tracked ultrasound probe. These points are then used by the atlas-based deformable model software to reconstruct the 3D model of the patient's bone. - In exemplary form, the software includes a plurality of bone models of the femur, tibia, and patella that are classified, for example, based upon ethnicity, gender, skeletal bone to be modeled, and the side of the body the bone is located. Each of these classifications is accounted for by the dropdown menus of the software so that the model initially chosen by the software most closely approximates the bone of the patient.
- After the software selects the bone model to approximate the bone of the patient, the ultrasound transducer probe is repositioned on the exterior of the skin and data points are generated and applied to the model bone (in this case a distal femur), numerically recorded and viewable in a data window, and ultimately utilized by the software to conform the bone model to the patient's actual bone shape. Obviously, a higher number of data points imposed on the model will generally result in a more accurate patient model. Nevertheless, in view of the model bones already taking into account numerous traits of the patient (ethnicity, gender, bone modeled, and body side of the bone), it is quite possible to construct an accurate patient-specific 3D model with as few as 150 data points, which typically can be taken by repositioning the probe over the bone for 30 seconds for each bone. In this example, it is preferable for the data to be acquired both while the knee is bent and extended to more accurately shape the ends of the bones. This same procedure is repeated for the remaining bones of the joint, in this case the proximal end of the tibia and the patella, in order for the software to combine the bones thereby forming the joint. Ultrasound will not be affected whether the patient has normal or prosthetic implant. The 3D model of the femur can be resected and attached with the implanted CAD model.
- Referring to
FIG. 7 , the JKT module tracks the kinematics of the knee joint using the patient-specific 3D bone models from the PEAUMR module. In this exemplary embodiment, motion tracking of the patient's knee joint bones is performed using one or more bone motion tracking braces. In exemplary form, the bone motion tracking brace includes pulse echo A-mode ultrasound transducers to transcutaneously localize points on the bones surface. Incidentally, the pulse echo A-mode ultrasound transducers may or may not be identical to the pulse echo A-mode ultrasound transducers used by the PEAUMR module. Commercially available transducers for use with the exemplary embodiments include, without limitation, the Olympus immersion unfocused 3.5 MHz transducer. The force sensing shoe detects the ground reactive pressures simultaneous with knee joint kinematic data acquisition. - Each ultrasound transducer is tracked using an accelerometer or a sensor-specific localizer (or any other appropriate inertial sensor). The resulting localized bone points generated from the outputs of the ultrasound transducers are used in combination with the patient specific 3D bone models to discern bone movement while the knee joint is taken through a range of motion. In exemplary form, three braces and a foot force sensing shoe are used to track knee joint kinematics and dynamic forces: (a) a first brace is positioned proximate the distal portion of the femur; (b) a second brace is positioned proximate the distal end of the tibia; and, (c) a third brace is positioned proximate the patella region.
- Referring to
FIG. 8 , an exemplary bone motion tacking brace includes a plurality of pulse echo A-mode ultrasound transducers for transcutaneous detection of the bone's surface and inertia-based localizers to track the motion of the ultrasound transducers, which in turn, track the bones motion. Each brace is wirelessly connected to a computer operative to perform computations and visualization in real-time showing movements of the patient-specific 3D bone models paralleling movements of the patient's actual knee joint in a time synchronized manner. Each exemplary brace include a rigid or semi-rigid body having a plurality (two or more) of complementary metal oxide semiconductor (CMOS) inertia-based sensors attached thereto. The position of each sensor and/or transducer is tracked by using the equation of motion: Fr+Fr*=0, where, Fr is a summation of all the generalized active forces in the system, and Fr* is a summation of all the generalized inertia forces in the system. The homogenous transformation between the localizer's reference coordinate frame and the world coordinate frame is calculated using the positions of multiple inertia sensors. The following equation calculates the linear movement of the transducer: v(n+1)=v(n)+a(n)dt and s(n+1)=s(n)+v(n)dt−0.5a(n)dt2, where s(n+1) is position at the current state, s(n) is the position from the previous state, v(n+1) is instantaneous velocity of the current state, v(n) is the velocity from previous state, and a(n) is the acceleration from the accelerometer and dt is the sampling time interval. The previous equations describe the dynamic motion and positioning of a point in 3D Euclidean space. Additionally information is needed to describe a 3D body orientation and motion. The orientation of the transducer can be described by using a gravity based accelerometer (example: ADXL-330, analog device) by extracting the tilting information from each pair of orthogonal axis. The acceleration output on x, y, or z due to gravity is equal to the following: Ai=(Vouti−Voff)/S, where Ai is the acceleration at x, y, or z axis, Vouti is the voltage output from the x, y, or z axis, Voff is the offset voltage, and S is the sensitivity of the accelerometer. The yaw, pitch and row can be calculated as shown in the following: -
- where pitch is ρ, which is x-axis relative to the ground, roll is φ, which is y-axis relative to the ground, and row is θ, which is z-axis relative to the ground. Since the accelerometer is based using gravity, the orientation does not require information from the previous state once the sensor is calibrated. The static calibration requires the resultant sum of accelerations from the 3 axis equal to 1 g. Alternatively, an orientation sensor that gives us yaw, pitch, and row information of the body are also commercially available (example: IDG-300, Invensense). The orientation of the transducer can then be resolved by using direction cosine matrix transformation:
-
- where C shorts for cosine and S shorts for sine.
- Referring to
FIG. 9 , an accelerometer based localizer is used to track each pulse echo A-mode ultrasound transducer mounted to the brace. The localizer comprises a plurality of nodes, with each node comprising a CMOS accelerometer and a temperature sensor for thermal drift comparison. Each node is integrated to minimize noise and distortion. The outputs of the accelerometers regarding the X, Y, and Z coordinates and temperature sensor are directed to a multiplexer that multiplexes the signals. Multiplexed outputs are amplified by an amplifier and then directed to an analog-to-digital converter. The digital conversion of the signal can be performed within or outside the CMOS sensors chip. Outputted digital signals are directed to a wireless transmitter by way of a parallel input/serial output device. - Referring to
FIG. 10 , each of the three exemplary design alternatives for the brace has a similar electronic architecture. An exemplary electronic architecture includes a high voltage amplifier circuit feeding a voltage multiplexer circuit to excite each ultrasound transducer and thereby acting as an analog switch. The echo signals from each transducer are multiplexed pursuant to a logic control directing the opening of the switches in the multiplexer circuit at precise intervals. An exemplary logic control is the MSP430 available from Texas Instruments. The output from the multiplexer circuit is amplified by an amplifier circuit, signal conditioned using a signal conditioning circuit, and digitized using an analog-to-digital converter. Electric power to the foregoing components is supplied by way of a battery, which also supplies power to a wireless transmitter module. In exemplary form, the wireless transmitter module utilizes the universal asynchronous receiver/transmitter (UART) protocol. The module includes a wireless transmitter circuit receiving the output of the first in-first out (FIFO) buffer of the analog-to-digital converter by way of a serial interface. An output from the wireless transmitter circuit is conveyed using a serial link coupled to an antenna. Signals conveyed through the antenna are broadcast for reception by a wireless receiver coupled to a controller computer. - Referring to
FIGS. 11 and 12 , an exemplary high voltage circuit is utilized to trigger and generate the excitation energy for the piezoelectric crystal in the transducer. Exemplary high voltage circuits for use in this embodiment include, without limitation, the pulsar integrated circuit (HV379) available from Supertex. - Referencing
FIG. 13 , an exemplary high voltage multiplexer is utilized to trigger and excite multiple piezoelectric transducers without increasing the number of high voltage circuits mentioned with regard toFIG. 11 . Exemplary high voltage multiplexers for use in this embodiment include, without limitation, the high voltage multiplexer (HV2221) available from Supertex. The advantage of using a high voltage multiplexer is the ability to use CMOS level control circuitry, thereby making the control logic compatible with virtually any microcontroller or field programmable gate array commercially available. - Referring to
FIG. 14 , an exemplary receiving circuit, which comprises the multiplexer circuit, the amplifier circuit, the signal conditioning circuit, and the analog-to-digital converter, is utilized to receive the echo signals from each transducer. Exemplary receiving circuits for use in this embodiment include, without limitation, the AD9271 8-channel ultrasound receiving integrated circuit available from Analog Devices. - Referring to
FIG. 15 , a first exemplary bone tracking brace includes a plurality of transducers mounted thereto. Each transducer is responsible for determining the location of a point on the surface of the bone for each motion tracking frame. Problems of locating and tracking the bone using ultrasound data are reduced as the motion of the bone relative to the skin is small compared to the gross joint motion. There are at least three approaches disclosed herein for tracking the motion of the ultrasound transducers themselves. The first approach, commonly referred to herein as the ITT (Individual transducer tracking) approach, involves each transducer in the brace having an inertia-based localizer to individually track each transducer. Using the ITT approach, in exemplary form, the transducers are held together by flexible length straps. - Referencing
FIG. 16 , a second approach, commonly referred to herein as the ITML (Inter-transducers Mechanical Links) approach, involves the transducers being connected to each other by movable mechanical links. Each mechanical link includes length and angle sensors that allow for detection of the movement of the transducers relative to one another and the relative translational motions of the links. Every two links are connected by a pivot pin that allows rotation and translation of the links relative to each other. An angle sensor is mounted to at least one link proximate the pivot pin to allow for detection of the angle between the links. The ITML approach features less localizers than the individual transducer tracking design. - Referring to
FIG. 17 , a third approach, commonly referred to herein as the RT (Rotating Transducer) approach, involves using a single ultrasound transducer that is mounted to a carriage. The carriage traverses along a track located on the inner circumference of the brace. For example, the carriage may be moved along the tack by a string loop that is wrapped around the drive shaft of a motor. When the transducer reaches the motor, the rotation direction of the motor is changed and the transducer moves in the opposite direction. - An inertia-based localizer is mounted to the transducer to track its motion. As the transducer rotates within the inner circumference of the brace, it collects data as to the outer circumferential topography of the bone surface. By using a single transducer, the RT approach includes the advantage of lower cost than the stationary transducer designs and higher accuracy due to the greater number of localized bone surface points for each tracking step, while maintaining a mechanical flexibility.
- Referring to
FIG. 18 , a third module of the exemplary diagnostic system, the vibration detection module, includes thin film accelerometers that detect the vibration produced by motion of the knee joint. Thin film accelerometers are used in lieu of sound sensors, because of better performance and less noise susceptibility. In exemplary form, the thin film accelerometers may be the same ones used for the localizer, as well as having the same circuitry for driving the accelerometers. The accelerometers are attached to the patients and communicatively connected to the kinematic tracking braces so the outputs from the accelerometers can be amplified, digitized, and sent wirelessly to the controller computer. - Referring to
FIG. 22 , X-ray video fluoroscopy and in-vivo measurements of dynamic knee kinematics are important for understanding the effects of joint injuries, diseases, and evaluating the outcome of surgical procedures. In exemplary form, using the two aforementioned techniques, six degrees of freedom (DOF) are determined between the femur and tibia, femur and patella, and tibia and patella that involve the position and orientation of each with respect to the other. The accuracy of this approach is within one degree of rotation and one mm of translation (except for translation parallel to the viewing direction). Although this approach is highly accurate, it constrains the patient to remain within the small working volume of the fluoroscope unit and subjects the patient to ionizing radiation for a prolonged period of time. For most dynamic activities where the joints are loaded such as running, jumping, or other dynamic activities, fluoroscopy is an unacceptable alternative. To address this deficiency in preexisting approaches, an exemplary system accurately measures joint motion during dynamic activities using a portable brace, such as those previously discussed herein. By using a portable brace having sensors mounted thereto, X-ray fluoroscopy may be omitted. - Implementation of joint movement visualization includes using the exemplary 3C model reconstruction using pulse-echo A-mode ultrasound system to measure vibrations produced to accurately localize the exact vibration center and causes for its occurrence. The interpretation of the vibration and kinematic data is a complicated task involving an in-depth understanding of data acquisition, training data sets and signal analysis, as well as the mechanical system characteristics. Vibrations generated through the implant components, bones, and/or soft tissues interaction result from a forced vibration induced by driving force leading to a dynamic response. The driving force can be associated with the impact following knee ligament instability, bone properties, and conditions. A normal, intact knee will have a distinct pattern of motion, coupled with distinct vibrational characteristics. Once degeneration or damage occurs to the knee joint, both the kinematic patterns and vibrational characteristics become altered. This altering, for each type of injury or degeneration, leads to distinct changes that can be captured using both kinematic and vibration determination.
- Referencing
FIG. 25 , a fourth module of the exemplary diagnostic system, the intelligent diagnosis module, which is a software module, is operative to diagnose ligament, other soft tissue, and bone injuries. From previous studies, normal and anterior cruciate ligament deficient (ACLD) knee subjects exhibit a similar pattern of posterior femoral translation during progressive knee flexion, but the subjects exhibit different axial rotation patterns of 30 degrees of knee flexion. Accordingly, the diagnostic module is a two stage device that includes a first stage involving motion measurement extraction, while a second stage classifies any injury that is detected. - This first stage includes acquisition of kinematic feature vectors using multiple physiological measurements taken from the patient while the patient moves the joint in question through a range of motion. Exemplary measurements include, without limitation, medical condyle anteroposterior motion (MAP) and lateral condyle anteroposterior (LAP), with the latter pertaining to the anterior-posterior A/P distance of the medial and lateral condyle points relative to the tibia geometric center. Other exemplary measurements include LSI (distance between the lateral femoral condyle and the lateral tibial plateau) and MSI (distance between the medial femoral condyle and the medial tibial plateau) which are S/I (superior/inferior) distances of the lateral and medial condyle points to the tibial plane. Further exemplary measurements include condyle separation, which is the horizontal (x-y plane) distance between the two minimum condyle points to the tibia (See
FIG. 21 ). Feature vectors also include the femoral position with respect to the tibia which is defined by three Euler angles and three translation components in addition to the vibration signal, and force data (seeFIGS. 19A, 19B, and 19C ).FIG. 20 is an exemplary graphical representation showing average ACLD medial and lateral condyle contact positions during a deep knee bend activity. - Referring to
FIG. 22 , the motion features vectors extracted from the kinematic and vibration analyses are output to a multilayer back propagation neural network for determining the injured ligament. - Referencing
FIG. 23 , an exemplary neural network classifier has multiple binary outputs. Each output is either a one or zero, with one corresponding to yes and zero corresponding to no. In this exemplary neural network classifier, each output represents the response of the neural network to a particular injury type; for example one output will represent the response for anterior cruciate ligament deficiency (ACLD), its state will be one if an ACL injury is detected, and zero otherwise. Obviously, the neural network may be significantly more sophisticated or less sophisticated, depending upon the underlying model of the joint in question. - Referring to
FIG. 24 , construction of the exemplary neural network (NN) includes formulating a supervised classifier using a training set of the kinematic and vibration data corresponding to normal and injured knee joints. The NN is trained with a set of vectors. Each vector consists of data (kinematics, vibrations and forces) collected from one joint. Fluoroscopy data can be used to calculate the kinematics. Once the NN is trained, it can be used to classify new cases and categorize the injury type using these kinematics, vibration and forces data. Those skilled in the art will readily understand that the types and classifications desired to be accommodated by the neural network necessarily include training the neural network on these very types and classifications. Exemplary types and classifications of injuries to a mammalian knee joint include, without limitation, osteoarthritic conditions, soft tissue damage, and abnormal growths. Likewise, the neural network also needs to be trained as to indicators of normal knee function. In this manner, once the neural network is trained, it has the capability to differentiate between and output diagnosis data concerning normal and abnormal knee conditions. - Referencing
FIG. 25 , the vibration and kinematics features of a person's knee joint are compiled and fed to the trained neural network. The trained neural network then diagnoses the condition of the patient's knee joint, identifying and degeneration by type and severity. - Exemplary embodiments may be adapted to collect data outside of a clinical setting. For example, an exemplary embodiment may be worn by a patient for an extended period of time while performing normal activities. For example, a patient may wear vibration sensors and/or a kinematics tracking brace during activities that are not reproducible in the office (for example, weight lifting, racquet ball etc.) that elicit the pain or symptom. In some embodiments, the patient may turn the device on immediately prior to the activity and/or the patient may mark the event when it occurs. This enables analysis of the data just a few seconds before the marked time to see what abnormal sounds or joint kinematic were occurring. Data may be stored on a portable hard drive (or any other portable storage device) and then may be downloaded to exemplary systems for analysis. The data can be transmitted and stored in a computer wirelessly. It can also be stored with a miniature memory drive if field data is desired. If the occurrence of the pain was more random, exemplary devices allow continuous gathering of data. In embodiments, the patient may mark the event. Devices capable of continuous monitoring may require a larger data storage capacity.
- It is also understood that while the exemplary embodiments have been described herein with respect to a knee joint, those skilled in the art will readily understand that the aforementioned embodiments may be easily adapted to other joints of a mammalian animal. For example, embodiments may be adapted for use on hips, ankles, toes, spines, shoulders, elbows, wrists, fingers, and temporomandibular joints.
- Following from the above description and invention summaries, it should be apparent to those of ordinary skill in the art that, while the methods and apparatuses herein described constitute exemplary embodiments of the present invention, the invention contained herein is not limited to this precise embodiment and that changes may be made to such embodiments without departing from the scope of the invention as defined by the claims. Additionally, it is to be understood that the invention is defined by the claims and it is not intended that that any limitations or elements describing the exemplary embodiments set forth herein are to be incorporated into the interpretation of any claim element unless such claim limitation is explicitly stated. Likewise, it is to be understood that it is not necessary to meet any or all of the identified advantages or objects of the invention disclosed herein in order to fall within the scope of any claims. Since the invention is defined by the claims and since inherent and/or unforeseen advantages of the present invention may exist even though they any not have been explicitly discussed herein.
Claims (21)
1.-19. (canceled)
20. A method of generating a three dimensional (3D), patient-specific bone model, the method comprising:
selecting a deformable three dimensional (3D) virtual bone model corresponding to a bone of a patient;
displaying the deformable 3D virtual bone model on a graphical display;
scanning the bone of the patient using an ultrasound transducer;
recording a plurality of data points during the scanning of the bone, where the plurality of data points each correspond to a bone surface point on the bone;
applying a plurality of bone surface points to the deformable 3D virtual bone model graphically displayed; and,
conforming the deformable 3D virtual bone model to the plurality of bone surface points to generate a 3D virtual patient-specific bone model of the bone;
displaying the 3D virtual patient-specific bone model of the bone on the graphical display.
21. The method of claim 20 , wherein selecting the deformable 3D virtual bone model includes accounting for at least one of a gender of the patient and an ethnicity of the patient.
22. The method of claim 20 , wherein selecting the deformable 3D virtual bone model includes accounting for a body side of the patient.
23. The method of claim 20 , wherein the plurality of bone surface points comprises no fewer than one hundred and fifty bone surface points.
24. The method of claim 20 , wherein scanning the bone of the patient using the ultrasound transducer includes repositioning the bone relative to another bone of the patient, where the bone and another bone comprise part of a patient joint.
25. The method of claim 20 , further comprising tracking motion of the bone relative to the ultrasound transducer while scanning the bone of the patient.
26. The method of claim 20 , wherein the deformable 3D virtual bone model is a statistical atlas based bone model.
27. The method of claim 20 , wherein:
scanning the bone of the patient using the ultrasound transducer includes tracking a 3D position of the ultrasound transducer to generate tracking data; and,
the plurality of data points are used with the tracking data to generate the plurality of bone surface points.
28. The method of claim 20 , wherein scanning the bone of the patient using the ultrasound transducer includes repositioning the ultrasound transducer on an exterior of skin covering the bone.
29. A method of generating a three dimensional (3D), patient-specific joint model, the method comprising:
selecting a first deformable three dimensional (3D) virtual bone model corresponding to a first bone of a patient joint;
displaying the first deformable 3D virtual bone model on a graphical display;
scanning the first bone of the patient joint using an ultrasound transducer;
recording a first plurality of data points during the scanning of the first bone, where the first plurality of data points correspond to a first plurality of bone surface points on the first bone;
applying the first plurality of bone surface points to the first deformable 3D virtual bone model graphically displayed;
conforming the first deformable 3D virtual bone model to the first plurality of bone surface points to generate a 3D virtual patient-specific bone model of the first bone;
selecting a second deformable three dimensional (3D) virtual bone model corresponding to a second bone of the patient joint;
displaying the second deformable 3D virtual bone model on the graphical display;
scanning the second bone of the patient joint using the ultrasound transducer;
recording a second plurality of data points during the scanning of the second bone, where the second plurality of data points correspond to a second plurality of bone surface points on the second bone;
applying the second plurality of bone surface points to the second deformable 3D virtual bone model graphically displayed; and,
conforming the second deformable 3D virtual bone model to the second plurality of bone surface points to generate a 3D virtual patient-specific bone model of the second bone.
30. The method of claim 29 , further comprising:
concurrently displaying the 3D virtual patient-specific bone model of the first bone and the 3D virtual patient-specific bone model of the second bone.
31. The method of claim 29 , wherein selecting at least one of the first deformable 3D virtual bone model and the second deformable 3D virtual bone model includes accounting for at least one of a gender of the patient and an ethnicity of the patient.
32. The method of claim 29 , wherein selecting at least one of the first deformable 3D virtual bone model and the second deformable 3D virtual bone model the deformable 3D virtual bone model includes accounting for a body side of the patient.
33. The method of claim 29 , wherein each of the first plurality of bone surface points and first the second plurality of bone surface points comprises no fewer than one hundred and fifty bone surface points.
34. The method of claim 29 , wherein scanning the first bone of the patient using the ultrasound transducer includes repositioning the first bone relative to the second bone of the patient, where the first bone and the second bone comprise part of a patient joint.
35. The method of claim 929, further comprising tracking motion of the first bone relative to the ultrasound transducer while scanning the first bone of the patient.
36. The method of claim 29 , wherein at least one of the first deformable 3D virtual bone model and the second deformable 3D virtual bone model is a statistical atlas based bone model.
37. The method of claim 29 , wherein:
scanning the first bone of the patient using the ultrasound transducer includes tracking a 3D position of the ultrasound transducer to generate first tracking data; and,
the first plurality of data points are used with the first tracking data to generate the first plurality of bone surface points.
38. The method of claim 37 , wherein:
scanning the second bone of the patient using the ultrasound transducer includes tracking the 3D position of the ultrasound transducer to generate second tracking data; and,
the second plurality of data points are used with the second tracking data to generate the second plurality of bone surface points.
39. The method of claim 29 , wherein scanning the first bone of the patient using the ultrasound transducer includes repositioning the ultrasound transducer on an exterior of skin covering the first bone.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US18/238,338 US11935648B1 (en) | 2009-02-02 | 2023-08-25 | Noninvasive diagnostic system |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/364,267 US8444564B2 (en) | 2009-02-02 | 2009-02-02 | Noninvasive diagnostic system |
US13/898,092 US11342071B2 (en) | 2009-02-02 | 2013-05-20 | Noninvasive diagnostic system |
US17/704,376 US11776686B2 (en) | 2009-02-02 | 2022-03-25 | Noninvasive diagnostic system |
US18/238,338 US11935648B1 (en) | 2009-02-02 | 2023-08-25 | Noninvasive diagnostic system |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/704,376 Continuation US11776686B2 (en) | 2009-02-02 | 2022-03-25 | Noninvasive diagnostic system |
Publications (2)
Publication Number | Publication Date |
---|---|
US20240079125A1 true US20240079125A1 (en) | 2024-03-07 |
US11935648B1 US11935648B1 (en) | 2024-03-19 |
Family
ID=42396088
Family Applications (8)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/364,267 Active 2030-11-26 US8444564B2 (en) | 2009-02-02 | 2009-02-02 | Noninvasive diagnostic system |
US13/196,701 Abandoned US20120029345A1 (en) | 2009-02-02 | 2011-08-02 | Noninvasive diagnostic system |
US13/841,402 Active 2030-06-06 US9642572B2 (en) | 2009-02-02 | 2013-03-15 | Motion Tracking system with inertial-based sensing units |
US13/898,092 Active 2032-10-16 US11342071B2 (en) | 2009-02-02 | 2013-05-20 | Noninvasive diagnostic system |
US15/478,148 Active US11004561B2 (en) | 2009-02-02 | 2017-04-03 | Motion tracking system with inertial-based sensing units |
US17/181,372 Pending US20210193313A1 (en) | 2009-02-02 | 2021-02-22 | Motion Tracking System with Inertial-Based Sensing Units |
US17/704,376 Active US11776686B2 (en) | 2009-02-02 | 2022-03-25 | Noninvasive diagnostic system |
US18/238,338 Active US11935648B1 (en) | 2009-02-02 | 2023-08-25 | Noninvasive diagnostic system |
Family Applications Before (7)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/364,267 Active 2030-11-26 US8444564B2 (en) | 2009-02-02 | 2009-02-02 | Noninvasive diagnostic system |
US13/196,701 Abandoned US20120029345A1 (en) | 2009-02-02 | 2011-08-02 | Noninvasive diagnostic system |
US13/841,402 Active 2030-06-06 US9642572B2 (en) | 2009-02-02 | 2013-03-15 | Motion Tracking system with inertial-based sensing units |
US13/898,092 Active 2032-10-16 US11342071B2 (en) | 2009-02-02 | 2013-05-20 | Noninvasive diagnostic system |
US15/478,148 Active US11004561B2 (en) | 2009-02-02 | 2017-04-03 | Motion tracking system with inertial-based sensing units |
US17/181,372 Pending US20210193313A1 (en) | 2009-02-02 | 2021-02-22 | Motion Tracking System with Inertial-Based Sensing Units |
US17/704,376 Active US11776686B2 (en) | 2009-02-02 | 2022-03-25 | Noninvasive diagnostic system |
Country Status (5)
Country | Link |
---|---|
US (8) | US8444564B2 (en) |
EP (2) | EP2391971B1 (en) |
JP (3) | JP5723788B2 (en) |
CA (5) | CA2977574C (en) |
WO (1) | WO2010088696A1 (en) |
Families Citing this family (327)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7559931B2 (en) | 2003-06-09 | 2009-07-14 | OrthAlign, Inc. | Surgical orientation system and method |
JP2007528243A (en) * | 2004-03-05 | 2007-10-11 | デピュー インターナショナル リミテッド | Orthopedic monitoring system, method and apparatus |
US8377066B2 (en) | 2006-02-27 | 2013-02-19 | Biomet Manufacturing Corp. | Patient-specific elbow guides and associated methods |
US9918740B2 (en) | 2006-02-27 | 2018-03-20 | Biomet Manufacturing, Llc | Backup surgical instrument system and method |
US8568487B2 (en) | 2006-02-27 | 2013-10-29 | Biomet Manufacturing, Llc | Patient-specific hip joint devices |
US9173661B2 (en) | 2006-02-27 | 2015-11-03 | Biomet Manufacturing, Llc | Patient specific alignment guide with cutting surface and laser indicator |
US10278711B2 (en) | 2006-02-27 | 2019-05-07 | Biomet Manufacturing, Llc | Patient-specific femoral guide |
US9289253B2 (en) | 2006-02-27 | 2016-03-22 | Biomet Manufacturing, Llc | Patient-specific shoulder guide |
US8608749B2 (en) | 2006-02-27 | 2013-12-17 | Biomet Manufacturing, Llc | Patient-specific acetabular guides and associated instruments |
US7967868B2 (en) | 2007-04-17 | 2011-06-28 | Biomet Manufacturing Corp. | Patient-modified implant and associated method |
US8591516B2 (en) | 2006-02-27 | 2013-11-26 | Biomet Manufacturing, Llc | Patient-specific orthopedic instruments |
US9907659B2 (en) | 2007-04-17 | 2018-03-06 | Biomet Manufacturing, Llc | Method and apparatus for manufacturing an implant |
US8337426B2 (en) | 2009-03-24 | 2012-12-25 | Biomet Manufacturing Corp. | Method and apparatus for aligning and securing an implant relative to a patient |
US9345548B2 (en) | 2006-02-27 | 2016-05-24 | Biomet Manufacturing, Llc | Patient-specific pre-operative planning |
US8535387B2 (en) | 2006-02-27 | 2013-09-17 | Biomet Manufacturing, Llc | Patient-specific tools and implants |
US20150335438A1 (en) | 2006-02-27 | 2015-11-26 | Biomet Manufacturing, Llc. | Patient-specific augments |
US8092465B2 (en) | 2006-06-09 | 2012-01-10 | Biomet Manufacturing Corp. | Patient specific knee alignment guide and associated method |
US8407067B2 (en) | 2007-04-17 | 2013-03-26 | Biomet Manufacturing Corp. | Method and apparatus for manufacturing an implant |
US9339278B2 (en) | 2006-02-27 | 2016-05-17 | Biomet Manufacturing, Llc | Patient-specific acetabular guides and associated instruments |
US9113971B2 (en) | 2006-02-27 | 2015-08-25 | Biomet Manufacturing, Llc | Femoral acetabular impingement guide |
US8608748B2 (en) | 2006-02-27 | 2013-12-17 | Biomet Manufacturing, Llc | Patient specific guides |
US8603180B2 (en) | 2006-02-27 | 2013-12-10 | Biomet Manufacturing, Llc | Patient-specific acetabular alignment guides |
PL2023812T3 (en) | 2006-05-19 | 2017-07-31 | The Queen's Medical Center | Motion tracking system for real time adaptive imaging and spectroscopy |
US9795399B2 (en) | 2006-06-09 | 2017-10-24 | Biomet Manufacturing, Llc | Patient-specific knee alignment guide and associated method |
GB2442441B (en) | 2006-10-03 | 2011-11-09 | Biomet Uk Ltd | Surgical instrument |
WO2008130454A1 (en) * | 2007-04-19 | 2008-10-30 | Mako Surgical Corp. | Implant planning using captured joint motion information |
US20100153081A1 (en) * | 2008-12-11 | 2010-06-17 | Mako Surgical Corp. | Implant planning for multiple implant components using constraints |
JP4378552B1 (en) * | 2008-07-22 | 2009-12-09 | 国立大学法人 東京大学 | Ultrasonic probe support device |
WO2010011978A1 (en) | 2008-07-24 | 2010-01-28 | OrthAlign, Inc. | Systems and methods for joint replacement |
CA2736525C (en) | 2008-09-10 | 2019-10-22 | OrthAlign, Inc. | Hip surgery systems and methods |
US8647287B2 (en) | 2008-12-07 | 2014-02-11 | Andrew Greenberg | Wireless synchronized movement monitoring apparatus and system |
US8444564B2 (en) | 2009-02-02 | 2013-05-21 | Jointvue, Llc | Noninvasive diagnostic system |
US8679186B2 (en) | 2010-06-29 | 2014-03-25 | Ortho Sensor Inc. | Hermetically sealed prosthetic component and method therefor |
US9259179B2 (en) | 2012-02-27 | 2016-02-16 | Orthosensor Inc. | Prosthetic knee joint measurement system including energy harvesting and method therefor |
US8701484B2 (en) | 2010-06-29 | 2014-04-22 | Orthosensor Inc. | Small form factor medical sensor structure and method therefor |
US8720270B2 (en) | 2010-06-29 | 2014-05-13 | Ortho Sensor Inc. | Prosthetic component for monitoring joint health |
US20100331733A1 (en) * | 2009-06-30 | 2010-12-30 | Orthosensor | Sensing device and method for an orthopedic joint |
US8826733B2 (en) | 2009-06-30 | 2014-09-09 | Orthosensor Inc | Sensored prosthetic component and method |
US9462964B2 (en) | 2011-09-23 | 2016-10-11 | Orthosensor Inc | Small form factor muscular-skeletal parameter measurement system |
US8714009B2 (en) | 2010-06-29 | 2014-05-06 | Orthosensor Inc. | Shielded capacitor sensor system for medical applications and method |
US8707782B2 (en) | 2009-06-30 | 2014-04-29 | Orthosensor Inc | Prosthetic component for monitoring synovial fluid and method |
US8118815B2 (en) | 2009-07-24 | 2012-02-21 | OrthAlign, Inc. | Systems and methods for joint replacement |
US10869771B2 (en) | 2009-07-24 | 2020-12-22 | OrthAlign, Inc. | Systems and methods for joint replacement |
DE102009028503B4 (en) | 2009-08-13 | 2013-11-14 | Biomet Manufacturing Corp. | Resection template for the resection of bones, method for producing such a resection template and operation set for performing knee joint surgery |
CA2825042C (en) * | 2010-01-21 | 2021-01-05 | OrthAlign, Inc. | Systems and methods for joint replacement |
US8632547B2 (en) | 2010-02-26 | 2014-01-21 | Biomet Sports Medicine, Llc | Patient-specific osteotomy devices and methods |
US8926530B2 (en) | 2011-09-23 | 2015-01-06 | Orthosensor Inc | Orthopedic insert measuring system for having a sterilized cavity |
EP2585835A1 (en) * | 2010-06-22 | 2013-05-01 | Stephen J. McGregor | Method of monitoring human body movement |
US20130144135A1 (en) * | 2011-08-02 | 2013-06-06 | Mohamed R. Mahfouz | Method and apparatus for three dimensional reconstruction of a joint using ultrasound |
US10512451B2 (en) | 2010-08-02 | 2019-12-24 | Jointvue, Llc | Method and apparatus for three dimensional reconstruction of a joint using ultrasound |
US9271744B2 (en) | 2010-09-29 | 2016-03-01 | Biomet Manufacturing, Llc | Patient-specific guide for partial acetabular socket replacement |
US20120123252A1 (en) * | 2010-11-16 | 2012-05-17 | Zebris Medical Gmbh | Imaging apparatus for large area imaging of a body portion |
US9968376B2 (en) | 2010-11-29 | 2018-05-15 | Biomet Manufacturing, Llc | Patient-specific orthopedic instruments |
US9111305B2 (en) * | 2010-12-17 | 2015-08-18 | Amazon Technologies, Inc. | Personal remote storage for purchased electronic content items |
US9521966B2 (en) * | 2012-05-17 | 2016-12-20 | Alan N. Schwartz | Localization of the parathyroid |
US9241745B2 (en) | 2011-03-07 | 2016-01-26 | Biomet Manufacturing, Llc | Patient-specific femoral version guide |
WO2012139007A1 (en) * | 2011-04-06 | 2012-10-11 | Northeastern University | Joint sensor devices and methods |
US8715289B2 (en) | 2011-04-15 | 2014-05-06 | Biomet Manufacturing, Llc | Patient-specific numerically controlled instrument |
US9675400B2 (en) | 2011-04-19 | 2017-06-13 | Biomet Manufacturing, Llc | Patient-specific fracture fixation instrumentation and method |
US8668700B2 (en) | 2011-04-29 | 2014-03-11 | Biomet Manufacturing, Llc | Patient-specific convertible guides |
US8956364B2 (en) | 2011-04-29 | 2015-02-17 | Biomet Manufacturing, Llc | Patient-specific partial knee guides and other instruments |
US8532807B2 (en) | 2011-06-06 | 2013-09-10 | Biomet Manufacturing, Llc | Pre-operative planning and manufacturing method for orthopedic procedure |
US9084618B2 (en) | 2011-06-13 | 2015-07-21 | Biomet Manufacturing, Llc | Drill guides for confirming alignment of patient-specific alignment guides |
US20130001121A1 (en) | 2011-07-01 | 2013-01-03 | Biomet Manufacturing Corp. | Backup kit for a patient-specific arthroplasty kit assembly |
US8764760B2 (en) | 2011-07-01 | 2014-07-01 | Biomet Manufacturing, Llc | Patient-specific bone-cutting guidance instruments and methods |
US8597365B2 (en) | 2011-08-04 | 2013-12-03 | Biomet Manufacturing, Llc | Patient-specific pelvic implants for acetabular reconstruction |
CA2845044C (en) | 2011-08-12 | 2023-03-28 | Jointvue, Llc | 3-d ultrasound imaging device and methods |
US9606209B2 (en) | 2011-08-26 | 2017-03-28 | Kineticor, Inc. | Methods, systems, and devices for intra-scan motion correction |
HUP1100471A2 (en) * | 2011-08-30 | 2013-04-29 | Bay Zoltan Alkalmazott Kutatasi Koezhasznu Nonprofit Kft | Method and instrument for detecting equilibrium and intelligent insole suitable for monitoring walking parameters |
US9295497B2 (en) | 2011-08-31 | 2016-03-29 | Biomet Manufacturing, Llc | Patient-specific sacroiliac and pedicle guides |
US9066734B2 (en) | 2011-08-31 | 2015-06-30 | Biomet Manufacturing, Llc | Patient-specific sacroiliac guides and associated methods |
US9414940B2 (en) | 2011-09-23 | 2016-08-16 | Orthosensor Inc. | Sensored head for a measurement tool for the muscular-skeletal system |
US9839374B2 (en) | 2011-09-23 | 2017-12-12 | Orthosensor Inc. | System and method for vertebral load and location sensing |
US8911448B2 (en) | 2011-09-23 | 2014-12-16 | Orthosensor, Inc | Device and method for enabling an orthopedic tool for parameter measurement |
US9386993B2 (en) | 2011-09-29 | 2016-07-12 | Biomet Manufacturing, Llc | Patient-specific femoroacetabular impingement instruments and methods |
CA3194212A1 (en) | 2011-10-14 | 2013-04-18 | Jointvue, Llc | Real-time 3-d ultrasound reconstruction of knee and its implications for patient specific implants and 3-d joint injections |
US9301812B2 (en) | 2011-10-27 | 2016-04-05 | Biomet Manufacturing, Llc | Methods for patient-specific shoulder arthroplasty |
ES2635542T3 (en) | 2011-10-27 | 2017-10-04 | Biomet Manufacturing, Llc | Glenoid guides specific to the patient |
US9451973B2 (en) | 2011-10-27 | 2016-09-27 | Biomet Manufacturing, Llc | Patient specific glenoid guide |
US9554910B2 (en) | 2011-10-27 | 2017-01-31 | Biomet Manufacturing, Llc | Patient-specific glenoid guide and implants |
KR20130046336A (en) | 2011-10-27 | 2013-05-07 | 삼성전자주식회사 | Multi-view device of display apparatus and contol method thereof, and display system |
DK2771712T3 (en) | 2011-10-28 | 2023-06-26 | Decision Sciences Int Corp | ENCODED SPREAD SPECTRUM WAVEFORMS IN ULTRASONIC IMAGES |
US11259744B2 (en) | 2011-11-15 | 2022-03-01 | Neurometrix, Inc. | Transcutaneous electrical nerve stimulator with automatic detection of leg orientation and leg motion for enhanced sleep analysis, including enhanced transcutaneous electrical nerve stimulation (TENS) using the same |
US11247040B2 (en) | 2011-11-15 | 2022-02-15 | Neurometrix, Inc. | Dynamic control of transcutaneous electrical nerve stimulation therapy using continuous sleep detection |
US10279179B2 (en) | 2013-04-15 | 2019-05-07 | Neurometrix, Inc. | Transcutaneous electrical nerve stimulator with automatic detection of user sleep-wake state |
US10335595B2 (en) * | 2011-11-15 | 2019-07-02 | Neurometrix, Inc. | Dynamic control of transcutaneous electrical nerve stimulation therapy using continuous sleep detection |
US9474898B2 (en) | 2013-03-29 | 2016-10-25 | Neurometrix, Inc. | Detecting cutaneous electrode peeling using electrode-skin impedance |
PT2780073T (en) | 2011-11-15 | 2017-12-18 | Neurometrix Inc | Apparatus for relieving pain using transcutaneous electrical nerve stimulation |
US9675801B2 (en) | 2011-11-15 | 2017-06-13 | Neurometrix, Inc. | Measuring the “on-skin” time of a transcutaneous electrical nerve stimulator (TENS) device in order to minimize skin irritation due to excessive uninterrupted wearing of the same |
US10112040B2 (en) | 2011-11-15 | 2018-10-30 | Neurometrix, Inc. | Transcutaneous electrical nerve stimulation using novel unbalanced biphasic waveform and novel electrode arrangement |
US9286355B2 (en) * | 2012-01-16 | 2016-03-15 | Emovi Inc. | Method and system for human joint treatment plan and personalized surgery planning using 3-D kinematics, fusion imaging and simulation |
NL2008437C2 (en) * | 2012-01-19 | 2013-07-22 | Clinical Graphics B V | Process to generate a computer-accessible medium comprising information on the functioning of a joint. |
US9237950B2 (en) | 2012-02-02 | 2016-01-19 | Biomet Manufacturing, Llc | Implant with patient-specific porous structure |
CA3114206A1 (en) | 2012-02-07 | 2013-08-15 | Jointvue, Llc | Three-dimensional guided injection device and methods |
US9844335B2 (en) | 2012-02-27 | 2017-12-19 | Orthosensor Inc | Measurement device for the muscular-skeletal system having load distribution plates |
US9271675B2 (en) | 2012-02-27 | 2016-03-01 | Orthosensor Inc. | Muscular-skeletal joint stability detection and method therefor |
US9622701B2 (en) | 2012-02-27 | 2017-04-18 | Orthosensor Inc | Muscular-skeletal joint stability detection and method therefor |
US9549742B2 (en) | 2012-05-18 | 2017-01-24 | OrthAlign, Inc. | Devices and methods for knee arthroplasty |
US9564682B2 (en) * | 2012-07-11 | 2017-02-07 | Digimarc Corporation | Body-worn phased-array antenna |
US9649160B2 (en) | 2012-08-14 | 2017-05-16 | OrthAlign, Inc. | Hip replacement navigation system and method |
US9992729B2 (en) * | 2012-10-22 | 2018-06-05 | The Nielsen Company (Us), Llc | Systems and methods for wirelessly modifying detection characteristics of portable devices |
US9237885B2 (en) | 2012-11-09 | 2016-01-19 | Orthosensor Inc. | Muscular-skeletal tracking system and method |
US9526437B2 (en) | 2012-11-21 | 2016-12-27 | i4c Innovations Inc. | Animal health and wellness monitoring using UWB radar |
US9060788B2 (en) | 2012-12-11 | 2015-06-23 | Biomet Manufacturing, Llc | Patient-specific acetabular guide for anterior approach |
US9204977B2 (en) | 2012-12-11 | 2015-12-08 | Biomet Manufacturing, Llc | Patient-specific acetabular guide for anterior approach |
US10314733B2 (en) * | 2012-12-20 | 2019-06-11 | Elwha Llc | Sensor-based control of active wearable system |
US9345609B2 (en) | 2013-01-11 | 2016-05-24 | Elwha Llc | Position sensing active torso support |
ES2672296T3 (en) * | 2012-12-31 | 2018-06-13 | Mako Surgical Corporation | Alignment systems using an ultrasonic probe |
US10327708B2 (en) | 2013-01-24 | 2019-06-25 | Kineticor, Inc. | Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan |
US9717461B2 (en) | 2013-01-24 | 2017-08-01 | Kineticor, Inc. | Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan |
US9305365B2 (en) | 2013-01-24 | 2016-04-05 | Kineticor, Inc. | Systems, devices, and methods for tracking moving targets |
CN109008972A (en) | 2013-02-01 | 2018-12-18 | 凯内蒂科尔股份有限公司 | The motion tracking system of real-time adaptive motion compensation in biomedical imaging |
EP2953548A1 (en) * | 2013-02-11 | 2015-12-16 | Koninklijke Philips N.V. | Ultrasound imaging system and method |
WO2014138435A1 (en) | 2013-03-07 | 2014-09-12 | The Regents Of The University Of California | System for health monitoring on prosthetic and fixation devices |
US9839438B2 (en) | 2013-03-11 | 2017-12-12 | Biomet Manufacturing, Llc | Patient-specific glenoid guide with a reusable guide holder |
US9579107B2 (en) | 2013-03-12 | 2017-02-28 | Biomet Manufacturing, Llc | Multi-point fit for patient specific guide |
US9826981B2 (en) | 2013-03-13 | 2017-11-28 | Biomet Manufacturing, Llc | Tangential fit of patient-specific guides |
US9498233B2 (en) | 2013-03-13 | 2016-11-22 | Biomet Manufacturing, Llc. | Universal acetabular guide and associated hardware |
US11086970B2 (en) * | 2013-03-13 | 2021-08-10 | Blue Belt Technologies, Inc. | Systems and methods for using generic anatomy models in surgical planning |
US10149617B2 (en) | 2013-03-15 | 2018-12-11 | i4c Innovations Inc. | Multiple sensors for monitoring health and wellness of an animal |
CA2906476A1 (en) * | 2013-03-15 | 2014-09-25 | Jointvue, Llc | Determination of joint condition based on vibration analysis |
ES2900643T3 (en) * | 2013-03-15 | 2022-03-17 | Joint Vue Llc | Motion tracking system with inertia-based detector units |
US9517145B2 (en) | 2013-03-15 | 2016-12-13 | Biomet Manufacturing, Llc | Guide alignment system and method |
US9936898B2 (en) | 2013-03-18 | 2018-04-10 | Orthosensor Inc. | Reference position tool for the muscular-skeletal system and method therefor |
US11793424B2 (en) | 2013-03-18 | 2023-10-24 | Orthosensor, Inc. | Kinetic assessment and alignment of the muscular-skeletal system and method therefor |
US10940311B2 (en) | 2013-03-29 | 2021-03-09 | Neurometrix, Inc. | Apparatus and method for button-free control of a wearable transcutaneous electrical nerve stimulator using interactive gestures and other means |
US9439797B2 (en) * | 2013-04-08 | 2016-09-13 | Elwha Llc | Apparatus, system, and method for controlling movement of an orthopedic joint prosthesis in a mammalian subject |
US10420666B2 (en) | 2013-04-08 | 2019-09-24 | Elwha Llc | Apparatus, system, and method for controlling movement of an orthopedic joint prosthesis in a mammalian subject |
WO2014172381A1 (en) | 2013-04-15 | 2014-10-23 | Neurometrix Inc. | Transcutaneous electrical nerves stimulator with automatic detection of user sleep-wake state |
US9417091B2 (en) * | 2013-05-13 | 2016-08-16 | The Johns Hopkins University | System and method for determining and correcting field sensors errors |
US20150124566A1 (en) | 2013-10-04 | 2015-05-07 | Thalmic Labs Inc. | Systems, articles and methods for wearable electronic devices employing contact sensors |
US11921471B2 (en) | 2013-08-16 | 2024-03-05 | Meta Platforms Technologies, Llc | Systems, articles, and methods for wearable devices having secondary power sources in links of a band for providing secondary power in addition to a primary power source |
US10042422B2 (en) | 2013-11-12 | 2018-08-07 | Thalmic Labs Inc. | Systems, articles, and methods for capacitive electromyography sensors |
US9844359B2 (en) | 2013-09-13 | 2017-12-19 | Decision Sciences Medical Company, LLC | Coherent spread-spectrum coded waveforms in synthetic aperture image formation |
WO2015052961A1 (en) * | 2013-10-08 | 2015-04-16 | 株式会社ソニー・コンピュータエンタテインメント | Information processing device |
US20150112349A1 (en) | 2013-10-21 | 2015-04-23 | Biomet Manufacturing, Llc | Ligament Guide Registration |
FR3012314B1 (en) * | 2013-10-30 | 2018-08-17 | Voice | DEVICE FOR EVALUATING THE MOBILITY OF AN ARTICULATION |
GB2519987B (en) * | 2013-11-04 | 2021-03-03 | Imperial College Innovations Ltd | Biomechanical activity monitoring |
WO2015081113A1 (en) | 2013-11-27 | 2015-06-04 | Cezar Morun | Systems, articles, and methods for electromyography sensors |
US10258256B2 (en) | 2014-12-09 | 2019-04-16 | TechMah Medical | Bone reconstruction and orthopedic implants |
AU2014363945B2 (en) * | 2013-12-09 | 2019-04-04 | Techmah Medical Llc | Bone reconstruction and orthopedic implants |
WO2015095383A1 (en) * | 2013-12-17 | 2015-06-25 | The Regents Of The University Of California | Diagnostic knee arthrometer for detecting acl structural changes |
JP6278577B2 (en) * | 2014-03-12 | 2018-02-14 | 古野電気株式会社 | Ultrasonic diagnostic apparatus and index calculation method |
WO2015148391A1 (en) | 2014-03-24 | 2015-10-01 | Thomas Michael Ernst | Systems, methods, and devices for removing prospective motion correction from medical imaging scans |
US10274509B1 (en) | 2014-04-09 | 2019-04-30 | Inertialwave | Inertial motion tracking device |
US10993639B2 (en) * | 2014-04-25 | 2021-05-04 | Massachusetts Institute Of Technology | Feedback method and wearable device to monitor and modulate knee adduction moment |
US10282488B2 (en) | 2014-04-25 | 2019-05-07 | Biomet Manufacturing, Llc | HTO guide with optional guided ACL/PCL tunnels |
US9408616B2 (en) | 2014-05-12 | 2016-08-09 | Biomet Manufacturing, Llc | Humeral cut guide |
EP3443925B1 (en) * | 2014-05-14 | 2021-02-24 | Stryker European Holdings I, LLC | Processor arrangement for tracking the position of a work target |
US9839436B2 (en) | 2014-06-03 | 2017-12-12 | Biomet Manufacturing, Llc | Patient-specific glenoid depth control |
US9575560B2 (en) | 2014-06-03 | 2017-02-21 | Google Inc. | Radar-based gesture-recognition through a wearable device |
US9561040B2 (en) | 2014-06-03 | 2017-02-07 | Biomet Manufacturing, Llc | Patient-specific glenoid depth control |
US9880632B2 (en) | 2014-06-19 | 2018-01-30 | Thalmic Labs Inc. | Systems, devices, and methods for gesture identification |
WO2016007936A2 (en) * | 2014-07-10 | 2016-01-14 | Mahfouz Mohamed R | Bone reconstruction and orthopedic implants |
US9734589B2 (en) | 2014-07-23 | 2017-08-15 | Kineticor, Inc. | Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan |
US9423318B2 (en) * | 2014-07-29 | 2016-08-23 | Honeywell International Inc. | Motion detection devices and systems |
US9811164B2 (en) | 2014-08-07 | 2017-11-07 | Google Inc. | Radar-based gesture sensing and data transmission |
US9778749B2 (en) | 2014-08-22 | 2017-10-03 | Google Inc. | Occluded gesture recognition |
US11169988B2 (en) | 2014-08-22 | 2021-11-09 | Google Llc | Radar recognition-aided search |
EP3185761B1 (en) * | 2014-08-29 | 2021-03-24 | Bionic Skins LLC | Instrument for determining viscoelastic properties of a body segment of a person |
US9950194B2 (en) | 2014-09-09 | 2018-04-24 | Mevion Medical Systems, Inc. | Patient positioning system |
US20170245830A1 (en) * | 2014-09-19 | 2017-08-31 | Think Surgical, Inc. | System and process for ultrasonic determination of long bone orientation |
US9826994B2 (en) | 2014-09-29 | 2017-11-28 | Biomet Manufacturing, Llc | Adjustable glenoid pin insertion guide |
US9833245B2 (en) | 2014-09-29 | 2017-12-05 | Biomet Sports Medicine, Llc | Tibial tubercule osteotomy |
US9600080B2 (en) | 2014-10-02 | 2017-03-21 | Google Inc. | Non-line-of-sight radar-based gesture recognition |
JP2017536954A (en) * | 2014-11-04 | 2017-12-14 | オステオイド サグリク テクノロジレリ アー.エス. | Method for integrating sensors and effectors in a custom 3D brace |
US10191544B2 (en) | 2014-12-06 | 2019-01-29 | Horsemoon Llc | Hand gesture recognition system for controlling electronically controlled devices |
US9648457B2 (en) * | 2014-12-16 | 2017-05-09 | Intel Corporation | Multi-signal geometric location sensing for access control |
KR102256676B1 (en) * | 2015-02-06 | 2021-05-26 | 삼성전자주식회사 | Multi-purpose device including mobile terminal and sensing device using radio-wave based sensor |
US10363149B2 (en) | 2015-02-20 | 2019-07-30 | OrthAlign, Inc. | Hip replacement navigation system and method |
US10743838B2 (en) | 2015-02-25 | 2020-08-18 | Decision Sciences Medical Company, LLC | Acoustic signal transmission couplants and coupling mediums |
WO2016146817A1 (en) * | 2015-03-19 | 2016-09-22 | Meloq Ab | Method and device for anatomical angle measurement |
US11272879B2 (en) | 2015-03-23 | 2022-03-15 | Consensus Orthopedics, Inc. | Systems and methods using a wearable device for monitoring an orthopedic implant and rehabilitation |
US11684260B2 (en) | 2015-03-23 | 2023-06-27 | Tracpatch Health, Inc. | System and methods with user interfaces for monitoring physical therapy and rehabilitation |
US10582891B2 (en) | 2015-03-23 | 2020-03-10 | Consensus Orthopedics, Inc. | System and methods for monitoring physical therapy and rehabilitation of joints |
WO2016154230A1 (en) * | 2015-03-23 | 2016-09-29 | Consensus Orthopedics, Inc. | Systems and methods for monitoring an orthopedic implant and rehabilitation |
JP6879927B2 (en) * | 2015-03-26 | 2021-06-02 | バイオメット マニュファクチャリング,リミティド ライアビリティ カンパニー | A system for planning and performing surgical procedures |
US9820868B2 (en) | 2015-03-30 | 2017-11-21 | Biomet Manufacturing, Llc | Method and apparatus for a pin apparatus |
US10310620B2 (en) | 2015-04-30 | 2019-06-04 | Google Llc | Type-agnostic RF signal representations |
EP3289434A1 (en) | 2015-04-30 | 2018-03-07 | Google LLC | Wide-field radar-based gesture recognition |
KR102002112B1 (en) | 2015-04-30 | 2019-07-19 | 구글 엘엘씨 | RF-based micro-motion tracking for gesture tracking and recognition |
EP3297520B1 (en) * | 2015-05-18 | 2022-11-02 | Vayu Technology Corp. | Devices for measuring human gait and related methods of use |
WO2016191753A1 (en) | 2015-05-27 | 2016-12-01 | Georgia Tech Research Corporation | Wearable technologies for joint health assessment |
US10088908B1 (en) | 2015-05-27 | 2018-10-02 | Google Llc | Gesture detection and interactions |
WO2016191813A1 (en) * | 2015-06-01 | 2016-12-08 | Latey Penelope Jane | Foot muscle biofeedback unit |
US20170225033A1 (en) * | 2015-06-23 | 2017-08-10 | Ipcomm Llc | Method and Apparatus for Analysis of Gait and to Provide Haptic and Visual Corrective Feedback |
US10568647B2 (en) | 2015-06-25 | 2020-02-25 | Biomet Manufacturing, Llc | Patient-specific humeral guide designs |
US10226262B2 (en) | 2015-06-25 | 2019-03-12 | Biomet Manufacturing, Llc | Patient-specific humeral guide designs |
CA2934366A1 (en) * | 2015-06-30 | 2016-12-30 | Ulterra Drilling Technologies, L.P. | Universal joint |
WO2017007518A1 (en) * | 2015-07-07 | 2017-01-12 | Obma Padraic R | Noninvasive medical monitoring device, system and method |
US9943247B2 (en) | 2015-07-28 | 2018-04-17 | The University Of Hawai'i | Systems, devices, and methods for detecting false movements for motion correction during a medical imaging scan |
US20180289354A1 (en) * | 2015-09-30 | 2018-10-11 | Koninklijke Philips N.V. | Ultrasound apparatus and method for determining a medical condition of a subject |
US10817065B1 (en) | 2015-10-06 | 2020-10-27 | Google Llc | Gesture recognition using multiple antenna |
AU2016334258B2 (en) | 2015-10-08 | 2021-07-01 | Decision Sciences Medical Company, LLC | Acoustic orthopedic tracking system and methods |
JP6449753B2 (en) * | 2015-11-05 | 2019-01-09 | 国立大学法人佐賀大学 | Joint inflammation detection device |
CN105468896B (en) * | 2015-11-13 | 2017-06-16 | 上海逸动医学科技有限公司 | Joint motions detecting system and method |
CN105902274B (en) * | 2016-04-08 | 2017-08-25 | 上海逸动医学科技有限公司 | Knee joint dynamic assessment method and system |
EP3171286A1 (en) * | 2015-11-17 | 2017-05-24 | Universitat De València, Estudi General | Methods for determining an identifier for use in methods for diagnosing haemophilic arthropathy, methods and apparatus for diagnosing |
US10716515B2 (en) | 2015-11-23 | 2020-07-21 | Kineticor, Inc. | Systems, devices, and methods for tracking and compensating for patient motion during a medical imaging scan |
EP3878367A1 (en) * | 2015-12-08 | 2021-09-15 | Kneevoice, Inc. | Assessing joint condition using acoustic sensors |
US10467534B1 (en) * | 2015-12-09 | 2019-11-05 | Roger Brent | Augmented reality procedural system |
ES2931044T3 (en) | 2015-12-16 | 2022-12-23 | Techmah Medical Llc | A method of calibrating an inertial measurement unit |
US10463279B2 (en) * | 2016-02-19 | 2019-11-05 | Trustees Of Dartmouth College | Movement monitoring systems and methods |
EP3422951B1 (en) * | 2016-02-29 | 2024-05-22 | Mohamed R. Mahfouz | Connected healthcare environment |
WO2017185108A2 (en) * | 2016-04-28 | 2017-11-02 | Medfit Beratungs-Und Beteiligunges.M.B.H | Dynamic ligament balancing system (dlb) |
WO2017192167A1 (en) | 2016-05-03 | 2017-11-09 | Google Llc | Connecting an electronic component to an interactive textile |
JP2017202236A (en) * | 2016-05-13 | 2017-11-16 | 花王株式会社 | Gait analysis method and gait analysis device |
WO2017200949A1 (en) | 2016-05-16 | 2017-11-23 | Google Llc | Interactive fabric |
US10314514B2 (en) * | 2016-05-29 | 2019-06-11 | Ankon Medical Technologies (Shanghai) Co., Ltd. | System and method for using a capsule device |
WO2017205983A1 (en) * | 2016-06-02 | 2017-12-07 | Bigmotion Technologies Inc. | Systems and methods for walking speed estimation |
US10078377B2 (en) | 2016-06-09 | 2018-09-18 | Microsoft Technology Licensing, Llc | Six DOF mixed reality input by fusing inertial handheld controller with hand tracking |
CN107510466B (en) * | 2016-06-15 | 2022-04-12 | 中慧医学成像有限公司 | Three-dimensional imaging method and system |
US10849551B2 (en) | 2016-06-24 | 2020-12-01 | Surgical Sensors Bvba | Integrated ligament strain measurement |
WO2017220173A1 (en) * | 2016-06-24 | 2017-12-28 | Surgical Sensors Bvba | Integrated ligament strain measurement |
US20170367644A1 (en) * | 2016-06-27 | 2017-12-28 | Claris Healthcare Inc. | Apparatus and Method for Monitoring Rehabilitation from Joint Surgery |
BR112019000548A2 (en) | 2016-07-13 | 2019-05-21 | GSK Consumer Healthcare S.A. | apparatus and method for controlling transcutaneous electrical neurostimulation. |
US11331045B1 (en) | 2018-01-25 | 2022-05-17 | Facebook Technologies, Llc | Systems and methods for mitigating neuromuscular signal artifacts |
EP3487395A4 (en) | 2016-07-25 | 2020-03-04 | CTRL-Labs Corporation | Methods and apparatus for predicting musculo-skeletal position information using wearable autonomous sensors |
WO2018022657A1 (en) | 2016-07-25 | 2018-02-01 | Ctrl-Labs Corporation | System and method for measuring the movements of articulated rigid bodies |
US10687759B2 (en) | 2018-05-29 | 2020-06-23 | Facebook Technologies, Llc | Shielding techniques for noise reduction in surface electromyography signal measurement and related systems and methods |
CN110312471B (en) | 2016-07-25 | 2022-04-29 | 脸谱科技有限责任公司 | Adaptive system for deriving control signals from neuromuscular activity measurements |
EP3487402B1 (en) | 2016-07-25 | 2021-05-05 | Facebook Technologies, LLC | Methods and apparatus for inferring user intent based on neuromuscular signals |
US10772519B2 (en) | 2018-05-25 | 2020-09-15 | Facebook Technologies, Llc | Methods and apparatus for providing sub-muscular control |
US11216069B2 (en) | 2018-05-08 | 2022-01-04 | Facebook Technologies, Llc | Systems and methods for improved speech recognition using neuromuscular information |
US20180028109A1 (en) * | 2016-07-27 | 2018-02-01 | Andrew TESNOW | System and method for a wearable knee injury prevention |
EP3508120B1 (en) | 2016-08-30 | 2022-06-08 | Fujitsu Limited | Information processing device, information processing system and information processing method |
CN106447713B (en) * | 2016-08-31 | 2019-05-28 | 北京维盛视通科技有限公司 | Method for automatic measurement and device based on cloud manikin |
JP6738250B2 (en) * | 2016-09-09 | 2020-08-12 | 花王株式会社 | Gait analysis method and gait analysis device |
JP6738249B2 (en) * | 2016-09-09 | 2020-08-12 | 花王株式会社 | Gait analysis method and gait analysis device |
WO2018051898A1 (en) * | 2016-09-14 | 2018-03-22 | Cyberdyne株式会社 | Device for producing knee joint correction tool, method for producing knee joint correction tool, device for assisting knee joint treatment, and method for assissting knee joint treatment |
WO2018081795A1 (en) | 2016-10-31 | 2018-05-03 | Zipline Medical, Inc. | Systems and methods for monitoring physical therapy of the knee and other joints |
WO2018085822A1 (en) | 2016-11-07 | 2018-05-11 | Synergistic Biosensors, LLC | Systems and methods for monitoring implantable devices for detection of implant failure utilizing wireless in vivo micro sensors |
FI127689B (en) * | 2016-11-07 | 2018-12-14 | Oulun Yliopisto | Arrangement for knee diagnostics |
US10579150B2 (en) | 2016-12-05 | 2020-03-03 | Google Llc | Concurrent detection of absolute distance and relative movement for sensing action gestures |
RU2019119439A (en) | 2016-12-23 | 2021-01-26 | ДжиЭсКей Консьюмер Хелткер С.А. | "Intelligent" electrode node for transcutaneous electrical nerve stimulation (TESN) |
US10120455B2 (en) * | 2016-12-28 | 2018-11-06 | Industrial Technology Research Institute | Control device and control method |
CN106725598B (en) * | 2016-12-28 | 2023-09-12 | 苏州科技城医院 | Heart ultrasonic system based on multiple percutaneous ultrasonic transducers and imaging method |
EP3568065B1 (en) * | 2017-01-10 | 2024-03-13 | Braindex S.A.S | Physiological sensor for near-infrared spectroscopy at various depths |
US11259743B2 (en) * | 2017-03-08 | 2022-03-01 | Strive Orthopedics, Inc. | Method for identifying human joint characteristics |
US10722310B2 (en) | 2017-03-13 | 2020-07-28 | Zimmer Biomet CMF and Thoracic, LLC | Virtual surgery planning system and method |
US10863995B2 (en) | 2017-03-14 | 2020-12-15 | OrthAlign, Inc. | Soft tissue measurement and balancing systems and methods |
WO2018169995A1 (en) | 2017-03-14 | 2018-09-20 | OrthAlign, Inc. | Hip replacement navigation systems and methods |
GB2560909B (en) * | 2017-03-27 | 2020-12-02 | 270 Vision Ltd | Movement sensor |
EP3606459B1 (en) * | 2017-04-07 | 2024-05-01 | Orthosoft ULC | Non-invasive system and method for tracking bones |
CN110545759B (en) * | 2017-05-18 | 2023-01-10 | 史密夫和内修有限公司 | System and method for determining the position and orientation of a joint replacement surgical implant |
US11058877B2 (en) | 2017-05-30 | 2021-07-13 | Neurometrix, Inc. | Apparatus and method for the automated control of transcutaneous electrical nerve stimulation based on current and forecasted weather conditions |
JP7227168B2 (en) | 2017-06-19 | 2023-02-21 | モハメド・アール・マーフーズ | Surgical Navigation of the Hip Using Fluoroscopy and Tracking Sensors |
US11000229B2 (en) * | 2017-08-03 | 2021-05-11 | Orthini, LLC | Systems, methods, and apparatuses for integrating a body joint rehabilitation regimen with a wearable movement capture device operable in conjunction with a cloud based computing environment |
AU2018332792A1 (en) | 2017-09-14 | 2020-05-07 | Howmedica Osteonics Corp. | Non-symmetrical insert sensing system and method therefor |
USD865986S1 (en) | 2017-09-21 | 2019-11-05 | Neurometrix, Inc. | Transcutaneous electrical nerve stimulation device strap |
GB201716123D0 (en) * | 2017-10-03 | 2017-11-15 | Virtualclinic Direct Ltd | Data capture device |
EP3697297A4 (en) | 2017-10-19 | 2020-12-16 | Facebook Technologies, Inc. | Systems and methods for identifying biological structures associated with neuromuscular source signals |
CA3079503A1 (en) | 2017-11-07 | 2019-05-16 | Djo, Llc | Brace having integrated remote patient monitoring technology and method of using same |
EP3551137A4 (en) | 2017-11-30 | 2020-08-26 | Bruin Biometrics, LLC | Implant evaluation using acoustic emissions |
CN108030512A (en) * | 2017-12-21 | 2018-05-15 | 福州大学 | A kind of supersonic array measuring method of ankle arthrosis degree of injury |
CN108245164B (en) * | 2017-12-22 | 2021-03-26 | 北京精密机电控制设备研究所 | Human body gait information acquisition and calculation method for wearable inertial device |
US11110281B2 (en) | 2018-01-04 | 2021-09-07 | Cardiac Pacemakers, Inc. | Secure transdermal communication with implanted device |
CN108175381A (en) * | 2018-01-10 | 2018-06-19 | 中山大学附属第医院 | A kind of knee joint endoprosthesis surface damage detecting system and its application method |
US10706693B1 (en) * | 2018-01-11 | 2020-07-07 | Facebook Technologies, Llc. | Haptic device for creating vibration-, pressure-, and shear-based haptic cues |
US11493993B2 (en) | 2019-09-04 | 2022-11-08 | Meta Platforms Technologies, Llc | Systems, methods, and interfaces for performing inputs based on neuromuscular control |
CN112005198A (en) | 2018-01-25 | 2020-11-27 | 脸谱科技有限责任公司 | Hand state reconstruction based on multiple inputs |
US10460455B2 (en) | 2018-01-25 | 2019-10-29 | Ctrl-Labs Corporation | Real-time processing of handstate representation model estimates |
WO2019148002A1 (en) | 2018-01-25 | 2019-08-01 | Ctrl-Labs Corporation | Techniques for anonymizing neuromuscular signal data |
US11907423B2 (en) | 2019-11-25 | 2024-02-20 | Meta Platforms Technologies, Llc | Systems and methods for contextualized interactions with an environment |
CN112074870A (en) | 2018-01-25 | 2020-12-11 | 脸谱科技有限责任公司 | Visualization of reconstructed hand state information |
US11481030B2 (en) | 2019-03-29 | 2022-10-25 | Meta Platforms Technologies, Llc | Methods and apparatus for gesture detection and classification |
WO2019147958A1 (en) | 2018-01-25 | 2019-08-01 | Ctrl-Labs Corporation | User-controlled tuning of handstate representation model parameters |
US11150730B1 (en) | 2019-04-30 | 2021-10-19 | Facebook Technologies, Llc | Devices, systems, and methods for controlling computing devices via neuromuscular signals of users |
US11961494B1 (en) | 2019-03-29 | 2024-04-16 | Meta Platforms Technologies, Llc | Electromagnetic interference reduction in extended reality environments |
US10970936B2 (en) | 2018-10-05 | 2021-04-06 | Facebook Technologies, Llc | Use of neuromuscular signals to provide enhanced interactions with physical objects in an augmented reality environment |
US10937414B2 (en) | 2018-05-08 | 2021-03-02 | Facebook Technologies, Llc | Systems and methods for text input using neuromuscular information |
EP3742961A4 (en) | 2018-01-25 | 2021-03-31 | Facebook Technologies, Inc. | Calibration techniques for handstate representation modeling using neuromuscular signals |
WO2019152566A1 (en) * | 2018-01-30 | 2019-08-08 | The Regents Of The University Of California | Systems and methods for subject specific kinematic mapping |
US10592001B2 (en) | 2018-05-08 | 2020-03-17 | Facebook Technologies, Llc | Systems and methods for improved speech recognition using neuromuscular information |
WO2019241701A1 (en) | 2018-06-14 | 2019-12-19 | Ctrl-Labs Corporation | User identification and authentication with neuromuscular signatures |
EP3810008A2 (en) * | 2018-06-19 | 2021-04-28 | Tornier, Inc. | Mixed reality-aided depth tracking in orthopedic surgical procedures |
US10940021B2 (en) * | 2018-06-20 | 2021-03-09 | Techmah Medical Llc | Methods and devices for knee surgery with inertial sensors |
US11510737B2 (en) | 2018-06-21 | 2022-11-29 | Mako Surgical Corp. | Patella tracking |
CN112424573A (en) * | 2018-06-22 | 2021-02-26 | 尹迪泰特有限责任公司 | Sensor device, use of a sensor device and method for detecting solid noise |
CA3104181A1 (en) * | 2018-06-26 | 2020-01-02 | Australian Institute of Robotic Orthopaedics Pty Ltd | Implant fit analysis |
US11045137B2 (en) | 2018-07-19 | 2021-06-29 | Facebook Technologies, Llc | Methods and apparatus for improved signal robustness for a wearable neuromuscular recording device |
GB2574074B (en) | 2018-07-27 | 2020-05-20 | Mclaren Applied Tech Ltd | Time synchronisation |
WO2020036958A1 (en) | 2018-08-13 | 2020-02-20 | Ctrl-Labs Corporation | Real-time spike detection and identification |
EP4241661A1 (en) | 2018-08-31 | 2023-09-13 | Facebook Technologies, LLC | Camera-guided interpretation of neuromuscular signals |
EP3853698A4 (en) | 2018-09-20 | 2021-11-17 | Facebook Technologies, LLC | Neuromuscular text entry, writing and drawing in augmented reality systems |
US10921764B2 (en) | 2018-09-26 | 2021-02-16 | Facebook Technologies, Llc | Neuromuscular control of physical objects in an environment |
JP7132816B2 (en) * | 2018-10-10 | 2022-09-07 | 大和ハウス工業株式会社 | Joint condition determination system |
US11510035B2 (en) | 2018-11-07 | 2022-11-22 | Kyle Craig | Wearable device for measuring body kinetics |
SG10201810156PA (en) * | 2018-11-14 | 2020-06-29 | Prec Medical Pte Ltd | Method and device for measuring anatomical movement of a joint |
CN109859592B (en) * | 2018-11-14 | 2020-12-08 | 华中科技大学 | Soft tissue injury simulation test device |
CN113423341A (en) | 2018-11-27 | 2021-09-21 | 脸谱科技有限责任公司 | Method and apparatus for automatic calibration of wearable electrode sensor system |
US11883661B2 (en) | 2018-12-07 | 2024-01-30 | Neurometrix, Inc. | Intelligent determination of therapeutic stimulation intensity for transcutaneous electrical nerve stimulation |
KR102550854B1 (en) * | 2018-12-13 | 2023-07-04 | 삼성전자주식회사 | Method and device for assisting walking |
CN111374674B (en) * | 2018-12-29 | 2023-02-10 | 西安思博探声生物科技有限公司 | Knee joint movement information processing equipment |
CN109801278B (en) * | 2019-01-21 | 2023-05-30 | 燕山大学 | Surface damage classification method for high-speed sliding electric contact movement pair |
US10905383B2 (en) | 2019-02-28 | 2021-02-02 | Facebook Technologies, Llc | Methods and apparatus for unsupervised one-shot machine learning for classification of human gestures and estimation of applied forces |
US20220071791A1 (en) * | 2019-03-20 | 2022-03-10 | Optima Molliter Srl | Orthosis brace with monitoring system |
US11154274B2 (en) | 2019-04-23 | 2021-10-26 | Decision Sciences Medical Company, LLC | Semi-rigid acoustic coupling articles for ultrasound diagnostic and treatment applications |
US11537702B2 (en) * | 2019-05-13 | 2022-12-27 | Cardiac Pacemakers, Inc. | Implanted medical device authentication based on comparison of internal IMU signal to external IMU signal |
US11911213B2 (en) | 2019-06-03 | 2024-02-27 | General Electric Company | Techniques for determining ultrasound probe motion |
US11497452B2 (en) * | 2019-06-20 | 2022-11-15 | The Hong Kong Polytechnic University | Predictive knee joint loading system |
KR102251925B1 (en) * | 2019-07-18 | 2021-05-13 | 경상국립대학교 산학협력단 | Apparatus and application for predicting musculoskeletal disorders |
US11877870B2 (en) | 2019-08-05 | 2024-01-23 | Consultation Semperform Inc | Systems, methods and apparatus for prevention of injury |
US11812978B2 (en) | 2019-10-15 | 2023-11-14 | Orthosensor Inc. | Knee balancing system using patient specific instruments |
GB2588237B (en) * | 2019-10-18 | 2023-12-27 | Mclaren Applied Ltd | Joint axis direction estimation |
GB2588236B (en) | 2019-10-18 | 2024-03-20 | Mclaren Applied Ltd | Gyroscope bias estimation |
US10842415B1 (en) * | 2019-10-25 | 2020-11-24 | Plethy, Inc. | Devices, systems, and methods for monitoring and assessing gait, stability, and/or balance of a user |
JP7037159B2 (en) * | 2019-11-08 | 2022-03-16 | 国立大学法人大阪大学 | Systems, programs, and methods for measuring a subject's jaw movements |
CN110772262B (en) * | 2019-12-05 | 2020-12-29 | 广东电网有限责任公司 | Comfort evaluation method for human body tower-climbing posture |
JP2023504213A (en) * | 2019-12-09 | 2023-02-01 | オースアライン・インコーポレイテッド | Cup alignment system and method |
US10863928B1 (en) | 2020-01-28 | 2020-12-15 | Consensus Orthopedics, Inc. | System and methods for monitoring the spine, balance, gait, or posture of a patient |
JP2023522918A (en) * | 2020-04-20 | 2023-06-01 | フォームス ラブス リミティッド | surgical system |
US11832934B1 (en) | 2020-05-04 | 2023-12-05 | Qingbin Zheng | Joint monitoring |
US20210378853A1 (en) * | 2020-06-09 | 2021-12-09 | National Cheng Kung University | Wearable interface for intelligent health promotion service system |
WO2022011054A1 (en) * | 2020-07-07 | 2022-01-13 | The General Hospital Corporation | Evaluating the stability of a joint in the foot and ankle complex via weight-bearing medical imaging |
CA3202517A1 (en) | 2020-11-13 | 2022-05-19 | Decision Sciences Medical Company, LLC | Systems and methods for synthetic aperture ultrasound imaging of an object |
EP4016536A1 (en) * | 2020-12-16 | 2022-06-22 | Polar Electro Oy | Biomechanical modelling of motion measurements |
CN113143256B (en) * | 2021-01-28 | 2023-09-26 | 上海电气集团股份有限公司 | Gait feature extraction method, lower limb evaluation and control method, device and medium |
CN112754516B (en) * | 2021-02-07 | 2022-04-08 | 河南省肿瘤医院 | Intelligent bowel sound positioning and collecting device |
JP2024512398A (en) * | 2021-03-07 | 2024-03-19 | リキッド ワイヤ エルエルシー | Apparatus, system, and method for monitoring and characterizing user behavior with flexible circuits |
US11868531B1 (en) | 2021-04-08 | 2024-01-09 | Meta Platforms Technologies, Llc | Wearable device providing for thumb-to-finger-based input gestures detected based on neuromuscular signals, and systems and methods of use thereof |
WO2022266254A1 (en) * | 2021-06-16 | 2022-12-22 | Kinisi Inc | Wearable imaging system for measuring bone displacement |
IT202100017267A1 (en) * | 2021-06-30 | 2022-12-30 | Scuola Superiore Santanna | POSITIONING DEVICE FOR ULTRASONIC PROBE |
WO2023031642A1 (en) * | 2021-08-30 | 2023-03-09 | Siemens Industry Software Ltd. | Method and system for determining a joint in a virtual kinematic device |
DE102021124873A1 (en) * | 2021-09-27 | 2023-03-30 | Aesculap Ag | Medical technology system and method |
WO2023148427A1 (en) * | 2022-02-03 | 2023-08-10 | Aikoa Technologies Oy | Method for training computing arrangement to provide prognosis of progression of tissue condition |
WO2023235859A2 (en) * | 2022-06-02 | 2023-12-07 | New York Society For The Relief Of The Ruptured And Crippled, Maintaining The Hospital For Special Surgery | Method for mechanical phenotyping of knees |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070249967A1 (en) * | 2006-03-21 | 2007-10-25 | Perception Raisonnement Action En Medecine | Computer-aided osteoplasty surgery system |
US20070270680A1 (en) * | 2006-03-22 | 2007-11-22 | Garrett Sheffer | Modeling method and apparatus for use in surgical navigation |
Family Cites Families (90)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3881164A (en) * | 1973-09-13 | 1975-04-29 | Commw Of Australia | Cross array ultrasonic transducer |
US4016750B1 (en) * | 1975-11-06 | 1994-04-05 | Stanford Research Inst | Ultrasonic imaging method and apparatus |
US5488952A (en) | 1982-02-24 | 1996-02-06 | Schoolman Scientific Corp. | Stereoscopically display three dimensional ultrasound imaging |
US4476873A (en) | 1982-09-03 | 1984-10-16 | Medtronic, Inc. | Ultrasound scanning system for skeletal imaging |
FR2694881B1 (en) * | 1992-07-31 | 1996-09-06 | Univ Joseph Fourier | METHOD FOR DETERMINING THE POSITION OF AN ORGAN. |
US6005916A (en) | 1992-10-14 | 1999-12-21 | Techniscan, Inc. | Apparatus and method for imaging with wavefields using inverse scattering techniques |
US5413116A (en) * | 1993-06-24 | 1995-05-09 | Bioresearch | Method and apparatus for diagnosing joints |
US5394875A (en) | 1993-10-21 | 1995-03-07 | Lewis; Judith T. | Automatic ultrasonic localization of targets implanted in a portion of the anatomy |
US5826578A (en) * | 1994-05-26 | 1998-10-27 | Curchod; Donald B. | Motion measurement apparatus |
US5919149A (en) * | 1996-03-19 | 1999-07-06 | Allum; John H. | Method and apparatus for angular position and velocity based determination of body sway for the diagnosis and rehabilitation of balance and gait disorders |
US5806521A (en) | 1996-03-26 | 1998-09-15 | Sandia Corporation | Composite ultrasound imaging apparatus and method |
US5771310A (en) | 1996-12-30 | 1998-06-23 | Shriners Hospitals For Children | Method and apparatus for recording three-dimensional topographies |
US6213958B1 (en) | 1996-08-29 | 2001-04-10 | Alan A. Winder | Method and apparatus for the acoustic emission monitoring detection, localization, and classification of metabolic bone disease |
US7468075B2 (en) * | 2001-05-25 | 2008-12-23 | Conformis, Inc. | Methods and compositions for articular repair |
US6205411B1 (en) * | 1997-02-21 | 2001-03-20 | Carnegie Mellon University | Computer-assisted surgery planner and intra-operative guidance system |
JPH10243937A (en) * | 1997-03-07 | 1998-09-14 | Mitsubishi Heavy Ind Ltd | Joint labile measurement device |
US6120453A (en) | 1997-11-17 | 2000-09-19 | Sharp; William A. | Three-dimensional ultrasound system based on the coordination of multiple ultrasonic transducers |
US6231585B1 (en) | 1997-11-20 | 2001-05-15 | Medivas, Llc | Device for stabilizing a treatment site and method of use |
DE69934215T2 (en) * | 1998-03-20 | 2007-10-25 | Barbara Ann Karmanos Cancer Institute, Detroit | Multi-dimensional identification and characterization of pathological tissues |
US6280387B1 (en) | 1998-05-06 | 2001-08-28 | Siemens Medical Systems, Inc. | Three-dimensional tissue/flow ultrasound imaging system |
US7239908B1 (en) * | 1998-09-14 | 2007-07-03 | The Board Of Trustees Of The Leland Stanford Junior University | Assessing the condition of a joint and devising treatment |
DE69922317D1 (en) | 1998-09-29 | 2005-01-05 | Koninkl Philips Electronics Nv | Image processing method for ultrasonic medical images of the bone structure, and a computer-aided surgery device |
JP2000251078A (en) * | 1998-12-22 | 2000-09-14 | Atr Media Integration & Communications Res Lab | Method and device for estimating three-dimensional posture of person, and method and device for estimating position of elbow of person |
US6106464A (en) | 1999-02-22 | 2000-08-22 | Vanderbilt University | Apparatus and method for bone surface-based registration of physical space with tomographic images and for guiding an instrument relative to anatomical sites in the image |
WO2000063719A1 (en) | 1999-04-20 | 2000-10-26 | Synthes Ag Chur | Device for the percutaneous obtainment of 3d-coordinates on the surface of a human or animal organ |
WO2001032114A1 (en) | 1999-11-02 | 2001-05-10 | Wizcare Ltd. | Skin-gripper |
MXPA02007262A (en) * | 2000-01-29 | 2002-12-09 | Paul E Thomson | Detection and quantification of joint and tissue inflammation. |
US6904123B2 (en) | 2000-08-29 | 2005-06-07 | Imaging Therapeutics, Inc. | Methods and devices for quantitative analysis of x-ray images |
WO2002022013A1 (en) * | 2000-09-14 | 2002-03-21 | The Board Of Trustees Of The Leland Stanford Junior University | Assessing condition of a joint and cartilage loss |
CA2325236A1 (en) | 2000-11-06 | 2002-05-06 | Cyril B. Frank | Auditory display of knee joint vibration signals |
US6561991B2 (en) * | 2000-12-19 | 2003-05-13 | The Research Foundation Of The State University Of New York (Suny) | Non-invasive method and system of quantifying human postural stability |
CA2333224A1 (en) | 2001-01-31 | 2002-07-31 | University Technologies International Inc. | Non-invasive diagnostic method and apparatus for musculoskeletal systems |
US20050165284A1 (en) | 2002-03-25 | 2005-07-28 | Amit Gefen | Method and system for determining a risk of ulcer onset |
US7117026B2 (en) * | 2002-06-12 | 2006-10-03 | Koninklijke Philips Electronics N.V. | Physiological model based non-rigid image registration |
US7981057B2 (en) * | 2002-10-11 | 2011-07-19 | Northrop Grumman Guidance And Electronics Company, Inc. | Joint motion sensing to make a determination of a positional change of an individual |
WO2004062495A2 (en) * | 2003-01-07 | 2004-07-29 | Imaging Therapeutics, Inc. | Methods of predicting musculoskeletal disease |
US7660623B2 (en) | 2003-01-30 | 2010-02-09 | Medtronic Navigation, Inc. | Six degree of freedom alignment display for medical procedures |
JP2004264060A (en) * | 2003-02-14 | 2004-09-24 | Akebono Brake Ind Co Ltd | Error correction method in attitude detector, and action measuring instrument using the same |
JP3932360B2 (en) * | 2003-03-04 | 2007-06-20 | 独立行政法人産業技術総合研究所 | Landmark extraction apparatus and landmark extraction method |
US20050043660A1 (en) * | 2003-03-31 | 2005-02-24 | Izex Technologies, Inc. | Orthoses |
WO2005007217A2 (en) * | 2003-07-10 | 2005-01-27 | Neurocom International, Inc. | Apparatus and method for characterizing contributions of forces associated with a body part of a subject |
US7454242B2 (en) | 2003-09-17 | 2008-11-18 | Elise Fear | Tissue sensing adaptive radar imaging for breast tumor detection |
US20050093859A1 (en) | 2003-11-04 | 2005-05-05 | Siemens Medical Solutions Usa, Inc. | Viewing direction dependent acquisition or processing for 3D ultrasound imaging |
US8265728B2 (en) | 2003-11-26 | 2012-09-11 | University Of Chicago | Automated method and system for the evaluation of disease and registration accuracy in the subtraction of temporally sequential medical images |
US20080269596A1 (en) * | 2004-03-10 | 2008-10-30 | Ian Revie | Orthpaedic Monitoring Systems, Methods, Implants and Instruments |
JP4455118B2 (en) * | 2004-03-30 | 2010-04-21 | 独立行政法人科学技術振興機構 | Delivery diagnosis support program, recording medium storing the program, and delivery diagnosis support method and apparatus. |
US7678052B2 (en) | 2004-04-13 | 2010-03-16 | General Electric Company | Method and apparatus for detecting anatomic structures |
JP4411384B2 (en) * | 2004-04-15 | 2010-02-10 | 独立行政法人放射線医学総合研究所 | Diagnostic system |
US7483732B2 (en) | 2004-04-15 | 2009-01-27 | Boston Scientific Scimed, Inc. | Magnetic resonance imaging of a medical device and proximate body tissue |
DE102004026525A1 (en) * | 2004-05-25 | 2005-12-22 | Aesculap Ag & Co. Kg | Method and device for the non-invasive determination of prominent structures of the human or animal body |
US20060052727A1 (en) | 2004-09-09 | 2006-03-09 | Laurence Palestrant | Activity monitoring device and weight management method utilizing same |
CN100573589C (en) | 2004-09-09 | 2009-12-23 | 皇家飞利浦电子股份有限公司 | The system that is used for the three-dimensional imaging of movable joint |
US20060161052A1 (en) * | 2004-12-08 | 2006-07-20 | Perception Raisonnement Action En Medecine | Computer assisted orthopaedic surgery system for ligament graft reconstruction |
WO2006085387A1 (en) * | 2005-02-08 | 2006-08-17 | Kouki Nagamune | Noninvasive moving body analytic system and its using method |
GB0504172D0 (en) * | 2005-03-01 | 2005-04-06 | King S College London | Surgical planning |
JP4304341B2 (en) * | 2005-03-17 | 2009-07-29 | 国立大学法人 新潟大学 | Three-dimensional shape measurement device and socket design device for prosthetic limbs based on the measurement data |
US20100100011A1 (en) | 2008-10-22 | 2010-04-22 | Martin Roche | System and Method for Orthopedic Alignment and Measurement |
CN101193603B (en) | 2005-06-06 | 2010-11-03 | 直观外科手术公司 | Laparoscopic ultrasound robotic surgical system |
US8092398B2 (en) * | 2005-08-09 | 2012-01-10 | Massachusetts Eye & Ear Infirmary | Multi-axis tilt estimation and fall remediation |
US7529398B2 (en) | 2005-08-09 | 2009-05-05 | Gil Zwirn | High resolution radio frequency medical imaging and therapy system |
GB2435614A (en) | 2006-03-01 | 2007-09-05 | Samuel George | Transducer holder for maintaining signal-receiving contact with a patient's body |
US8676293B2 (en) * | 2006-04-13 | 2014-03-18 | Aecc Enterprises Ltd. | Devices, systems and methods for measuring and evaluating the motion and function of joint structures and associated muscles, determining suitability for orthopedic intervention, and evaluating efficacy of orthopedic intervention |
US20080009722A1 (en) | 2006-05-11 | 2008-01-10 | Constantine Simopoulos | Multi-planar reconstruction for ultrasound volume data |
US7578799B2 (en) * | 2006-06-30 | 2009-08-25 | Ossur Hf | Intelligent orthosis |
US7769422B2 (en) | 2006-09-29 | 2010-08-03 | Depuy Products, Inc. | Apparatus and method for monitoring the position of an orthopaedic prosthesis |
WO2008074151A1 (en) * | 2006-12-20 | 2008-06-26 | Mcmaster University | System and method of assessing the condition of a joint |
US20080194997A1 (en) * | 2007-02-08 | 2008-08-14 | Rehabilitation Institute Of Chicago | System and method for diagnosing and treating patellar maltracking and malalignment |
US20080221487A1 (en) * | 2007-03-07 | 2008-09-11 | Motek Bv | Method for real time interactive visualization of muscle forces and joint torques in the human body |
EP1970005B1 (en) * | 2007-03-15 | 2012-10-03 | Xsens Holding B.V. | A system and a method for motion tracking using a calibration unit |
US7920731B2 (en) | 2007-03-27 | 2011-04-05 | Siemens Medical Solutions Usa, Inc. | Bleeding detection using a blanket ultrasound device |
CA2687116C (en) * | 2007-05-14 | 2015-05-26 | Queen's University At Kingston | Patient-specific surgical guidance tool and method of use |
US8089417B2 (en) | 2007-06-01 | 2012-01-03 | The Royal Institution For The Advancement Of Learning/Mcgill University | Microwave scanning system and miniaturized microwave antenna |
US8771188B2 (en) | 2007-06-20 | 2014-07-08 | Perception Raisonnement Action En Medecine | Ultrasonic bone motion tracking system |
JP5061281B2 (en) * | 2007-08-20 | 2012-10-31 | 国立大学法人広島大学 | Knee joint rotation angle measuring device |
EP2194836B1 (en) | 2007-09-25 | 2015-11-04 | Perception Raisonnement Action En Medecine | Apparatus for assisting cartilage diagnostic and therapeutic procedures |
JP5416900B2 (en) | 2007-11-22 | 2014-02-12 | 株式会社東芝 | Ultrasonic diagnostic apparatus and puncture support control program |
WO2009117832A1 (en) * | 2008-03-25 | 2009-10-01 | Orthosoft Inc. | Tracking system and method |
US8377073B2 (en) * | 2008-04-21 | 2013-02-19 | Ray Wasielewski | Method of designing orthopedic implants using in vivo data |
EP2212994B1 (en) | 2008-05-19 | 2015-10-28 | Hittite Microwave Corporation | Multiple input variable gain amplifier |
US20100125229A1 (en) * | 2008-07-11 | 2010-05-20 | University Of Delaware | Controllable Joint Brace |
US8444564B2 (en) | 2009-02-02 | 2013-05-21 | Jointvue, Llc | Noninvasive diagnostic system |
JP5377166B2 (en) | 2009-09-01 | 2013-12-25 | 古野電気株式会社 | Ultrasound bone analyzer |
US20110125016A1 (en) | 2009-11-25 | 2011-05-26 | Siemens Medical Solutions Usa, Inc. | Fetal rendering in medical diagnostic ultrasound |
US8979758B2 (en) | 2010-06-29 | 2015-03-17 | Orthosensor Inc | Sensing module for orthopedic load sensing insert device |
US20130144135A1 (en) | 2011-08-02 | 2013-06-06 | Mohamed R. Mahfouz | Method and apparatus for three dimensional reconstruction of a joint using ultrasound |
CA2807288C (en) | 2010-08-02 | 2021-10-26 | Joint Vue, LLC | Method and apparatus for three dimensional reconstruction of a joint using ultrasound |
CA2845044C (en) | 2011-08-12 | 2023-03-28 | Jointvue, Llc | 3-d ultrasound imaging device and methods |
CA3194212A1 (en) | 2011-10-14 | 2013-04-18 | Jointvue, Llc | Real-time 3-d ultrasound reconstruction of knee and its implications for patient specific implants and 3-d joint injections |
CA2906476A1 (en) | 2013-03-15 | 2014-09-25 | Jointvue, Llc | Determination of joint condition based on vibration analysis |
ES2900643T3 (en) | 2013-03-15 | 2022-03-17 | Joint Vue Llc | Motion tracking system with inertia-based detector units |
-
2009
- 2009-02-02 US US12/364,267 patent/US8444564B2/en active Active
-
2010
- 2010-02-02 JP JP2011548412A patent/JP5723788B2/en active Active
- 2010-02-02 CA CA2977574A patent/CA2977574C/en active Active
- 2010-02-02 CA CA2751422A patent/CA2751422C/en active Active
- 2010-02-02 CA CA3170396A patent/CA3170396C/en active Active
- 2010-02-02 WO PCT/US2010/022939 patent/WO2010088696A1/en active Application Filing
- 2010-02-02 EP EP10736584.3A patent/EP2391971B1/en active Active
- 2010-02-02 CA CA3192190A patent/CA3192190A1/en active Pending
- 2010-02-02 EP EP21199275.5A patent/EP3968220A1/en active Pending
- 2010-02-02 CA CA3049975A patent/CA3049975C/en active Active
-
2011
- 2011-08-02 US US13/196,701 patent/US20120029345A1/en not_active Abandoned
-
2013
- 2013-03-15 US US13/841,402 patent/US9642572B2/en active Active
- 2013-05-20 US US13/898,092 patent/US11342071B2/en active Active
-
2014
- 2014-12-18 JP JP2014256357A patent/JP6005715B2/en active Active
-
2016
- 2016-09-07 JP JP2016174550A patent/JP6404286B2/en active Active
-
2017
- 2017-04-03 US US15/478,148 patent/US11004561B2/en active Active
-
2021
- 2021-02-22 US US17/181,372 patent/US20210193313A1/en active Pending
-
2022
- 2022-03-25 US US17/704,376 patent/US11776686B2/en active Active
-
2023
- 2023-08-25 US US18/238,338 patent/US11935648B1/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070249967A1 (en) * | 2006-03-21 | 2007-10-25 | Perception Raisonnement Action En Medecine | Computer-aided osteoplasty surgery system |
US20070270680A1 (en) * | 2006-03-22 | 2007-11-22 | Garrett Sheffer | Modeling method and apparatus for use in surgical navigation |
Also Published As
Publication number | Publication date |
---|---|
JP6005715B2 (en) | 2016-10-12 |
US20100198067A1 (en) | 2010-08-05 |
US11342071B2 (en) | 2022-05-24 |
US20130217998A1 (en) | 2013-08-22 |
CA3049975A1 (en) | 2010-08-05 |
EP2391971A1 (en) | 2011-12-07 |
JP2015109972A (en) | 2015-06-18 |
JP6404286B2 (en) | 2018-10-10 |
JP2012516719A (en) | 2012-07-26 |
CA2977574A1 (en) | 2010-08-05 |
WO2010088696A1 (en) | 2010-08-05 |
EP2391971A4 (en) | 2015-10-21 |
US11004561B2 (en) | 2021-05-11 |
CA2751422C (en) | 2017-10-17 |
CA3170396C (en) | 2023-04-18 |
CA3049975C (en) | 2022-10-11 |
US9642572B2 (en) | 2017-05-09 |
US11776686B2 (en) | 2023-10-03 |
US20120029345A1 (en) | 2012-02-02 |
CA2751422A1 (en) | 2010-08-05 |
CA3192190A1 (en) | 2010-08-05 |
CA3170396A1 (en) | 2010-08-05 |
JP2016202974A (en) | 2016-12-08 |
EP3968220A1 (en) | 2022-03-16 |
US20210193313A1 (en) | 2021-06-24 |
JP5723788B2 (en) | 2015-05-27 |
US8444564B2 (en) | 2013-05-21 |
US11935648B1 (en) | 2024-03-19 |
CA2977574C (en) | 2019-07-23 |
US20130253379A1 (en) | 2013-09-26 |
US20220215947A1 (en) | 2022-07-07 |
US20170296115A1 (en) | 2017-10-19 |
EP2391971B1 (en) | 2021-11-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11935648B1 (en) | Noninvasive diagnostic system | |
US20130211259A1 (en) | Determination of joint condition based on vibration analysis | |
US20020147415A1 (en) | Method for simultaneous anatomical and functional mapping of a joint | |
JPWO2006085387A1 (en) | Non-invasive moving body analysis system and method of use thereof | |
WO2014150780A2 (en) | Determination of joint condition based on vibration analysis | |
Zhang et al. | Techniques for in vivo measurement of ligament and tendon strain: a review | |
US20130185310A1 (en) | Method and system for human joint treatment plan and personalized surgery planning using 3-d kinematics, fusion imaging and simulation | |
Sati et al. | Computer assisted knee surgery: diagnostics and planning of knee surgery | |
Safaei et al. | Vibration stimulation as a non-invasive approach to monitor the severity of meniscus tears | |
Blokker | Development and assessment of a micro-CT based system for quantifying loaded knee joint kinematics and tissue mechanics | |
FitzPatrick et al. | Synchronized acoustic emission and gait analysis of total hip replacement patients | |
WO2023099936A1 (en) | A wearable to assess knee joint integrity using non-contact acoustic sensors | |
Zingde | In vivo mechanics of cam-post engagement in fixed and mobile bearing tka and vibroarthrography of the knee joint | |
Klets | Subject-specific musculoskeletal modeling of the lower extremities in persons with unilateral cerebral palsy | |
Okamoto et al. | Study of anatomical landmark sampling error effect on motion measurement reproducibility for orthopaedic physical examination assisting system | |
Sheehan | Cross-platform Comparison of Imaging Technologies for Measuring Musculoskeletal Motion Richard M. Smith, BA1 and Frances T. Sheehan, PhD1 | |
na Opeenvolgende | Jildert Apperloo | |
Seisler | Normative 3D Patellofemoral and Tibiofemoral Kinematics: a Dynamic, in vivo Study |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |