GB2623770A - Ultrasound imaging - Google Patents
Ultrasound imaging Download PDFInfo
- Publication number
- GB2623770A GB2623770A GB2215789.5A GB202215789A GB2623770A GB 2623770 A GB2623770 A GB 2623770A GB 202215789 A GB202215789 A GB 202215789A GB 2623770 A GB2623770 A GB 2623770A
- Authority
- GB
- United Kingdom
- Prior art keywords
- ultrasound
- probe
- images
- captured
- ultrasound images
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000012285 ultrasound imaging Methods 0.000 title description 3
- 238000002604 ultrasonography Methods 0.000 claims abstract description 177
- 239000000523 sample Substances 0.000 claims abstract description 169
- 238000003384 imaging method Methods 0.000 claims abstract description 39
- 238000000034 method Methods 0.000 claims description 44
- 238000005259 measurement Methods 0.000 claims description 28
- 238000000926 separation method Methods 0.000 claims description 16
- 230000033001 locomotion Effects 0.000 description 9
- 238000013473 artificial intelligence Methods 0.000 description 5
- 238000004891 communication Methods 0.000 description 4
- 238000010801 machine learning Methods 0.000 description 4
- 210000000056 organ Anatomy 0.000 description 4
- 238000004590 computer program Methods 0.000 description 3
- 230000000644 propagated effect Effects 0.000 description 3
- 210000001015 abdomen Anatomy 0.000 description 2
- 238000002059 diagnostic imaging Methods 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 230000011218 segmentation Effects 0.000 description 2
- 239000007787 solid Substances 0.000 description 2
- 210000001835 viscera Anatomy 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 210000003484 anatomy Anatomy 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 210000004204 blood vessel Anatomy 0.000 description 1
- 210000002318 cardia Anatomy 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 238000012790 confirmation Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000002955 isolation Methods 0.000 description 1
- 210000004185 liver Anatomy 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000029058 respiratory gaseous exchange Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 239000000758 substrate Substances 0.000 description 1
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/42—Details of probe positioning or probe attachment to the patient
- A61B8/4245—Details of probe positioning or probe attachment to the patient involving determining the position of the probe, e.g. with respect to an external reference frame or to the patient
- A61B8/4254—Details of probe positioning or probe attachment to the patient involving determining the position of the probe, e.g. with respect to an external reference frame or to the patient using sensors mounted on the probe
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/13—Tomography
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/42—Details of probe positioning or probe attachment to the patient
- A61B8/4245—Details of probe positioning or probe attachment to the patient involving determining the position of the probe, e.g. with respect to an external reference frame or to the patient
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/44—Constructional features of the ultrasonic, sonic or infrasonic diagnostic device
- A61B8/4444—Constructional features of the ultrasonic, sonic or infrasonic diagnostic device related to the probe
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/48—Diagnostic techniques
- A61B8/483—Diagnostic techniques involving the acquisition of a 3D volume of data
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/52—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/5215—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data
- A61B8/5238—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data for combining image data of patient, e.g. merging several images from different acquisition modes into one image
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S15/00—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
- G01S15/88—Sonar systems specially adapted for specific applications
- G01S15/89—Sonar systems specially adapted for specific applications for mapping or imaging
- G01S15/8906—Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S15/00—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
- G01S15/88—Sonar systems specially adapted for specific applications
- G01S15/89—Sonar systems specially adapted for specific applications for mapping or imaging
- G01S15/8906—Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques
- G01S15/8934—Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques using a dynamic transducer configuration
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S15/00—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
- G01S15/88—Sonar systems specially adapted for specific applications
- G01S15/89—Sonar systems specially adapted for specific applications for mapping or imaging
- G01S15/8906—Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques
- G01S15/8934—Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques using a dynamic transducer configuration
- G01S15/8936—Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques using a dynamic transducer configuration using transducers mounted for mechanical movement in three dimensions
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S15/00—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
- G01S15/88—Sonar systems specially adapted for specific applications
- G01S15/89—Sonar systems specially adapted for specific applications for mapping or imaging
- G01S15/8906—Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques
- G01S15/8993—Three dimensional imaging systems
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S15/00—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
- G01S15/88—Sonar systems specially adapted for specific applications
- G01S15/89—Sonar systems specially adapted for specific applications for mapping or imaging
- G01S15/8906—Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques
- G01S15/8995—Combining images from different aspect angles, e.g. spatial compounding
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/52—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
- G01S7/52017—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 particularly adapted to short-range imaging
- G01S7/52085—Details related to the ultrasound signal acquisition, e.g. scan sequences
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/46—Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/54—Control of the diagnostic device
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Remote Sensing (AREA)
- Radar, Positioning & Navigation (AREA)
- Acoustics & Sound (AREA)
- Radiology & Medical Imaging (AREA)
- Veterinary Medicine (AREA)
- Heart & Thoracic Surgery (AREA)
- Medical Informatics (AREA)
- Molecular Biology (AREA)
- Surgery (AREA)
- Animal Behavior & Ethology (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Pathology (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Computer Networks & Wireless Communication (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Ultra Sonic Daignosis Equipment (AREA)
Abstract
A subject 150 is imaged by capturing a series of images with an ultrasound probe 100, determining a respective orientation of the probe 100 for each image with an accelerometer (110, fig 1a), and arranging the series of images into a three dimensional image based on the probe orientation. The ultrasound system may determine whether there are any gaps (330, fig 3) between the arranged images (320, fig 3), and if the gaps 330 exceed a threshold instructions are output for further imaging. The accelerometer 110 may be positioned at the acoustic lens end 120 of the probe 100. One of the images (310, fig 3) may be selected as a reference image and the remaining images 320 arranged relative to the reference image 310 based on the difference in orientation. The system may identify a feature within the images, compare the position of the feature in each image, determine a lateral offset between the images and translate one or more of the images to compensate for the offset.
Description
Ultrasound Imaging FIELD
[0001] Embodiments described herein relate to methods and systems for three-dimensionally imaging a subject.
BACKGROUND
[0002] Three dimensional ultrasound is a medical imaging technique which is often used in foetal, cardia or intra-vascular applications. Existing three-dimensional ultrasound imaging techniques involve controlling the direction of a sequence of ultrasound images using a beam steering means or a motorised gimbal that is integrated with an ultrasound probe. Such sophisticated methods may be expensive, cover only limited angular ranges, and require a specially trained medical professional to perform them, limiting their accessibility.
[0003] An aim of the present invention is to provide cost-effective means for reliably obtaining a high field-of-view three-dimensional ultrasound image without requiring a specialised medical professional.
SUMMARY OF THE INVENTION
[0004] According to an embodiment, there is provided a method of imaging a subject, the method comprising: capturing a series of ultrasound images using an ultrasound probe; determining a respective orientation of the ultrasound probe when each of the series of ultrasound images is captured, each orientation being derived from measurements of an accelerometer comprised by the ultrasound probe; arranging each of the series of captured ultrasound images in a three-dimensional arrangement based on the orientation of the ultrasound probe when that ultrasound image was captured.
[0005] Such a method may enable a three-dimensional representation of the subject to be captured by arranging the three-dimensional images in their associated orientations.
By varying the orientation of the ultrasound probe a larger amount as the series of images are captured, a three-dimensional representation with a larger field of view may be obtained.
[0006] In some embodiments, the method further comprises determining whether any gaps between the arranged ultrasound images exceed a threshold separation; and if one or more gaps between the arranged ultrasound images exceed the threshold separation are identified, outputting instructions to perform further imaging.
[0007] Identifying gaps and outputting instructions to perform further imaging if they are identified may ensure the obtained three-dimensional representation is not limited by ineffective manipulation of the ultrasound probe while the images are captured.
[0008] In some embodiments, the instructions to perform further imaging are output by or to a graphical user interface [0009] In some embodiments, the instructions to perform further imaging comprise instructions indicating one or more orientations in which to position the ultrasound probe while performing the further imaging, the one or more orientations in which to position the ultrasound prove corresponding to orientations of the one or more gaps.
[0010] In some embodiments, the method further comprises positioning the ultrasound probe on a region of interest of the subject before capturing the series of ultrasound images and reorienting the ultrasound probe while holding it on the region of interest while the series of ultrasound images are captured.
[0011] The ultrasound probe may be manually positioned and reoriented by a user manipulating the probe, while other steps may be performed by an ultrasound system, for example, by executing computer instructions.
[0012] In some embodiments, positioning the ultrasound probe on the region of interest comprises positioning the probe on the subject, capturing an initial ultrasound image using the ultrasound probe, identifying one or more features of the subject in the initial image, and using the one or more identified features to determine whether the probe is correctly positioned on the region of interest.
[0013] Such steps may assist a non-medical-professional user in positioning the ultrasound probe to image the region of interest. Whether the probe is correctly positioned on the region of interest may be performed using an artificial intelligence or machine learning model which may be configured and/or trained to perform semantic segmentation. The identity, position, orientation and/or absence of features from the one or more identified features may be used to determine whether the probe is correctly positioned. In some embodiments, a series of initial ultrasound images may be captured, have features identified therein and used to determine whether the probe is correctly positioned, for example, until the probe is determined to be correctly positioned.
[0014] In some such embodiments, the method further comprises, when the probe is determined not to be correctly positioned, outputting instructions to reposition the ultrasound probe [0015] In some embodiments, arranging the series of captured ultrasound images in the three-dimensional arrangement comprises selecting one of the images as a reference image and arranging the remaining images relative to the reference image in the arrangement, based on the differences in the orientations in which they were captured and the orientation in which the reference image was captured.
[0016] In some such embodiments, the method further comprises identifying a feature of the subject in a plurality of the captured ultrasound images, comparing the position of the identified feature in the plurality of the captured ultrasound images to determine a lateral offset between the plurality of captured ultrasound images, and translating one or more of the plurality of captured ultrasound images to compensate for the lateral offset.
[0017] According to embodiments of the invention, there is provided a system for imaging a subject, the system comprising: an ultrasound probe configured to capture a series of ultrasound images; an accelerometer, comprised by the ultrasound probe; an output device for providing instructions to a user; and a processor, the processor configured to: use measurements of the accelerometer to derive a respective orientation of the ultrasound probe when each of the series of ultrasound images is captured; arrange each of the series of captured ultrasound images in a three-dimensional arrangement based on the orientation of the ultrasound probe when that ultrasound image was captured [0018] In some embodiments, the processor is further configured to: determine whether any gaps between the arranged ultrasound images exceed a threshold separation; and when one or more gaps between the arranged ultrasound images exceed the threshold separation are identified, outputting instructions to perform further imaging using the output device.
[0019] In some embodiments, the output device is an electronic display configured to display the instructions using a graphical user interface.
[0020] In some embodiments, the instructions to perform further imaging comprise instructions indicating one or more orientations in which to position the ultrasound probe while performing the further imaging, the one or more orientations in which to position the ultrasound prove corresponding to orientations of the one or more gaps.
[0021] In some embodiments, the ultrasound probe is further configured to capture an initial ultrasound image before capturing the series of ultrasound images, and wherein the processor is further configured to identify one or more features of the subject in the initial image, and use the identify and/or position of the one or more features to determine whether the probe is correctly positioned on the region of interest.
[0022] In some embodiments, arranging the series of captured ultrasound images in the three-dimensional arrangement comprises selecting one of the images as a reference image and arranging the remaining images relative to the reference image in the arrangement, based on the differences in the orientations in which they were captured and the orientation in which the reference image was captured.
[0023] In some embodiments, the system is further configured to: identify a feature of the subject in a plurality of the captured ultrasound images; compare the position of the identified feature in the plurality of the captured ultrasound images to determine a lateral offset between the plurality of captured ultrasound images, and translate one or more of the plurality of captured ultrasound images to compensate for the lateral offset.
[0024] According to embodiments of the invention, there is provided storage medium comprising computer instructions executable by one or more processors, the computer instructions when executed by the one or more processors causing the one or more processors to perform a method according as described above. The method may comprise any of the optional features described herein.
BRIEF DESCRIPTION OF THE FIGURES [0025] Fig. la shows an ultrasound probe for use in imaging a subject; [0026] Fig. lb shows a user holding the ultrasound probe of Fig. la against a subject and rotating it through a range of orientations; [0027] Fig. lc is a diagram showing a range of motion through which the ultrasound probe of Fig. 1 may be moved in use; [0028] Fig. 2a is a flowchart of an embodiment of a method for imaging a subject; [0029] Fig. 2b shows additional optional steps of the method of Fig. 2a; and [0030] Fig. 3 shows a cross sectional view of a three-dimensional arrangement.
DETAILED DESCRIPTION
[0031] Referring to figures 1 to 3 generally, embodiments of methods and systems for three-dimensionally imaging a subject are described, in which an ultrasound probe comprising an accelerometer captures both ultrasound images and orientation data, from which a three-dimensional image is constructed.
[0032] In such embodiments, an ultrasound probe comprises at least one accelerometer, configured to detect changes in the orientation of the probe. The accelerometer may be arranged adjacent an end of the probe at which ultrasonic acoustic waves are emitted and detected, for example proximate to, or adjacent, an acoustic lens of the probe.
[0033] In example embodiments described herein, the accelerometer is configured to detect changes in orientations of the probe in three or more dimensions, for example, the accelerometer may be a three-degrees of freedom (300F) accelerometer. The accelerometer may be configured to record orientations of the probe, or changes therein, as sets of associated pitch, yaw and roll angles thereof.
[0034] In alternative embodiments, the accelerometer may be configured to detect changes in orientations of the probe in fewer dimensions, for example, in only a single dimension. For example, in some embodiments the accelerometer may be configured to detect changes about only a single axis of the probe, in which case a user may only tilt the probe about such an axis in use. The accelerometer is preferably configured to detect rotation of the probe about at least a short transverse axis of the probe parallel to an imaging plane thereof.
[0035] Existing methods of 3D imaging a subject frequently require more sophisticated accelerometers with more than three degrees of freedom, such as six-degrees of freedom accelerometers that detect changes in their position as well changes in their orientation. However embodiments described herein allow changes in position of the accelerometer and associated ultrasound probe to be determined from features in captured ultrasound images and fewer-degree-of-freedom orientation changes, without requiring such accelerometers, which may be more expensive, complex and/or bulky.
[0036] Fig. la shows an example of an ultrasound probe 100 for use in embodiments described herein. The illustrated probe is a curvilinear phased array ultrasound probe, however it will be appreciated that any other ultrasound probe may be used, such as a linear ultrasound probe. The probe is configured to be held and manipulated by a user, and/or by other suitable machinery, enabling the orientation of the probe to be varied in use, for example while held in contact with a subject. In use, the probe captures ultrasound images in orientations corresponding to the orientation in which it is held.
[0037] The illustrated probe 100 comprises an accelerometer 110 arranged adjacent a distal end, or "nose end", 120 of the probe 100, at which the acoustic lens of the probe 100 is located.
[0038] Such an ultrasound probe 100 defines or is comprised by a system configured to three-dimensionally image a subject. The system may further comprise a computing device, such as a terminal, which may be configured to control the ultrasound probe and/or to view, output or manipulate images captured by the probe, or composites thereof. The system may be configured to perform an imaging method as described herein, for example by executing computer instructions stored in one or more storage media comprised by or associated with the system using one or more processors comprised by the system, for example, a non-transitory storage media such as a memory comprised by the computing device.
[0039] In some embodiments, the ultrasound probe 100 and/or the computing device of the system is configured to output instructions and/or other information for a user. Such instructions and/or information may comprise visual, audio, haptic, and/or pictographic elements. The instructions may be output directly by the system, for example using an electronic visual display, graphical user interface (GUI), speaker or other output device comprised by, or connected to, the ultrasound probe and/or computing device, or may be output to another computing device in communication with the system, such as a personal computer or smartphone, in order to be output to a user by said other computing device.
[0040] Fig. lb shows an example of the probe 100 being used to image a subject 150.
A user 140 holds the probe, such that its acoustic lens end 120 is pressed against or into the exterior of the subject 150, such as a portion of their own body. In this position, an imaging plane 130 in which the probe 100 captures ultrasound images extends into the interior of the subject 150.
[0041] In use, the user holds the acoustic lens end 120 of the probe 100 in the same position on the subject 150 while tilting the probe 100 about an axis 160 parallel to the imaging plane 130 and tangential to the surface of the subject 150, such that tilting the probe 130 sweeps the imaging plane through the subject 150. In the illustrated example, this axis 160 is a lateral axis of the probe 100, perpendicular to the length of its handle. The reorientation of the probe 100 and the imaging plane 130 as the probe is tilted about the axis 160 are shown with double-headed arrows. Tilting an ultrasound probe 100 about lateral axis 160 of the probe 100 in this manner is sometimes referred to as "fanning" the probe 100. As the probe 100 is tilted, the accelerometer 120 measures the changes in orientation of the probe 100 (and by extension, of the imaging plane 130).
[0042] Fig. 1c shows an example of the probe 100 being used to image the abdomen of a human patient 150. As in the illustrated example, the probe 100 may be arranged such that its imaging plane 130 is parallel to a longitudinal axis or midline of the body 150, and the probe 100 may be rotated about an axis 160 substantially parallel to the longitudinal axis or midline, such that all of the series of images are substantially parallel to said longitudinal axis or midline, and such that the ultrasound probe is rotated within a transverse plane 170 substantially perpendicular to the probe's imaging plane throughout the motion. Capturing a series of ultrasound images in this arrangement may advantageously facilitate identifying, and compensating for, motion of organs of the patient 150 relative to the probe 100 within such planes, for example as a consequence of the patient's breathing or heartbeat. In other examples, the ultrasound probe may be rotated within other planes such as sagittal or coronal planes.
[0043] Fig. 2a is a flowchart 200 of a method of imaging a subject 150. In some embodiments, the subject 150 is a patient's body, such as a human patient's body, and the methods or systems are medical imaging methods or systems. However it will be appreciated that in alternative embodiments, the subject may be a non-human animal body or an inanimate device or object.
[0044] During measurement steps 210, 215, of the method 200, an ultrasound probe 100 comprising an accelerometer for detecting changes in its orientation is located on a region of interest on the subject 150. In some embodiments, the probe 100 is located in a target location on the region of interest, for example near a centre thereof. In some embodiments, the method 200 may comprise positioning the ultrasound probe 200 on the region of interest before the measurement steps 210, 215 are performed.
[0045] In some embodiments, instructions on where to position the probe 100 on and/or within the region of interest are output by the system. For example, in image of a probe 100 in position may be presented to the user. Providing instructions on positioning the ultrasound probe 100 may enable a non-medical-professional user to use the ultrasound probe to obtain a three-dimensional ultrasound medical image by following them. The region of interest may be input or selected by a user, for example using a user interface comprised by the probe, or a terminal or computing device in communication therewith.
[0046] In some embodiments, positioning the ultrasound probe 100 comprises detecting that the ultrasound probe 100 is positioned on the region of interest of the subject 150, or on a target location thereof. For example, a user 140 may move the probe across the surface of the subject 150 until the probe 100 is detected to be on the region of interest and/or target location. The probe 100 may be moved across the surface by sliding it over the surface, with its acoustic lens end 120 held against the surface, and/or with a longitudinal axis of the probe 100 extending substantially perpendicularly to the surface.
[0047] In some such embodiments, as the probe 100 is in the process of being positioned on the region of interest, one or more initial ultrasound images are captured using the ultrasound probe 100 and whether the probe is correctly positioned or not may be determined based on the contents of such initial ultrasound images. In some embodiments, such initial ultrasound images may be captured continuously and/or at a pre-set frequency.
[0048] In some embodiments, the system is configured to identify the presence, absence, and/or position of one or more features of the subject in each of the initial ultrasound images captured during positioning of the ultrasound image. For example, using an image recognition artificial intelligence or machine learning model to identify any instances of the one or more features in the initial images. The system may be further configured to use the identity, position, and/or absence of the one or more identified features within an ultrasound image to determine/ whether the probe 100 was correctly positioned on the region of interest and/or a target location thereon when that ultrasound image was captured.
[0049] When the system determines that the probe 100 is not correctly positioned when an ultrasound image is captured, or after a pre-set period has elapsed and/or a pre-set number of images have been captured (for example, following initialisation of the system, provision of instructions to the user, or from some other initial time) it may output an indication that that the probe is incorrectly positioned and/or instructions to, or how to, move the probe 100 to the correct position. In some embodiments, the actual location of the probe is determined from the one or more features, and the instructions may indicate a direction in which to move the probe 100 from the determined location to the correct position. When the system determines that the probe 100 is correctly positioned, or that the probe has been correctly positioned for a pre-set time period and/or number of images, a confirmation or other indication may be provided to the user and/or the next stage 210, 215 of the method may be performed.
[0050] In measurement steps 210, 215 of the method 200, after the ultrasound probe 100 has been positioned on the region of interest, a measurement phase is performed, during which a series of ultrasound images are captured 210 and orientation data is collected 215.
[0051] The measurement phase may be initiated when the system determines that the probe 100 has been correctly positioned as described above. In alternative embodiments, the measurement phase is initiated by a user, for example using one or more controls on the ultrasound probe or on a terminal or computing device in communication therewith.
[0052] Each captured 210 ultrasound image may be a two-dimensional planar image or "slice" in a plane corresponding to the orientation of the probe 100 while it was captured. As the series of ultrasound images are captured 210, a user 140 manipulates the probe 100 to reorient it relative to the subject 150. When reorienting the probe 100, the user 140 preferably holds the acoustic lens end 120 of the acoustic probe 100 on the same location on the subject 150.
[0053] In some embodiments, the probe may rotated in substantially a single direction, for example by fanning the probe as described above with reference to Fig. 1 b. In other embodiments, the probe may be rotated in multiple different directions, thereby capturing images in a greater variety of different orientations.
[0054] In some embodiments in which the subject 150 is a patient's body, the probe 100 may be rotated about an axis parallel to the patient's midline. For example, as described above with reference to Fig. 1c.
[0055] When each of the series of ultrasound images is captured 210, a respective orientation in which the probe is arranged while said ultrasound image of the series is being captured is determined from measurements of the accelerometer and may be recorded. The determined orientations may be relative to an absolute coordinate system, relative to each other, and/or relative to some reference orientation, such as an initial orientation when the first image of the series was captured, or when the probe was activated. The respective orientation determined for each ultrasound image defines the plane in which each ultrasound image is located. The captured images and their respective orientations may be stored in a memory.
[0056] In some embodiments, ultrasound images of the series and their associated orientations are captured continuously, for example at a set frequency. In alternative embodiments, an ultrasound image may instead be captured whenever the orientation of the probe measured by the accelerometer 110 differs from any orientation in which a previous ultrasound image of the series was captured, for example, by more than a threshold amount. In some embodiments, each ultrasound image and/or its associated orientation may be timestamped.
[0057] During the measurement phase, a user may tilt or otherwise reorient the ultrasound probe, while holding it in contact with the region of interest. In some embodiments, instructions to reorient the probe 100, or a specific motion in which to reorient the probe 100 may be output by the system, for example, after the measurement phase is automatically initiated upon detection that the probe 100 has been correctly positioned as described above.
[0058] The measurement phase may continue for a fixed period of time, until a fixed number or range of orientations of ultrasound images have been captured, and/or until manually ended by a user, for example using one or more controls on the ultrasound probe 100 or on a terminal or computing device in communication therewith.
[0059] After the measurement phase 210, 215, in a second stage of the method 200, the series of captured ultrasound images are assembled 220 into a three-dimensional arrangement of images based on the recorded orientations in which they were captured.
[0060] Assembling the series of captured images into the arrangement may comprise selecting one of the ultrasound images as a reference image for the arrangement. The selected reference image may be the image of a portion of the subject closest to one end or extreme of the subject, which may be determined based on the associated orientation of the probe 100 when it was captured. For example, in some embodiments in which the images are of a portion of a patient's body, the selected image is the most cranial (located closest towards the top of the patient's head) and medial (located closest to the midline of the patient's body, which extends from the top of the patient's head towards their feet).
[0061] After an image is selected as a reference image, the remaining images of the series may arranged 220 relative to the selected image in the arrangement, based on the difference between their orientations (corresponding to the orientations of the probe when they were captured) and the orientation of the selected reference image.
[0062] Fig. 3 shows a cross section of an example a three-dimensional image arrangement 300 of ultrasound images constructed after a measurement phase as described above is performed, in which the probe is rotated in a single direction about a single axis. The arrangement comprises a reference image 310 which is arranged horizontally within the arrangement 300. The selected reference image is the most cranial (located closest towards the top of the patient's head) and medial (located closest to the midline of the patient's body) image of the series. The remaining images 320 are arranged below the reference image 310 relative thereto based on the differences in their orientations and the orientation of the reference image.
[0063] The example illustrated arrangement 300 comprises ultrasound images which are all separated by angles within a single plane, as captured by an ultrasound probe 100 only being rotated in a single direction during the measurement steps 210, 215. In other examples, the ultrasound probe may be rotated in multiple directions around multiple different axes, producing a more complex arrangement of ultrasound images.
[0064] Arrangements 300 of ultrasound images constructed using the method described herein can cover very large fields of view. In situations, such arrangements can have fields of view of greater than 180°, for example, where an ultrasound is pressed into a relatively soft or yielding portion of the subject.
[0065] As shown in Fig. 3, a constructed arrangement 300 of ultrasound images may comprise differently sized gaps between adjacent images. Fig. 2b shows additional steps 230, 240 that may be performed in some embodiments of a method 200 as described above, in order to prevent the image loosing quality or information as a result of excessively large such gaps [0066] After the arrangement is constructed 220, in a third stage of the method 200 any gaps between images 310, 320 within the arrangement 300 larger than a threshold separation may be identified 230. Such gaps may define areas of the region of interest of the subject that have been insufficiently imaged during the initial measurement phase and of which additional images are to be captured. The threshold separation may depend upon the region of interest and/or the subject that is being imaged.
[0067] The gaps between images 310, 320 within the arrangement may be one-dimensional angular separations between adjacent images, which may be compared to a threshold one-dimensional angular separation, or may two-dimensional solid angles, which may be bounded by images and/or edges of the region of interest, which may be compared to a threshold two-dimensional solid angle.
[0068] In the example three-dimensional image arrangement 300 shown in Fig. 3, the gaps are one-dimensional separations that can be compared to a one-dimensional threshold angle. The illustrated example arrangement 300 includes one gap 330 with a separation between adjacent images exceeds the threshold. Such a gap results in a lack of information and/or detail on this portion of the imaged region of interest of the subject.
[0069] When one or more gaps 330 between images 310, 320 within the arrangement 300 larger than a threshold separation are identified 230, instructions to perform further imaging may be output 240 by the system, in a fourth stage of the method 200. Such further imaging may then be performed in an additional measurement phase.
[0070] The instructions output 240 when a gap 330 exceeding the threshold may instruct a user to initiate and/or perform an additional measurement phase. For example, the instructions may comprise a user interface element enabling a user to initiate an additional measurement phase. Alternatively, or additionally, identifying one or more gaps 330 exceeding the threshold may automatically initiate an additional measurement phase if and/or when the ultrasound probe 100 is arranged to perform another measurement phase.
[0071] In some embodiments, the instructions output when one or more gaps 330 exceeding the threshold are identified 230 may simply instruct the user to perform additional measurements while moving the ultrasound probe or to repeat their previous movement with the ultrasound probe 100.
[0072] In other embodiments, the instructions may instruct a user to reorient the probe 100 to and/or through one or more specific orientations corresponding to the threshold-exceeding gaps 330, in one or more specific directions, and/or at specific speeds. For example, the instructions may indicate one or more orientations in which to hold the ultrasound probe while performing the further imaging, the one or more orientation in which to hold the ultrasound prove corresponding to orientations of the one or more gaps. In some embodiments, during the additional measurement phase, the system may output an indication when images have been captured covering each and/or all of the identified gaps 330.
[0073] When a gap 330 is identified in an arrangement 300 comprising ultrasound images with angular separations in a single dimension, as in the example shown in Fig. 3. The instructions may be to rotate the probe about the same axis as it was rotated about in the initial measurement steps 210, 215. In some examples, the instructions may specify to rotate the probe in the opposite direction about the axis, and/or to rotate the probe more slowly than in the initial measurement steps 210, 215.
[0074] In some embodiments, the system may be configured to identify one or more features of the subject in a plurality or all of the ultrasound images captured during the measurement phase, for example, using an artificial intelligence or machine learning model.
[0075] The identified features may be anatomical structures of an imaged patient's body, such as individual organs and/or blood vessels, and/or parts thereof. In some embodiments, identifying features may comprise identifying parts of said images showing any of a preselected group of structures, which may correspond to the region of the subject being imaged. Alternatively, identifying structures in the first and second images may comprise identifying geometric structures within the images, which may not be part of a preselected group.
[0076] An artificial intelligence or machine learning model with which the features are identified may be an artificial intelligence configured and/or trained to perform semantic segmentation.
[0077] After the one or more features are identified in the images, the positions of features that have been identified in more than one of the images may be compared between those images to determine a lateral offset between those images. Lateral motion herein refers to movement within the plane of the ultrasound image, such as in a direction towards or away from the ultrasound probe 100, or in a direction across the plane of the ultrasound image.
[0078] Such lateral offsets may be motion artefacts resulting from the the subject moving during the capturing of the images. For example, when a patient whose abdomen is being imaged breathes in and out, this may displace a probe held against their skin inwards and outwards relative to some of their internal organs, and/or may displace one or more imaged internal organs within their body. For example a patient's liver will move downwards within their body as they breathe in and will move upwards as they breathe out. This can result in organs or parts thereof being located in different positions within the patient's body in different ultrasound images of the series.
[0079] As described above, in some embodiments, the probe is arranged and pivoted such that its imaging plane remains parallel to the midline or longitudinal axis of the patient's body, for example, as shown in Fig. 1c. In such embodiments, movement of organs upwards, downwards, towards the probe, and/or away from the probe between images of the series will be visible.
[0080] After such a lateral offset between images has been identified, images that have been laterally offset with respect to each other may be translated within the three-dimensional arrangement to remove such lateral offsets.
[0081] In some embodiments, the three-dimensional arrangement may be, or may be reformatted into, a three-dimensional image format comprising an array of voxels. Such a format may facilitate isolation and/or analysis of individual parts of the subject, which may be represented by groups or clusters of voxels within the array. Alternatively, the three-dimensional arrangement may be reformatted into a three dimensional array or a point cloud.
[0082] Implementations of the subject matter and the operations described in this specification can be realized in digital electronic circuitry, or in computer software, firmware, or hardware, including the structures disclosed in this specification and their structural equivalents, or in combinations of one or more of them. Implementations of the subject matter described in this specification can be realized using one or more computer programs, i.e., one or more modules of computer program instructions, encoded on computer storage medium for execution by, or to control the operation of, data processing apparatus. Alternatively or in addition, the program instructions can be encoded on an artificially generated propagated signal, e.g., a machine-generated electrical, optical, or electromagnetic signal that is generated to encode information for transmission to suitable receiver apparatus for execution by a data processing apparatus. A computer storage medium can be, or be included in, a computer-readable storage device, a computer-readable storage substrate, a random or serial access memory array or device, or a combination of one or more of them. Moreover, while a computer storage medium is not a propagated signal, a computer storage medium can be a source or destination of computer program instructions encoded in an artificially generated propagated signal. The computer storage medium can also be, or be included in, one or more separate physical components or media (e.g., multiple CDs, disks, or other storage devices).
[0083] While certain embodiments have been described, these embodiments have been presented by way of example only, and are not intended to limit the scope of the inventions. Indeed the novel methods and systems described herein may be embodied in a variety of other forms; furthermore, various omissions, substitutions and changes in the form of methods and systems described herein may be made without departing from the spirit of the inventions. The accompanying claims and their equivalents are intended to cover such forms of modifications as would fall within the scope and spirit of the inventions.
Claims (17)
- CLAIMS 2. 3. 4. 5. 6.A method of imaging a subject, the method comprising: capturing a series of ultrasound images using an ultrasound probe; determining a respective orientation of the ultrasound probe when each of the series of ultrasound images is captured, each orientation being derived from measurements of an accelerometer comprised by the ultrasound probe; arranging each of the series of captured ultrasound images in a three-dimensional arrangement based on the orientation of the ultrasound probe when that ultrasound image was captured.
- A method according to claim 1 further comprising: determining whether any gaps between the arranged ultrasound images exceed a threshold separation; and when one or more gaps between the arranged ultrasound images exceed the threshold separation are identified, outputting instructions to perform further imaging.
- A method according to claim 2, wherein the instructions to perform further imaging are output by or to a graphical user interface.
- A method according to claim 2 or claim 3, wherein the instructions to perform further imaging comprise instructions indicating one or more orientations in which to position the ultrasound probe while performing the further imaging, the one or more orientations in which to position the ultrasound prove corresponding to orientations of the one or more gaps.
- A method according to any preceding claim, further comprising: positioning the ultrasound probe on a region of interest of the subject before capturing the series of ultrasound images; reorienting the ultrasound probe while holding it on the region of interest while the series of ultrasound images are captured.
- A method according to claim 5, wherein positioning the ultrasound probe on the region of interest comprises positioning the probe on the subject, capturing an initial ultrasound image using the ultrasound probe, identifying one or more features of the subject in the initial image, and using the one or more identified features to determine whether the probe is correctly positioned on the region of interest.
- A method according to claim 6, the method comprising, when the probe is determined not to be correctly positioned, outputting instructions to reposition the ultrasound probe.
- 8. A method according to any preceding claim, wherein arranging the series of captured ultrasound images in the three-dimensional arrangement comprises selecting one of the images as a reference image and arranging the remaining images relative to the reference image in the arrangement, based on the differences in the orientations in which they were captured and the orientation in which the reference image was captured.
- 9. A method according to any preceding claim, further comprising: identifying a feature of the subject in a plurality of the captured ultrasound images, comparing the position of the identified feature in the plurality of the captured ultrasound images to determine a lateral offset between the plurality of captured ultrasound images, and translating one or more of the plurality of captured ultrasound images to compensate for the lateral offset.
- 10. A system for imaging a subject, the system comprising: an ultrasound probe configured to capture a series of ultrasound images; an accelerometer, comprised by the ultrasound probe; an output device for providing instructions to a user; and a processor, the processor configured to: use measurements of the accelerometer to derive a respective orientation of the ultrasound probe when each of the series of ultrasound images is captured; arrange each of the series of captured ultrasound images in a three-dimensional arrangement based on the orientation of the ultrasound probe when that ultrasound image was captured.
- 11. A system according to claim 10, the processor further configured to: 12. 13. 14. 15. 16.determine whether any gaps between the arranged ultrasound images exceed a threshold separation; and when one or more gaps between the arranged ultrasound images exceed the threshold separation are identified, outputting instructions to perform further imaging using the output device.
- A system according to claim 11, wherein the output device is an electronic display configured to display the instructions using a graphical user interface.
- A system according to any of claims 9 to 12, wherein the instructions to perform further imaging comprise instructions indicating one or more orientations in which to position the ultrasound probe while performing the further imaging, the one or more orientations in which to position the ultrasound prove corresponding to orientations of the one or more gaps.
- A system according to any of claims 10 to 13, wherein the ultrasound probe is further configured to capture an initial ultrasound image before capturing the series of ultrasound images, and wherein the processor is further configured to identify one or more features of the subject in the initial image, and use the one or more identified features to determine whether the probe is correctly positioned on the region of interest.
- A system according to any of claims 10 to 14, wherein arranging the series of captured ultrasound images in the three-dimensional arrangement comprises selecting one of the images as a reference image and arranging the remaining images relative to the reference image in the arrangement, based on the differences in the orientations in which they were captured and the orientation in which the reference image was captured.
- A system according to any of claims 10 to 15, wherein the system is further configured to: identify a feature of the subject in a plurality of the captured ultrasound images; compare the position of the identified feature in the plurality of the captured ultrasound images to determine a lateral offset between the plurality of captured ultrasound images, and translate one or more of the plurality of captured ultrasound images to compensate for the lateral offset.
- 17. A storage medium comprising computer instructions executable by one or more processors, the computer instructions when executed by the one or more processors causing the one or more processors to perform a method according to any of claims 1 to 9.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB2215789.5A GB2623770A (en) | 2022-10-25 | 2022-10-25 | Ultrasound imaging |
PCT/GB2023/052794 WO2024089423A1 (en) | 2022-10-25 | 2023-10-25 | System and method for three-dimensional imaging |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB2215789.5A GB2623770A (en) | 2022-10-25 | 2022-10-25 | Ultrasound imaging |
Publications (2)
Publication Number | Publication Date |
---|---|
GB202215789D0 GB202215789D0 (en) | 2022-12-07 |
GB2623770A true GB2623770A (en) | 2024-05-01 |
Family
ID=84818606
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
GB2215789.5A Pending GB2623770A (en) | 2022-10-25 | 2022-10-25 | Ultrasound imaging |
Country Status (1)
Country | Link |
---|---|
GB (1) | GB2623770A (en) |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0487339A1 (en) * | 1990-11-22 | 1992-05-27 | Advanced Technology Laboratories, Inc. | Acquisition and display of ultrasonic images from sequentially orientated image planes |
WO2006127142A2 (en) * | 2005-03-30 | 2006-11-30 | Worcester Polytechnic Institute | Free-hand three-dimensional ultrasound diagnostic imaging with position and angle determination sensors |
US20080146941A1 (en) * | 2006-12-13 | 2008-06-19 | Ep Medsystems, Inc. | Catheter Position Tracking for Intracardiac Catheters |
WO2016201006A1 (en) * | 2015-06-08 | 2016-12-15 | The Board Of Trustees Of The Leland Stanford Junior University | 3d ultrasound imaging, associated methods, devices, and systems |
WO2017200515A1 (en) * | 2016-05-16 | 2017-11-23 | Analogic Corporation | 3-d us volume from 2-d images from freehand rotation and/or translation of ultrasound probe |
CN114869332A (en) * | 2022-05-12 | 2022-08-09 | 亳州联岐医疗科技有限公司 | Three-dimensional ultrasonic image recombination method based on nine-axis sensor |
-
2022
- 2022-10-25 GB GB2215789.5A patent/GB2623770A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0487339A1 (en) * | 1990-11-22 | 1992-05-27 | Advanced Technology Laboratories, Inc. | Acquisition and display of ultrasonic images from sequentially orientated image planes |
WO2006127142A2 (en) * | 2005-03-30 | 2006-11-30 | Worcester Polytechnic Institute | Free-hand three-dimensional ultrasound diagnostic imaging with position and angle determination sensors |
US20080146941A1 (en) * | 2006-12-13 | 2008-06-19 | Ep Medsystems, Inc. | Catheter Position Tracking for Intracardiac Catheters |
WO2016201006A1 (en) * | 2015-06-08 | 2016-12-15 | The Board Of Trustees Of The Leland Stanford Junior University | 3d ultrasound imaging, associated methods, devices, and systems |
WO2017200515A1 (en) * | 2016-05-16 | 2017-11-23 | Analogic Corporation | 3-d us volume from 2-d images from freehand rotation and/or translation of ultrasound probe |
CN114869332A (en) * | 2022-05-12 | 2022-08-09 | 亳州联岐医疗科技有限公司 | Three-dimensional ultrasonic image recombination method based on nine-axis sensor |
Also Published As
Publication number | Publication date |
---|---|
GB202215789D0 (en) | 2022-12-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11992369B2 (en) | Intelligent ultrasound system for detecting image artefacts | |
US10219782B2 (en) | Position correlated ultrasonic imaging | |
US10362941B2 (en) | Method and apparatus for performing registration of medical images | |
US8756033B2 (en) | Ultrasonic diagnostic imaging system and control method thereof | |
US10542955B2 (en) | Method and apparatus for medical image registration | |
JP6334141B2 (en) | Method and apparatus for navigating a CT scan by a marker | |
US20160317122A1 (en) | In-device fusion of optical and inertial positional tracking of ultrasound probes | |
TWI605795B (en) | Method and system of determining probe position in surgical site | |
CN111035408B (en) | Method and system for enhanced visualization of ultrasound probe positioning feedback | |
WO2012091144A4 (en) | Information processing apparatus and method | |
KR20150091748A (en) | Scan position guide method of three dimentional ultrasound system | |
US20200305837A1 (en) | System and method for guided ultrasound imaging | |
US12211221B2 (en) | Systems and methods for enhancement of 3D imagery and navigation via integration of patient motion data | |
KR20150129506A (en) | Method and Appartus for registering medical images | |
CN108024789B (en) | Inter-volume lesion detection and image preparation | |
GB2623770A (en) | Ultrasound imaging | |
JP6501796B2 (en) | Acquisition Orientation Dependent Features for Model-Based Segmentation of Ultrasound Images | |
WO2024089423A1 (en) | System and method for three-dimensional imaging | |
CN114782537A (en) | Human carotid artery localization method and device based on 3D vision | |
JP7275261B2 (en) | 3D ULTRASOUND IMAGE GENERATING APPARATUS, METHOD, AND PROGRAM | |
US12249084B2 (en) | Systems and methods for medical imagery enhancement by use of image guidance system | |
WO2024225133A1 (en) | Image processing device, measurement system, image processing method, and program | |
US20230200775A1 (en) | Ultrasonic imaging system | |
TW202322766A (en) | Ultrasonic imaging system including an ultrasonic probe, a first characteristic pattern, a second characteristic pattern, a storage unit, an image capture unit, a display unit, and a processing unit | |
CN118447059A (en) | Medical image registration method and device |