CN113208645A - Method and system for providing enhanced ultrasound images - Google Patents
Method and system for providing enhanced ultrasound images Download PDFInfo
- Publication number
- CN113208645A CN113208645A CN202110174601.4A CN202110174601A CN113208645A CN 113208645 A CN113208645 A CN 113208645A CN 202110174601 A CN202110174601 A CN 202110174601A CN 113208645 A CN113208645 A CN 113208645A
- Authority
- CN
- China
- Prior art keywords
- acoustic power
- ultrasound image
- ultrasound
- image
- training
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000002604 ultrasonography Methods 0.000 title claims abstract description 275
- 238000000034 method Methods 0.000 title claims abstract description 52
- 238000012545 processing Methods 0.000 claims abstract description 36
- 238000012549 training Methods 0.000 claims description 60
- 238000013473 artificial intelligence Methods 0.000 claims description 28
- 239000002872 contrast media Substances 0.000 claims description 19
- 230000008569 process Effects 0.000 claims description 18
- 239000003814 drug Substances 0.000 claims description 8
- 229940079593 drug Drugs 0.000 claims description 8
- 238000004590 computer program Methods 0.000 claims description 5
- 238000009877 rendering Methods 0.000 claims description 4
- 230000002708 enhancing effect Effects 0.000 abstract description 8
- 210000002569 neuron Anatomy 0.000 description 20
- 210000003484 anatomy Anatomy 0.000 description 13
- 238000013528 artificial neural network Methods 0.000 description 12
- 239000000523 sample Substances 0.000 description 11
- 230000006870 function Effects 0.000 description 10
- 238000002607 contrast-enhanced ultrasound Methods 0.000 description 8
- 230000015654 memory Effects 0.000 description 8
- 238000010586 diagram Methods 0.000 description 7
- 210000001519 tissue Anatomy 0.000 description 6
- 238000013527 convolutional neural network Methods 0.000 description 4
- 238000003384 imaging method Methods 0.000 description 4
- 210000000056 organ Anatomy 0.000 description 4
- 238000012285 ultrasound imaging Methods 0.000 description 4
- 238000004891 communication Methods 0.000 description 3
- 238000002592 echocardiography Methods 0.000 description 3
- 238000010801 machine learning Methods 0.000 description 3
- 210000004204 blood vessel Anatomy 0.000 description 2
- 210000000988 bone and bone Anatomy 0.000 description 2
- 238000013500 data storage Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 239000011159 matrix material Substances 0.000 description 2
- 230000001537 neural effect Effects 0.000 description 2
- 230000002123 temporal effect Effects 0.000 description 2
- 230000002792 vascular Effects 0.000 description 2
- 241001465754 Metazoa Species 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 210000000601 blood cell Anatomy 0.000 description 1
- 230000017531 blood circulation Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 230000006378 damage Effects 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 238000002059 diagnostic imaging Methods 0.000 description 1
- 238000002091 elastography Methods 0.000 description 1
- 230000007717 exclusion Effects 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000002156 mixing Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 210000005036 nerve Anatomy 0.000 description 1
- 230000035515 penetration Effects 0.000 description 1
- 230000010412 perfusion Effects 0.000 description 1
- 230000001105 regulatory effect Effects 0.000 description 1
- 230000010076 replication Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000002099 shear wave elastography Methods 0.000 description 1
- 210000004872 soft tissue Anatomy 0.000 description 1
- 238000000527 sonication Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/52—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/5215—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/48—Diagnostic techniques
- A61B8/481—Diagnostic techniques involving the use of contrast agent, e.g. microbubbles introduced into the bloodstream
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/46—Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
- A61B8/461—Displaying means of special interest
- A61B8/463—Displaying means of special interest characterised by displaying multiple images or images and diagnostic data on one display
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/52—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/5207—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of raw data to produce diagnostic data, e.g. for generating an image
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/52—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/5215—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data
- A61B8/5238—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data for combining image data of patient, e.g. merging several images from different acquisition modes into one image
- A61B8/5246—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data for combining image data of patient, e.g. merging several images from different acquisition modes into one image combining images from the same or different imaging techniques, e.g. color Doppler and B-mode
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/58—Testing, adjusting or calibrating the diagnostic device
- A61B8/587—Calibration phantoms
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/776—Validation; Performance evaluation
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/52—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
- G01S7/52017—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 particularly adapted to short-range imaging
- G01S7/5205—Means for monitoring or calibrating
- G01S7/52052—Means for monitoring or calibrating with simulation of echoes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
Abstract
The invention provides a method and system for providing an enhanced ultrasound image. The present invention provides a method and system for enhancing ultrasound images acquired at low acoustic power to simulate acquisition at high acoustic power. The method may include acquiring, by an ultrasound system, a first ultrasound image at a first acoustic power. The method may include processing, by at least one processor, the first ultrasound image to generate a second ultrasound image simulating a second acoustic power greater than the first acoustic power. The method may include presenting, at the display system, a second ultrasound image simulating second acoustic power.
Description
Technical Field
Certain embodiments relate to ultrasound imaging. More particularly, certain embodiments relate to methods and systems for providing enhanced ultrasound images simulating acquisition at high acoustic power by processing ultrasound images acquired at low acoustic power.
Background
Ultrasound imaging is a medical imaging technique for imaging organs and soft tissue in the human body. Ultrasound imaging uses real-time, non-invasive high frequency sound waves to produce a series of two-dimensional (2D) images and/or three-dimensional (3D) images.
Currently, the U.S. Food and Drug Administration (FDA) requires diagnostic ultrasound imaging systems to operate under certain acoustic output limits for safety reasons. However, when the transmission acoustic power of the same scan target is reduced, the quality of the ultrasound image deteriorates. Contrast Enhanced Ultrasound (CEUS) involves the use of microbubble contrast agents and specialized imaging techniques to better visualize blood flow and tissue perfusion information. Thus, CEUS transmits much lower acoustic power (e.g., about 10%) than conventional limits to avoid destruction of microbubbles of contrast agents. In CEUS, B-mode images with low transmit power are typically displayed alongside contrast-enhanced images to provide anatomical reference. Due to the low transmit power, the reference image typically suffers from less penetration, less contrast, lower spatial resolution, and higher background noise than B-mode images acquired at typical acoustic power.
Furthermore, the american medical ultrasound society (AIUM) and British Medical Ultrasound Society (BMUS) guidelines advocate reducing the acoustic power of obstetrical ultrasound scans, as well as reducing the regulatory limits of ophthalmic ultrasound scans. These ultrasound images are also of lower quality than images acquired at higher acoustic power.
Further limitations and disadvantages of conventional and traditional approaches will become apparent to one of skill in the art, through comparison of such systems with some aspects of the present disclosure as set forth in the remainder of the present application with reference to the drawings.
Disclosure of Invention
A system and/or method for enhancing ultrasound images acquired at low acoustic power to simulate acquisition at high acoustic power, substantially as shown in and/or described in connection with at least one of the figures, as set forth more completely in the claims.
These and other advantages, aspects, and novel features of the present disclosure, as well as details of an illustrated embodiment thereof, will be more fully understood from the following description and drawings.
Drawings
Fig. 1 is a block diagram of an exemplary ultrasound system and training system operable to enhance ultrasound images acquired at low acoustic power to simulate acquisition at high acoustic power, in accordance with various embodiments.
Fig. 2 is a flow diagram illustrating exemplary steps that may be used to enhance ultrasound images acquired at low acoustic power to simulate acquisition at high acoustic power, in accordance with various embodiments.
Fig. 3 is a flow diagram illustrating exemplary steps that may be used to enhance ultrasound images acquired at low acoustic power to simulate acquisition at high acoustic power, in accordance with various embodiments.
Detailed Description
Certain embodiments may be found in methods and systems for enhancing ultrasound images acquired at low acoustic power to simulate acquisition at high acoustic power. Various embodiments have the technical effect of providing higher quality ultrasound images by enhancing images acquired with low or limited acoustic output without destroying contrast agents or exceeding safety limits. Aspects of the present disclosure have the technical effect of enhancing image quality for ultrasound applications and/or modes requiring lower or limited transmitted acoustic power.
The foregoing summary, as well as the following detailed description of certain embodiments, will be better understood when read in conjunction with the appended drawings. To the extent that the figures illustrate diagrams of the functional blocks of various embodiments, the functional blocks are not necessarily indicative of the division between hardware circuitry. Thus, for example, one or more of the functional blocks (e.g., processors or memories) may be implemented in a single piece of hardware (e.g., a general purpose signal processor or a block of random access memory, hard disk, or the like) or multiple pieces of hardware. Similarly, the programs may be stand alone programs, may be incorporated as subroutines in an operating system, may be functions in an installed software package, and the like. It should be understood that the various embodiments are not limited to the arrangements and instrumentality shown in the drawings. It is to be further understood that the embodiments may be combined, or that other embodiments may be utilized and that structural, logical, and electrical changes may be made without departing from the scope of the various embodiments. The following detailed description is, therefore, not to be taken in a limiting sense, and the scope of the present disclosure is defined by the appended claims and their equivalents.
As used herein, an element or step recited in the singular and proceeded with the word "a" or "an" should be understood as not excluding plural said elements or steps, unless such exclusion is explicitly recited. Furthermore, references to "exemplary embodiments," "various embodiments," "certain embodiments," "representative embodiments," etc., are not intended to be interpreted as excluding the existence of additional embodiments that also incorporate the recited features. Moreover, unless explicitly stated to the contrary, embodiments "comprising," "including," or "having" an element or a plurality of elements having a particular property may include additional elements not having that property.
In addition, as used herein, the term "image" broadly refers to both a viewable image and data representing a viewable image. However, many embodiments generate (or are configured to generate) at least one viewable image. Further, as used herein, the phrase "image" is used to refer to ultrasound modes, such as B-mode (2D mode), M-mode, three-dimensional (3D) mode, CF mode, PW doppler, CW doppler, Contrast Enhanced Ultrasound (CEUS), and/or sub-modes of B-mode and/or CF, such as harmonic imaging, Shear Wave Elastography (SWEI), strain elastography, TVI, PDI, B-flow, MVI, UGAP, and in some cases MM, CM, TVD, where "image" and/or "plane" includes a single beam or multiple beams.
Further, as used herein, the term processor or processing unit refers to any type of processing unit that can perform the required computations required by the various embodiments, such as single core or multi-core: a CPU, an Accelerated Processing Unit (APU), a Graphics Processing Unit (GPU), a DSP, an FPGA, an ASIC, or a combination thereof.
It should be noted that various embodiments of generating or forming images described herein may include processes for forming images that include beamforming in some embodiments, and do not include beamforming in other embodiments. For example, the image may be formed without beamforming, such as by multiplying a matrix of demodulated data by a matrix of coefficients, such that the product is an image, and wherein the process does not form any "beams. In addition, the formation of an image may be performed using a combination of channels (e.g., synthetic aperture techniques) that may result from more than one transmit event.
In various embodiments, for example, sonication is performed in software, firmware, hardware, or a combination thereof to form an image, including ultrasound beamforming, such as receive beamforming. One specific implementation of an ultrasound system having a software beamformer architecture formed in accordance with various embodiments is shown in figure 1.
Fig. 1 is a block diagram of an exemplary ultrasound system 100 and training system 200 operable to enhance ultrasound images acquired at low acoustic power to simulate acquisition at high acoustic power, in accordance with various embodiments. Referring to fig. 1, an ultrasound system 100 and a training system 200 are shown. Ultrasound system 100 includes a transmitter 102, an ultrasound probe 104, a transmit beamformer 110, a receiver 118, a receive beamformer 120, an A/D converter 122, an RF processor 124, an RF/IQ buffer 126, a user input device 130, a signal processor 132, an image buffer 136, a display system 134, and an archive 138.
The transmitter 102 may comprise suitable logic, circuitry, interfaces and/or code that may be operable to drive the ultrasound probe 104. The ultrasound probe 104 may include a two-dimensional (2D) array of piezoelectric elements. The ultrasound probe 104 may include a set of transmit transducer elements 106 and a set of receive transducer elements 108 that generally constitute the same elements. In certain embodiments, the ultrasound probe 104 is operable to acquire ultrasound image data covering at least a substantial portion of an anatomical structure, such as a heart, a blood vessel, or any suitable anatomical structure.
The transmit beamformer 110 may comprise suitable logic, circuitry, interfaces and/or code that may be operable to control the transmitter 102 to drive the set of transmit transducer elements 106 through the transmit sub-aperture beamformer 114 to transmit ultrasonic transmit signals into a region of interest (e.g., a human, an animal, a subsurface cavity, a physical structure, etc.). The transmitted ultrasound signals may be backscattered from structures in the object of interest, such as blood cells or tissue, to generate echoes. The echoes are received by the receiving transducer elements 108.
A set of receive transducer elements 108 in the ultrasound probe 104 is operable to convert the received echoes to analog signals, sub-aperture beamformed by a receive sub-aperture beamformer 116, and then transmitted to a receiver 118. The receiver 118 may comprise suitable logic, circuitry, interfaces and/or code that may be operable to receive signals from the receive sub-aperture beamformer 116. The analog signal may be transmitted to one of a/D converters 122.
The plurality of a/D converters 122 may comprise suitable logic, circuitry, interfaces and/or code that may be operable to convert analog signals from the receiver 118 to corresponding digital signals. A plurality of a/D converters 122 are disposed between the receiver 118 and the RF processor 124. The present disclosure is not limited in this respect, though. Thus, in some embodiments, multiple a/D converters 122 may be integrated within receiver 118.
The RF processor 124 may comprise suitable logic, circuitry, interfaces and/or code that may be operable to demodulate digital signals output by the plurality of a/D converters 122. According to one embodiment, the RF processor 124 may include a complex demodulator (not shown) that may be used to demodulate the digital signals to form I/Q data pairs representative of corresponding echo signals. The RF or I/Q signal data may then be passed to an RF/IQ buffer 126. The RF/IQ buffer 126 may comprise suitable logic, circuitry, interfaces and/or code that may be operable to provide temporary storage of RF or I/Q signal data generated by the RF processor 124.
The receive beamformer 120 may comprise suitable logic, circuitry, interfaces and/or code that may be operable to perform digital beamforming processing to, for example, sum delayed channel signals received from the RF processor 124 via the RF/IQ buffer 126 and output a beamformed signal. The resulting processed information may be the beam summation signal output from the receive beamformer 120 and passed to the signal processor 132. According to some embodiments, the receiver 118, the plurality of a/D converters 122, the RF processor 124, and the beamformer 120 may be integrated into a single beamformer, which may be digital. In various embodiments, the ultrasound system 100 includes a plurality of receive beamformers 120.
The user input device 130 may be used to enter patient data, scan parameters, settings, select protocols and/or templates, etc. In an exemplary embodiment, the user input device 130 is operable to configure, manage and/or control the operation of one or more components and/or modules in the ultrasound system 100. In this regard, the user input device 130 may be used to configure, manage and/or control the operation of the transmitter 102, ultrasound probe 104, transmit beamformer 110, receiver 118, receive beamformer 120, RF processor 124, RF/IQ buffer 126, user input device 130, signal processor 132, image buffer 136, display system 134 and/or archive 138. User input device 130 may include one or more buttons, one or more rotary encoders, a touch screen, motion tracking, voice recognition, a mouse device, a keyboard, a camera, and/or any other device capable of receiving user instructions. In certain embodiments, for example, one or more of the user input devices 130 may be integrated into other components (such as the display system 134 or the ultrasound probe 104). For example, the user input device 130 may include a touch screen display.
The signal processor 132 may comprise suitable logic, circuitry, interfaces and/or code that may be operable to process the ultrasound scan data (i.e., the summed IQ signals) to generate an ultrasound image for presentation on the display system 134. The signal processor 132 is operable to perform one or more processing operations according to a plurality of selectable ultrasound modalities on the acquired ultrasound scan data. In an exemplary embodiment, the signal processor 132 may be used to perform display processing and/or control processing, and the like. As echo signals are received, acquired ultrasound scan data may be processed in real-time during a scan session. Additionally or alternatively, the ultrasound scan data may be temporarily stored in the RF/IQ buffer 126 during a scan session and processed in a less real-time manner in an online operation or an offline operation. In various implementations, the processed image data may be presented at display system 134 and/or may be stored at archive 138. Archive 138 may be a local archive, Picture Archiving and Communication System (PACS), or any suitable device for storing images and related information.
The signal processor 132 may be one or more central processing units, graphics processing units, microprocessors, microcontrollers, or the like. For example, the signal processor 132 may be an integrated component, or may be distributed in various locations. In an exemplary embodiment, the signal processor 132 may include an analog acoustic power enhancement processor 140 and may be capable of receiving input information from the user input device 130 and/or the profile 138, generating output that may be displayed by the display system 134, and manipulating the output in response to input information from the user input device 130, and the like. The signal processor 132 and the analog acoustic power enhancement processor 140 may be capable of performing, for example, any of the methods and/or sets of instructions discussed herein in accordance with various embodiments.
The ultrasound system 100 is operable to continuously acquire ultrasound scan data at a frame rate appropriate for the imaging situation in question. Typical frame rates are in the range of 20-120, but may be lower or higher. The acquired ultrasound scan data may be displayed on the display system 134 at the same frame rate, or at a slower or faster display rate. An image buffer 136 is included for storing processed frames of acquired ultrasound scan data that are not scheduled for immediate display. Preferably, the image buffer 136 has sufficient capacity to store at least several minutes of frames of ultrasound scan data. The frames of ultrasound scan data are stored in a manner that is easily retrievable therefrom according to their acquisition order or time. The image buffer 136 may be embodied as any known data storage medium.
The signal processor 132 may comprise an analog acoustic power enhancement processor 140 comprising suitable logic, circuitry, interfaces and/or code that may be operable to process the low acoustic power ultrasound image to generate an ultrasound image that simulates high acoustic power. For example, the analog acoustic power enhancement processor 140 may receive ultrasound images acquired at low acoustic power, such as Contrast Enhanced Ultrasound (CEUS) reference images, obstetrical ultrasound images, ophthalmic ultrasound images, images with acoustic power levels within the U.S. Food and Drug Administration (FDA) limits, and/or any suitable ultrasound images acquired at low acoustic power. The analog acoustic power enhancement processor 140 may be configured to generate ultrasound images that enhance contrast resolution, spatial resolution, reduce noise, and the like, such that the generated ultrasound images appear to be acquired at higher acoustic power. In representative embodiments, the simulated acoustic power enhancement processor 140 may include an artificial intelligence model/deep neural network (e.g., a convolutional neural network) and/or may utilize any suitable form of artificial intelligence image processing techniques or machine learning processing functions configured to process ultrasound images acquired at low acoustic power to generate ultrasound images with simulated higher acoustic power.
The analog acoustic power enhancement processor 140 may comprise suitable logic, circuitry, interfaces and/or code that may be operable to enhance ultrasound images acquired at low acoustic power to simulate acquisition at high acoustic power. In various embodiments, the simulated acoustic power enhancement processor 140 may deploy a deep neural network (e.g., an artificial intelligence model) that may be composed of, for example, an input layer, an output layer, and one or more hidden layers between the input layer and the output layer. Each layer may be made up of a plurality of processing nodes, which may be referred to as neurons. For example, the simulated acoustic power enhancement processor 140 may infer an artificial intelligence model that includes an input layer with neurons for each pixel or group of pixels from a scan plane of the anatomical structure. The output layer may have neurons corresponding to enhanced pixels that simulate the acquisition at higher acoustic power of the generated ultrasound image. For example, the output layer may generate an enhanced ultrasound image with reduced noise and increased contrast resolution and spatial resolution to simulate a greater acoustic power than that of the acquisition of the input ultrasound image. Each neuron of each layer may perform a processing function and pass the processed ultrasound image information to one neuron of a plurality of neurons of a downstream layer for further processing. For example, neurons of the first layer may learn to identify structural edges in the ultrasound image data. Neurons of the second layer may learn to recognize shapes based on detected edges from the first layer. Neurons of the third layer may learn the location of the identified shape relative to landmarks in the ultrasound image data. The processing of the inferred deep neural network (e.g., convolutional neural network) performed by the analog acoustic power enhancement processor 140 may enhance the low acoustic power ultrasound image data with a high degree of probability.
In an exemplary embodiment, the analog acoustic power enhancement processor 140 may comprise suitable logic, circuitry, interfaces and/or code that may be operable to process the low acoustic power ultrasound image by registering and matching the low acoustic power ultrasound image with a previously acquired ultrasound image of the same patient anatomy acquired at a higher acoustic power. The simulated acoustic power enhancement processor 140 may be configured to blend and/or combine the registered and matched low acoustic power images and high acoustic power images to generate an enhanced ultrasound image simulating an acoustic power higher than that of the low acoustic power ultrasound image.
The analog acoustic power enhancement processor 140 may comprise suitable logic, circuitry, interfaces and/or code that may be operable to register and match the low acoustic power ultrasound images with the high acoustic power ultrasound images. In various embodiments, the analog acoustic power enhancement processor 140 may deploy a deep neural network that may be composed of, for example, an input layer, an output layer, and one or more hidden layers between the input and output layers. Each layer may be made up of a plurality of processing nodes, which may be referred to as neurons. For example, the simulated acoustic power enhancement processor 140 may infer an artificial intelligence model that includes an input layer with neurons for each pixel or group of pixels from a scan plane of the anatomical structure. The output layer may have neurons corresponding to the locations of the landmarks and/or identified anatomical structures. For example, the output layer may include neurons for neural locations, vascular locations, bone locations, organ locations, or any suitable anatomical locations. Each neuron of each layer may perform a processing function and pass the processed ultrasound image information to one neuron of a plurality of neurons of a downstream layer for further processing. For example, neurons of the first layer may learn to identify structural edges in the ultrasound image data. Neurons of the second layer may learn to recognize shapes based on detected edges from the first layer. Neurons of the third layer may learn the location of the identified shape relative to landmarks in the ultrasound image data. The process of deploying a deep neural network (e.g., a convolutional neural network) performed by the analog acoustic power enhancement processor 140 can register and match anatomical structures in the low acoustic power ultrasound images and the high acoustic power ultrasound images with a high degree of probability.
Still referring to FIG. 1, the display system 134 may be any device capable of communicating visual information to a user. For example, the display system 134 may include a liquid crystal display, a light emitting diode display, and/or any suitable display or displays. The display system 134 may be operable to present ultrasound images and/or any suitable information. For example, the display system 134 may be operable to simulate an enhanced ultrasound image acquired at high acoustic power, or any suitable information, in the presentation.
The archive 138 may be one or more computer-readable memories integrated with the ultrasound system 100 and/or communicatively coupled (e.g., over a network) to the ultrasound system 100, such as a Picture Archiving and Communication System (PACS), a server, a hard disk, a floppy disk, a CD-ROM, a DVD, a compact storage device, a flash memory, a random access memory, a read-only memory, an electrically erasable and programmable read-only memory, and/or any suitable memory. The archive 138 may include, for example, a database, library, information set, or other memory accessed by the signal processor 132 and/or incorporated into the signal processor 132. For example, the archive 138 can store data temporarily or permanently. The archive 138 may be capable of storing medical image data, data generated by the signal processor 132, and/or instructions readable by the signal processor 132, among others. In various embodiments, the archive 138 stores, for example, ultrasound images acquired at low acoustic power, ultrasound images acquired at high acoustic power, enhanced ultrasound images simulating high acoustic power, an artificial intelligence model deployable to generate the enhanced ultrasound images, an artificial intelligence model for registering and matching the high acoustic power ultrasound images and the low acoustic power ultrasound images, and instructions for blending and/or combining the registered and matched high acoustic power ultrasound images and low acoustic power ultrasound images.
The components of the ultrasound system 100 may be implemented in software, hardware, firmware, etc. The various components of the ultrasound system 100 may be communicatively connected. The components of the ultrasound system 100 may be implemented separately and/or integrated in various forms. For example, the display system 134 and the user input device 130 may be integrated as a touch screen display.
Still referring to fig. 1, training system 200 may include a training engine 210 and a training database 220. The training engine 160 may comprise suitable logic, circuitry, interfaces and/or code that may be operable to train neurons of a deep neural network (e.g., an artificial intelligence model) that are inferred (i.e., deployed) by the simulated acoustic power enhancement processor 140. For example, an artificial intelligence model inferred by the simulated acoustic power enhancement processor 140 may be trained to automatically process low acoustic power ultrasound images to generate ultrasound images that simulate acquisition at higher acoustic power by: the contrast resolution, spatial resolution, and noise of received ultrasound images acquired at low acoustic power are enhanced. For example, the training engine 210 may train a deep neural network deployed by the simulated acoustic power enhancement processor 140 using a database 220 of classified paired ultrasound images of various structures. The pair of ultrasound images may include a first ultrasound image acquired at a low acoustic power and a second ultrasound image of the same anatomy acquired at a high acoustic power. In various embodiments, a tissue phantom may be used to simulate the anatomy of a high acoustic power ultrasound image such that FDA limits may be exceeded during an ultrasound scan. For example, the artificial intelligence model inferred by the simulated acoustic power enhancement processor 140 may be trained by the training engine 210 with paired ultrasound images having different acoustic power acquisition levels to train the model deployed by the simulated acoustic power enhancement processor 140 with respect to characteristics of image data acquired at the different acoustic power levels, such as appearance of structural edges, appearance of edge-based structural shapes, appearance and/or presence of noise, and so forth.
As another example, a deep neural network inferred by the analog acoustic power enhancement processor 140 may be trained to automatically register and match ultrasound images acquired at low acoustic power with ultrasound images acquired at high acoustic power. For example, the training engine 210 may train a deep neural network inferred by the analog acoustic power enhancement processor 140 using a database 220 of classified ultrasound images of various structures. The ultrasound images may include images of the anatomy acquired at different acoustic power levels. For example, the deep neural network inferred by the simulated acoustic power enhancement processor 140 may be trained by the training engine 210 with ultrasound images of various anatomical structures at different acoustic power acquisition levels to train the deep neural network deployed by the simulated acoustic power enhancement processor 140 with respect to characteristics of the anatomical structures depicted in the acquired image data (such as appearance of structural edges, appearance of edge-based structural shapes, location of shapes with respect to landmarks in the ultrasound image data, and so forth). In exemplary embodiments, the structural location may include a neural location, a vascular location, a bone location, an organ location, or any suitable anatomical location. The structural information may include information about the edges, shape, and location of organs, nerves, blood vessels, tissues, and the like.
In various embodiments, the database of training images 220 may be a Picture Archiving and Communication System (PACS) or any suitable data storage medium. In certain embodiments, the training engine 210 and/or the training image database 220 may be a remote system communicatively coupled to the ultrasound system 100 via a wired or wireless connection, as shown in fig. 1. Additionally and/or alternatively, components or all of the training system 200 may be integrated with the ultrasound system 100 in various forms.
Fig. 2 is a flow diagram 300 illustrating exemplary steps 302 and 306 that may be used to enhance ultrasound images acquired at low acoustic power to simulate acquisition at high acoustic power, in accordance with various embodiments. Referring to fig. 2, a flowchart 300 is shown that includes exemplary steps 302-306. Certain embodiments may omit one or more steps, and/or perform steps in a different order than the order listed, and/or combine certain steps discussed below. For example, some steps may not be performed in certain embodiments. As another example, certain steps may be performed in a different temporal order than listed below, including concurrently.
At step 302, the ultrasound system 100 acquires ultrasound images at low acoustic power. For example, the ultrasound system 100 may acquire ultrasound images with the ultrasound probe 104 positioned at a scanning location over the region of interest. The acoustic power applied to acquire the images may be low level, such as within FDA limits and/or according to AIUM and/or BMUS guidelines. The acquired ultrasound image may be a CEUS image, an obstetrical ultrasound image, an ophthalmic ultrasound image, or any suitable image.
At step 304, the signal processor 132 of the ultrasound system 100 may process the low acoustic power ultrasound image by inferring an artificial intelligence model to generate a second ultrasound image that simulates high acoustic power. For example, the analog acoustic power enhancement processor 140 of the signal processor 132 may be configured to enhance the contrast resolution, spatial resolution, reduce noise thereof, etc. of the ultrasound image acquired at step 302 to generate an ultrasound image having the appearance of an image acquired at an acoustic power higher than the acoustic power at which the acquired ultrasound image was obtained. The simulated acoustic power enhancement processor 140 may deploy artificial intelligence models, artificial intelligence image processing algorithms, and/or may utilize any suitable form of artificial intelligence image processing techniques or machine learning processing functions configured to process low acoustic power ultrasound images to generate enhanced ultrasound images simulating acquisition at higher acoustic power. For example, the ultrasound image acquired at step 302 may be acquired at an acoustic power below the FDA limit, and the ultrasound image generated at step 304 may simulate an acoustic power exceeding the FDA limit. As another example, the ultrasound image acquired at step 302 may be acquired with a microbubble contrast agent, and the ultrasound image generated at step 304 may be a simulated acoustic power that will disrupt microbubbles of the contrast agent.
At step 306, a second ultrasound image simulating high acoustic power may be presented at display system 134. For example, the analog acoustic power enhancement processor 140 of the signal processor 132 may present the enhanced ultrasound image generated at step 304 that simulates the acquisition at a higher acoustic power level at the display system 134.
Fig. 3 is a flow diagram 400 illustrating exemplary steps 402 and 410 that may be used to enhance ultrasound images acquired at low acoustic power to simulate acquisition at high acoustic power, in accordance with various embodiments. Referring to fig. 3, a flow chart 400 is shown that includes exemplary steps 402 through 410. Certain embodiments may omit one or more steps, and/or perform steps in a different order than the order listed, and/or combine certain steps discussed below. For example, some steps may not be performed in certain embodiments. As another example, certain steps may be performed in a different temporal order than listed below, including concurrently.
At step 402, the ultrasound system 100 acquires ultrasound images at high acoustic power. For example, the ultrasound system 100 may acquire ultrasound images with the ultrasound probe 104 positioned at a scanning location over the region of interest. The acoustic power applied to acquire the images may be at a high level, such as at or near the highest level allowed by FDA limits and/or prior to the introduction of microbubble contrast agents in CEUS examinations.
At step 404, the ultrasound system 100 acquires ultrasound images at low acoustic power. For example, the ultrasound system 100 may acquire ultrasound images with the ultrasound probe 104 positioned at a scanning location over the region of interest. The acoustic power applied to acquire the image may be lower than the acoustic power applied at step 402. For example, the acquired ultrasound image may be a CEUS image (i.e., after the microbubble contrast agent has been introduced) or any suitable image. In various embodiments, the image acquired at step 404 has the same patient and the same anatomy as the image acquired at step 402.
At step 406, the signal processor 132 of the ultrasound system 100 may register and match the low acoustic power ultrasound image and the high acoustic power ultrasound image. For example, the analog acoustic power enhancement processor 140 of the signal processor 132 may be configured to register and match the low acoustic power ultrasound image acquired at step 404 with the high acoustic power ultrasound image acquired at step 402. The analog acoustic power enhancement processor 140 may include artificial intelligence image processing algorithms, one or more deep neural networks (e.g., convolutional neural networks), and/or may utilize any suitable form of artificial intelligence image analysis techniques or machine learning processing functions configured to analyze, register, and match the low acoustic power ultrasound images and the high acoustic power ultrasound images.
At step 408, the signal processor 132 of the ultrasound system 100 may combine the low acoustic power ultrasound image and the high acoustic power ultrasound image to generate an ultrasound image simulating an acoustic power greater than the low acoustic power. For example, the analog acoustic power enhancement processor 140 of the signal processor 132 may be configured to combine and/or blend the low acoustic power ultrasound image and the high acoustic power ultrasound image registered and matched at step 406 to generate an enhanced ultrasound image that simulates an acoustic power greater than that of the low acoustic power ultrasound image.
At step 410, the generated ultrasound image simulating an acoustic power greater than the low acoustic power may be presented at display system 134. For example, the analog acoustic power enhancement processor 140 of the signal processor 132 may present the enhanced ultrasound image generated at step 408 that simulates the acquisition at a higher acoustic power level at the display system 134.
Aspects of the present disclosure provide methods 300, 400 and system 100 for enhancing ultrasound images acquired at low acoustic power to simulate acquisition at high acoustic power. According to various embodiments, the method 300, 400 may include acquiring 302, 404 a first ultrasound image by the ultrasound system 100 at a first acoustic power. The method 300, 400 may include processing 304, 406, 408, by the at least one processor 132, 140, the first ultrasound image to generate a second ultrasound image simulating a second acoustic power greater than the first acoustic power. The method 300, 400 may include rendering 306, 410 a second ultrasound image simulating second acoustic power at the display system 134.
In an exemplary embodiment, the artificial intelligence model may be inferred by the at least one processor 132, 140 to process the first ultrasound image to generate a second ultrasound image simulating the second acoustic power. In representative embodiments, the first ultrasound image may be acquired with a microbubble contrast agent and the simulated second acoustic power may be the power that will disrupt the microbubble contrast agent. In certain embodiments, the first acoustic power may be within Food and Drug Administration (FDA) limits. The second acoustic power may exceed the FDA limit. In various embodiments, the method 300 may include training an artificial intelligence model inferred by the at least one processor 132, 140 based on the pair-wise training images. Each pair of training images in the pair of training images may include a first training image acquired at an acoustic power and a second training image acquired at an acoustic power higher than the acoustic power of the first training image. In an exemplary embodiment, a second training image of each of the pairs of training images may be acquired from the mock tissue phantom. In a representative embodiment, the method 400 may include acquiring 402, by the ultrasound system 100, a third ultrasound image at a third acoustic power greater than the first acoustic power at the same region of interest prior to acquiring the first ultrasound image. The method 400 may include registering and matching 406, by the at least one processor 132, 140, the first ultrasound image with the third ultrasound image. Processing the first ultrasound image may include combining 408, by the at least one processor 132, 140, the first ultrasound image with the second ultrasound image to generate a second ultrasound image simulating a second acoustic power greater than the first acoustic power.
Various embodiments provide a system 100 for enhancing ultrasound images acquired at low acoustic power to simulate acquisition at high acoustic power. The system 100 may include an ultrasound system 100, at least one processor 132, 140, and a display system 134. The ultrasound system 100 may be configured to acquire a first ultrasound image at a first acoustic power. The at least one processor 132, 140 may be configured to process the first ultrasound image to generate a second ultrasound image simulating a second acoustic power greater than the first acoustic power. The display system 134 may be configured to present a second ultrasound image simulating the second acoustic power.
In a representative embodiment, the at least one processor 132, 140 may be configured to infer an artificial intelligence model to process the first ultrasound image to generate a second ultrasound image simulating the second acoustic power. In an exemplary embodiment, the ultrasound system 100 may be configured to acquire a first ultrasound image with a microbubble contrast agent and the simulated second acoustic power is the power that will break the microbubble contrast agent. In certain embodiments, the first acoustic power may be within Food and Drug Administration (FDA) limits. The second acoustic power may exceed the FDA limit. In various embodiments, the artificial intelligence model inferred by the at least one processor 132, 140 may be trained based on pairs of training images. Each pair of training images in the pair of training images may include a first training image acquired at an acoustic power and a second training image acquired at an acoustic power higher than the acoustic power of the first training image. In representative embodiments, the second training image of each of the pairs of training images may be acquired from the mock tissue phantom. In an exemplary embodiment, the ultrasound system 100 may be configured to acquire a third ultrasound image at a third acoustic power greater than the first acoustic power at the same region of interest prior to acquiring the first ultrasound image. The at least one processor 132, 140 may be configured to register and match the first ultrasound image with the third ultrasound image. The at least one processor 132, 140 may be configured to process the first ultrasound image by: the first ultrasound image is combined with the second ultrasound image to generate a second ultrasound image simulating a second acoustic power greater than the first acoustic power.
Certain embodiments provide a non-transitory computer readable medium having stored thereon a computer program having at least one code segment. The at least one code section is executable by a machine for causing the machine to perform steps 300, 400. Steps 300, 400 may include receiving 302, 404 a first ultrasound image acquired at a first acoustic power. Steps 300, 400 may include processing 304, 406, 408 the first ultrasound image to generate a second ultrasound image simulating a second acoustic power greater than the first acoustic power. Steps 300, 400 may include rendering 306, 410 a second ultrasound image simulating second acoustic power at a display system.
In various embodiments, processing 304 the first ultrasound image to generate a second ultrasound image simulating the second acoustic power is performed by inferring an artificial intelligence model. In certain embodiments, the first ultrasound image may be acquired with a microbubble contrast agent and the simulated second acoustic power may be the power that will disrupt the microbubble contrast agent. In representative embodiments, the first acoustic power may be within Food and Drug Administration (FDA) limits. The second acoustic power may exceed the FDA limit. In an exemplary embodiment, steps 300, 400 may include training an artificial intelligence model based on the pair-wise training images. Each pair of training images in the pair of training images may include a first training image acquired at an acoustic power and a second training image acquired at an acoustic power higher than the acoustic power of the first training image. In various embodiments, step 400 may include receiving 402 a third ultrasound image acquired at a third acoustic power greater than the first acoustic power at the same region of interest prior to receiving the first ultrasound image. Step 400 may include registering and matching 406 the first ultrasound image with the third ultrasound image. The processing 406, 408 of the first ultrasound image may include combining 408 the first ultrasound image with the second ultrasound image to generate a second ultrasound image simulating a second acoustic power greater than the first acoustic power.
As used herein, the term "circuitry" refers to physical electronic components (i.e., hardware) as well as configurable hardware, any software and/or firmware ("code") executed by and/or otherwise associated with hardware. For example, as used herein, a particular processor and memory may comprise first "circuitry" when executing one or more first codes and may comprise second "circuitry" when executing one or more second codes. As used herein, "and/or" means any one or more of the items in the list joined by "and/or". For example, "x and/or y" represents any element of the three-element set { (x), (y), (x, y) }. As another example, "x, y, and/or z" represents any element of the seven-element set { (x), (y), (z), (x, y), (x, z), (y, z), (x, y, z) }. The term "exemplary", as used herein, means serving as a non-limiting example, instance, or illustration. As used herein, the terms "e.g., (e.g.)" and "e.g., (for example)" bring forth a list of one or more non-limiting examples, instances, or illustrations. As used herein, a circuit is "operable to" and/or "configured to" perform a function whenever the circuit includes the necessary hardware and code (if needed) to perform the function, regardless of whether execution of the function is disabled or not enabled by certain user-configurable settings.
Other embodiments may provide a computer-readable device and/or a non-transitory computer-readable medium, and/or a machine-readable device and/or a non-transitory machine-readable medium having stored thereon machine code and/or a computer program having at least one code section executable by a machine and/or a computer, thereby causing the machine and/or the computer to perform steps as described herein for enhancing ultrasound images acquired at low acoustic power to simulate acquisition at high acoustic power.
Accordingly, the present disclosure may be realized in hardware, software, or a combination of hardware and software. The present disclosure may be realized in a centralized fashion in at least one computer system, or in a distributed fashion where different elements are spread across several interconnected computer systems. Any kind of computer system or other apparatus adapted for carrying out the methods described herein is suited.
Various embodiments may also be embedded in a computer program product, which comprises all the features enabling the implementation of the methods described herein, and which when loaded in a computer system is able to carry out these methods. Computer program in the present context means any expression, in any language, code or notation, of a set of instructions intended to cause a system having an information processing capability to perform a particular function either directly or after either or both of the following: a) conversion to another language, code or notation; b) replication takes place in different physical forms.
While the disclosure has been described with reference to certain embodiments, it will be understood by those skilled in the art that various changes may be made and equivalents may be substituted without departing from the scope of the disclosure. In addition, many modifications may be made to adapt a particular situation or material to the teachings of the disclosure without departing from the scope thereof. Therefore, it is intended that the disclosure not be limited to the particular embodiments disclosed, but that the disclosure will include all embodiments falling within the scope of the appended claims.
Claims (20)
1. A method, the method comprising:
acquiring, by an ultrasound system, a first ultrasound image at a first acoustic power;
processing, by at least one processor, the first ultrasound image to generate a second ultrasound image simulating a second acoustic power greater than the first acoustic power; and
rendering, at a display system, the second ultrasound image simulating the second acoustic power.
2. The method of claim 1, wherein an artificial intelligence model is inferred by the at least one processor to process the first ultrasound image to generate the second ultrasound image simulating the second acoustic power.
3. The method of claim 2, wherein the first ultrasound image is acquired with a microbubble contrast agent and the simulated second acoustic power is the power that will disrupt the microbubble contrast agent.
4. The method of claim 2, wherein the first acoustic power is within Food and Drug Administration (FDA) limits, and wherein the second acoustic power exceeds the FDA limits.
5. The method of claim 2, comprising training the artificial intelligence model inferred by the at least one processor based on pairs of training images, each pair of the pairs of training images comprising a first training image acquired at an acoustic power and a second training image acquired at an acoustic power higher than the acoustic power of the first training image.
6. The method of claim 5, wherein the second training image of each of the pairs of training images is acquired from a mock tissue phantom.
7. The method of claim 1, comprising:
acquiring, by the ultrasound system, a third ultrasound image at a third acoustic power greater than the first acoustic power at the same region of interest prior to acquiring the first ultrasound image, and
registering and matching, by the at least one processor, the first ultrasound image with the third ultrasound image,
wherein the processing the first ultrasound image comprises combining, by the at least one processor, the first ultrasound image with the second ultrasound image to generate the second ultrasound image simulating the second acoustic power greater than the first acoustic power.
8. A system, the system comprising:
an ultrasound system configured to acquire a first ultrasound image at a first acoustic power;
at least one processor configured to process the first ultrasound image to generate a second ultrasound image simulating a second acoustic power greater than the first acoustic power; and
a display system configured to present the second ultrasound image simulating the second acoustic power.
9. The system of claim 8, wherein the at least one processor is configured to infer an artificial intelligence model to process the first ultrasound image to generate the second ultrasound image simulating the second acoustic power.
10. The system of claim 9, wherein the ultrasound system is configured to acquire the first ultrasound image with a microbubble contrast agent, and the simulated second acoustic power is a power that will disrupt the microbubble contrast agent.
11. The system of claim 9, wherein the first acoustic power is within Food and Drug Administration (FDA) limits, and wherein the second acoustic power exceeds the FDA limits.
12. The system of claim 9, wherein the artificial intelligence model inferred by the at least one processor is trained based on pairs of training images, each pair of the pairs of training images comprising a first training image acquired at an acoustic power and a second training image acquired at an acoustic power higher than the acoustic power of the first training image.
13. The system of claim 12, wherein the second training image of each of the pair of training images is acquired from a mock tissue phantom.
14. The system of claim 8, wherein:
the ultrasound system is configured to acquire a third ultrasound image at a third acoustic power greater than the first acoustic power before acquiring the first ultrasound image at the same region of interest,
the at least one processor is configured to register and match the first ultrasound image with the third ultrasound image, and
the at least one processor is configured to process the first ultrasound image by: combining the first ultrasound image with the second ultrasound image to generate the second ultrasound image simulating the second acoustic power greater than the first acoustic power.
15. A non-transitory computer readable medium having stored thereon a computer program having at least one code section executable by a machine to cause the machine to perform steps comprising:
receiving a first ultrasound image acquired at a first acoustic power;
processing the first ultrasound image to generate a second ultrasound image simulating a second acoustic power greater than the first acoustic power; and
rendering, at a display system, the second ultrasound image simulating the second acoustic power.
16. The non-transitory computer-readable medium of claim 15, wherein the processing the first ultrasound image to generate the second ultrasound image simulating the second acoustic power is performed by inferring an artificial intelligence model.
17. The non-transitory computer readable medium of claim 16, wherein the first ultrasound image is acquired with a microbubble contrast agent and the simulated second acoustic power is a power that will disrupt the microbubble contrast agent.
18. The non-transitory computer-readable medium of claim 16, wherein the first acoustic power is within Food and Drug Administration (FDA) limits, and wherein the second acoustic power exceeds the FDA limits.
19. The non-transitory computer-readable medium of claim 16, comprising training the artificial intelligence model based on paired training images, each pair of the paired training images comprising a first training image acquired at an acoustic power and a second training image acquired at an acoustic power higher than the acoustic power of the first training image.
20. The non-transitory computer readable medium of claim 15, comprising:
receiving a third ultrasound image acquired at a third acoustic power greater than the first acoustic power at the same region of interest prior to receiving the first ultrasound image, and
registering and matching the first ultrasound image with the third ultrasound image,
wherein the processing the first ultrasound image comprises combining the first ultrasound image with the second ultrasound image to generate the second ultrasound image simulating the second acoustic power greater than the first acoustic power.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/783,831 | 2020-02-06 | ||
US16/783,831 US20210244387A1 (en) | 2020-02-06 | 2020-02-06 | Method and system for providing enhanced ultrasound images simulating acquisition at high acoustic power by processing ultrasound images acquired at low acoustic power |
Publications (1)
Publication Number | Publication Date |
---|---|
CN113208645A true CN113208645A (en) | 2021-08-06 |
Family
ID=77084665
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110174601.4A Pending CN113208645A (en) | 2020-02-06 | 2021-02-07 | Method and system for providing enhanced ultrasound images |
Country Status (2)
Country | Link |
---|---|
US (1) | US20210244387A1 (en) |
CN (1) | CN113208645A (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP7428020B2 (en) * | 2020-03-06 | 2024-02-06 | コニカミノルタ株式会社 | Ultrasonic diagnostic equipment, ultrasound image generation method, program and model training device |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107610195A (en) * | 2017-07-28 | 2018-01-19 | 上海联影医疗科技有限公司 | The system and method for image conversion |
CN109620288A (en) * | 2019-01-16 | 2019-04-16 | 王立坤 | A kind of department of general surgery's abdominal ultrasonic diagnostic device and its application method |
CN110719755A (en) * | 2017-06-08 | 2020-01-21 | 皇家飞利浦有限公司 | Ultrasonic imaging method |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5833613A (en) * | 1996-09-27 | 1998-11-10 | Advanced Technology Laboratories, Inc. | Ultrasonic diagnostic imaging with contrast agents |
IT1306193B1 (en) * | 1999-01-05 | 2001-05-30 | Lorenzo Manzoni | USE OF ULTRASOUNDS FOR THE TREATMENT OF DADECOMPRESSION SYMPTOMATOLOGY. |
CN104116523B (en) * | 2013-04-25 | 2016-08-03 | 深圳迈瑞生物医疗电子股份有限公司 | A kind of ultrasonic image analyzes system and the method for analysis thereof |
WO2018127498A1 (en) * | 2017-01-05 | 2018-07-12 | Koninklijke Philips N.V. | Ultrasound imaging system with a neural network for image formation and tissue characterization |
US20200060652A1 (en) * | 2018-08-23 | 2020-02-27 | The Board Of Trustees Of The Leland Stanford Junior University | Neural networks-assisted contrast ultrasound imaging |
-
2020
- 2020-02-06 US US16/783,831 patent/US20210244387A1/en active Pending
-
2021
- 2021-02-07 CN CN202110174601.4A patent/CN113208645A/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110719755A (en) * | 2017-06-08 | 2020-01-21 | 皇家飞利浦有限公司 | Ultrasonic imaging method |
CN107610195A (en) * | 2017-07-28 | 2018-01-19 | 上海联影医疗科技有限公司 | The system and method for image conversion |
CN109620288A (en) * | 2019-01-16 | 2019-04-16 | 王立坤 | A kind of department of general surgery's abdominal ultrasonic diagnostic device and its application method |
Also Published As
Publication number | Publication date |
---|---|
US20210244387A1 (en) | 2021-08-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11364013B2 (en) | Ultrasound imaging system with a neural network for image formation and tissue characterization | |
US11903768B2 (en) | Method and system for providing ultrasound image enhancement by automatically adjusting beamformer parameters based on ultrasound image analysis | |
US20220071595A1 (en) | Method and system for adapting user interface elements based on real-time anatomical structure recognition in acquired ultrasound image views | |
CN113116387A (en) | Method and system for providing guided workflow through a series of ultrasound image acquisitions | |
CN113558658A (en) | Method and system for providing ultrasound image | |
CN113208645A (en) | Method and system for providing enhanced ultrasound images | |
CN112515747A (en) | Method and system for analyzing ultrasound scenes to provide needle guidance and warning | |
US8858442B2 (en) | Ultrasonic diagnostic apparatus and ultrasonic image processing apparatus | |
US20220061809A1 (en) | Method and system for providing an anatomic orientation indicator with a patient-specific model of an anatomical structure of interest extracted from a three-dimensional ultrasound volume | |
US11903898B2 (en) | Ultrasound imaging with real-time visual feedback for cardiopulmonary resuscitation (CPR) compressions | |
US11219430B2 (en) | Method and system for automatically providing artifact warnings in pulsed-wave doppler imaging | |
JP2022543540A (en) | Ultrasound system sound power control using image data | |
US20210390685A1 (en) | Method and system for providing clutter suppression in vessels depicted in b-mode ultrasound images | |
US20230148998A1 (en) | Method and system for dynamically adjusting imaging parameters during an ultrasound scan | |
US20230084230A1 (en) | Method and system for reconstructing high resolution versions of low resolution images of a cine loop sequence | |
US20230404533A1 (en) | System and method for automatically tracking a minimal hiatal dimension plane of an ultrasound volume in real-time during a pelvic floor examination | |
US20220160334A1 (en) | Method and system for enhanced visualization of a pleural line by automatically detecting and marking the pleural line in images of a lung ultrasound scan | |
US20230342917A1 (en) | Method and system for automatic segmentation and phase prediction in ultrasound images depicting anatomical structures that change over a patient menstrual cycle | |
JP5449738B2 (en) | Ultrasonic image diagnosis apparatus, image processing apparatus, and ultrasonic image diagnosis support program | |
US20240041430A1 (en) | Method and system for defining a boundary of a region of interest by applying threshold values to outputs of a probabilistic automatic segmentation model based on user-selected segmentation sensitivity levels | |
US20220280133A1 (en) | Method and system for automatically detecting an ultrasound image view and focus to provide measurement suitability feedback | |
US20230248331A1 (en) | Method and system for automatic two-dimensional standard view detection in transesophageal ultrasound images | |
JP2008220415A (en) | Ultrasonic diagnostic system | |
CN116369971A (en) | Method and system for automatically setting pitch angle of mechanical swing type ultrasonic probe | |
JP2023108812A (en) | Information processing device, ultrasonic diagnostic device, and method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |