US12094080B2 - System and method for magnifying an image based on trained neural networks - Google Patents
System and method for magnifying an image based on trained neural networks Download PDFInfo
- Publication number
- US12094080B2 US12094080B2 US17/943,724 US202217943724A US12094080B2 US 12094080 B2 US12094080 B2 US 12094080B2 US 202217943724 A US202217943724 A US 202217943724A US 12094080 B2 US12094080 B2 US 12094080B2
- Authority
- US
- United States
- Prior art keywords
- image
- roi
- neural network
- magnification
- network model
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/46—Arrangements for interfacing with the operator or the patient
- A61B6/467—Arrangements for interfacing with the operator or the patient characterised by special input means
- A61B6/469—Arrangements for interfacing with the operator or the patient characterised by special input means for selecting a region of interest [ROI]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4046—Scaling of whole images or parts thereof, e.g. expanding or contracting using neural networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/0482—Interaction with lists of selectable items, e.g. menus
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/24—Indexing scheme for image data processing or generation, in general involving graphical user interfaces [GUIs]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
Definitions
- the aspects of the disclosed embodiments relate generally to the field of image processing, and more specifically, to a system and method for magnifying an image based on trained neural networks.
- radiology for example, mammography
- pathology for example, whole slide image (WSI)
- WSI whole slide image
- the combination of radiology and pathology may form the core of cancer diagnosis.
- the region of interest (ROI) at a site may be a mass (or a clot) of tissues (referred to as a tumor) or an abnormal deposition of calcium phosphates or other calcific salts (referred to as calcification).
- the boundary of such ROI may be a key factor for cancer diagnosis.
- the ROI is substantially small as compared to the entire medical image.
- a physician or a technician may be required to use a viewer to manually select the ROI in the medical image and then magnify such medical image by a zooming factor (such as 2 ⁇ , 4 ⁇ , and the like).
- a zooming factor such as 2 ⁇ , 4 ⁇ , and the like.
- traditional interpolation techniques such as bicubic interpolation or bi-linear interpolation, may be used.
- the boundary of the ROI may appear blurry, which may not be acceptable, as a clear visualization of such ROI is of utmost significance in medical findings and proper diagnosis. Therefore, there is required a method that performs magnification with clear visualization of the medical image (or the ROI).
- Systems and/or methods are provided for magnifying an image based on trained neural networks, substantially as shown in and/or described in connection with at least one of the figures, as set forth more completely in the claims.
- the aspects of the disclosed embodiments are directed to a magnification system.
- the magnification system includes a hardware processor.
- the hardware processor is configured to receive a first user input associated with a selection of a region of interest (ROI) within an input image of a site and receive a second user input associated with a first magnification factor of the selected ROI.
- the first magnification factor is associated with a magnification of the ROI in the input image.
- the hardware processor is also configured to modify the ROI based on an application of a first neural network model on the ROI.
- the modification of the ROI corresponds to a magnified image that is predicted in accordance with the first magnification factor.
- the hardware processor is also configured to control a display device to display the modified ROI.
- the aspects of the disclosed embodiments are directed to a computer implemented method.
- the method includes receiving, by a hardware processor, a first user input associated with a selection of a region of interest (ROI) within an input image of a site and receiving, by the hardware processor, a second user input associated with a first magnification factor of the selected ROI.
- the first magnification factor is associated with a magnification of the ROI in the input image.
- the method also includes modifying, by the hardware processor, the ROI based on an application of a first neural network model on the ROI, wherein the ROI is magnified in accordance with the first magnification factor and controlling, by the hardware processor, a display device to display the modified ROI.
- FIG. 1 is a block diagram that illustrates an exemplary environment for magnifying an image based on trained neural networks, in accordance with an exemplary embodiment of the disclosure.
- FIG. 2 is a block diagram that illustrates an exemplary magnification system for magnifying an image based on trained neural networks, in accordance with an embodiment of the disclosure.
- FIG. 3 A is a flowchart that illustrates exemplary operations for magnifying an image based on trained neural networks, in accordance with an embodiment of the disclosure.
- FIG. 3 B is a flowchart that illustrates training of a neural network model from the set of neural network models for magnifying an image, in accordance with an embodiment of the disclosure.
- FIG. 4 is a diagram that illustrates training of a neural network model for magnifying an image, in accordance with an embodiment of the disclosure.
- FIG. 5 is a schematic block diagram illustrating an example of a hardware implementation for a magnification system used for magnifying an image based on trained neural networks, in accordance with an exemplary embodiment of the disclosure.
- the aspects of the disclosed embodiments are generally directed to a method and system for magnifying an image based on trained neural networks.
- the various aspects of the disclosed embodiments provide a method and system that may correspond to a solution that utilizes a set of neural network models, each of which may be trained to magnify the image, in accordance with a specific magnification factor.
- Conventional image viewers may be based on conventional interpolation algorithms, where the user is required to magnify the region of interest (ROI) for several times folding to properly visualize the ROI.
- ROI region of interest
- Such interpolation of an image or the ROI introduces blurriness in the image each time the ROI is magnified.
- magnification system of the disclosed embodiments provides an improved workflow for a viewer using an artificial intelligence (AI)-based zoom-in algorithm which provide a clear visualization even after multifold magnification of the image or the ROI.
- AI artificial intelligence
- the disclosed magnification system, or an AI image viewer, in a super-resolution network can utilize an unsupervised learning method to intelligently generate a higher resolution image or image patch (or ROI) for a substantially clear visualization.
- FIG. 1 is a block diagram that illustrates an exemplary environment for magnifying an image based on trained neural networks, in accordance with an exemplary embodiment of the disclosure.
- a network environment 100 which may include a magnification system 102 , a set of neural network models 104 , a display device 106 , a server 108 , and a communication network 110 .
- the set of neural network models 104 may include a first neural network model 104 A, a second neural network model 104 B, up to an Nth neural network model 104 N.
- an input image 112 of a site 114 There is further shown an input image 112 of a site 114 , a region of interest (ROI) 116 within the input image 112 , and a modified ROI 118 .
- the display device 106 may be integrated within the magnification system 102 .
- the magnification system 102 includes a hardware processor 202 .
- the hardware processor 202 is configured to receive a first user input associated with a selection of a region of interest (ROI) within an input image of a site and receive a second user input associated with a first magnification factor of the selected ROI.
- the first magnification factor is associated with a magnification of the ROI in the input image.
- the hardware processor 202 is also configured to modify the ROI based on an application of a first neural network model on the ROI. The modification of the ROI corresponds to a magnified image that is predicted in accordance with the first magnification factor.
- the hardware processor 202 is also configured to control a display device 106 to display the modified ROI.
- the magnification system 102 may comprise suitable logic, circuitry, interfaces, and/or code that may be configured to magnify the ROI 116 within the input image 112 .
- the magnification system 102 may be further configured to apply the first neural network model 104 A from the set of neural network models 104 on the ROI 116 to magnify the ROI 116 .
- the magnification system 102 may be further configured to control the display device 106 to display the modified ROI 118 .
- Examples of the magnification system 102 may include, but are not limited to, a computing device, a mainframe machine, a server, a computer workstation, a smartphone, a cellular phone, a mobile phone, a gaming device, and/or a consumer electronic (CE) device with image processing capabilities.
- CE consumer electronic
- Each of the set of neural network models 104 may be a computational network or a system of artificial neurons, arranged in a plurality of layers, as nodes.
- the plurality of layers of each of the set of neural network models 104 may include an input layer, one or more hidden layers, and an output layer.
- Each layer of the plurality of layers may include one or more nodes (or artificial neurons).
- Outputs of all nodes in the input layer may be coupled to at least one node of hidden layer(s).
- inputs of each hidden layer may be coupled to outputs of at least one node in other layers of the corresponding neural network model.
- Outputs of each hidden layer may be coupled to inputs of at least one node in other layers of the corresponding neural network model.
- Node(s) in the final layer may receive inputs from at least one hidden layer to output a result.
- the number of layers and the number of nodes in each layer may be determined from hyper-parameters of the corresponding neural network model. Such hyper-parameters may be set before or while training the corresponding neural network model on a training dataset.
- Each of the set of neural network models 104 may correspond to a mathematical function (for example, a sigmoid function or a rectified linear unit) with a set of parameters, tunable during training of the network.
- the set of parameters may include, for example, a weight parameter, a regularization parameter, and the like.
- Each node may use the mathematical function to compute an output based on one or more inputs from nodes in other layer(s) (for example, previous layer(s)) of the corresponding neural network model. All or some of the nodes of the each of the set of neural network models 104 may correspond to the same or different mathematical function.
- one or more parameters of each node of the corresponding neural network model may be updated based on whether an output of the final layer for a given input (from the training dataset) matches a correct result based on a loss function for the corresponding neural network model.
- the above process may be repeated for the same or a different input until a minima of loss function may be achieved, and a training error may be minimized.
- Several methods for training are known in art, for example, gradient descent, stochastic gradient descent, batch gradient descent, gradient boost, meta-heuristics, and the like.
- Each of the set of neural network models 104 may include electronic data, such as a software program, code of the software program, libraries, applications, scripts, or other logic or instructions for execution by a processing device, such as hardware processor.
- Each of the set of neural network models 104 may include code and routines configured to enable a computing device, such as the magnification system 102 , to perform one or more operations.
- each of the set of neural network models 104 may be implemented using hardware including a processor, a microprocessor, a field-programmable gate array (FPGA), or an application-specific integrated circuit (ASIC) to perform or control performance of one or more operations.
- each of the set of neural network models 104 may be implemented using a combination of hardware and software.
- the set of neural network models 104 is shown as a separate entity from the magnification system 102 , the disclosure is not so limited. Accordingly, in some embodiments, the set of neural network models 104 may be integrated within the magnification system 102 , without deviation from scope of the disclosure. In an embodiment, the set of neural network models 104 may be stored in the server 108 .
- Examples of the each of the set of neural network models 104 may include, but are not limited to, a deep neural network (DNN), a convolutional neural network (CNN), a CNN-recurrent neural network (CNN-RNN), R-CNN, Fast R-CNN, Faster R-CNN, an artificial neural network (ANN), (You Only Look Once) YOLO network, a fully connected neural network, and/or a combination of such networks.
- DNN deep neural network
- CNN convolutional neural network
- CNN-RNN CNN-recurrent neural network
- R-CNN Fast R-CNN
- Faster R-CNN Faster R-CNN
- ANN artificial neural network
- the display device 106 may comprise suitable logic, circuitry, and interfaces that may be configured to display the modified ROI 118 .
- the display device 106 may be a touch screen which may enable the user to provide user inputs (such as a first user input, a second user input, and a third user input) via the display device 106 .
- the touch screen may be at least one of a resistive touch screen, a capacitive touch screen, or a thermal touch screen.
- the display device 106 may be realized through several known technologies.
- Examples of such technologies may include, but are not limited to, at least one of a Liquid Crystal Display (LCD) display, a Light Emitting Diode (LED) display, a plasma display, or an Organic LED (OLED) display technology, or other display devices.
- LCD Liquid Crystal Display
- LED Light Emitting Diode
- plasma display a plasma display
- OLED Organic LED
- the server 108 may include suitable logic, circuitry, and interfaces, and/or code that may be configured to store the input image 112 of the site 114 .
- the server 108 may be further configured to store one or more user inputs.
- the server 108 may be further configured to store the set of neural network models 104 .
- the server 108 may be configured to train each of the set of neural network models 104 .
- the server 108 may be implemented as a cloud server and may execute operations through web applications, cloud applications, HTTP requests, repository operations, file transfer, and the like.
- Other example implementations of the server 108 may include, but are not limited to, a database server, a file server, a web server, a media server, an application server, a mainframe server, or a cloud computing server.
- the server 108 may be implemented as a plurality of distributed cloud-based resources by use of several technologies that are well known to those ordinarily skilled in the art. A person with ordinary skill in the art will understand that the scope of the disclosure may not be limited to the implementation of the server 108 and the magnification system 102 as two separate entities. In certain embodiments, the functionalities of the server 108 can be incorporated in its entirety or at least partially in the magnification system 102 , without a departure from the scope of the disclosure.
- the communication network 110 may include a communication medium through which the magnification system 102 , the display device 106 , and the server 108 may communicate with each other.
- the communication network 110 may be one of a wired connection or a wireless connection. Examples of the communication network 110 may include, but are not limited to, the Internet, a cloud network, a Wireless Fidelity (Wi-Fi) network, a Personal Area Network (PAN), a Local Area Network (LAN), or a Metropolitan Area Network (MAN).
- Various devices in the network environment 100 may be configured to connect to the communication network 110 in accordance with various wired and wireless communication protocols.
- wired and wireless communication protocols may include, but are not limited to, at least one of a Transmission Control Protocol and Internet Protocol (TCP/IP), User Datagram Protocol (UDP), Hypertext Transfer Protocol (HTTP), File Transfer Protocol (FTP), Zig Bee, EDGE, IEEE 802.11, light fidelity (Li-Fi), 802.16, IEEE 802.11s, IEEE 802.11g, multi-hop communication, wireless access point (AP), device to device communication, cellular communication protocols, and Bluetooth (BT) communication protocols.
- TCP/IP Transmission Control Protocol and Internet Protocol
- UDP User Datagram Protocol
- HTTP Hypertext Transfer Protocol
- FTP File Transfer Protocol
- Zig Bee EDGE
- AP wireless access point
- BT Bluetooth
- the magnification system 102 may be configured to receive a first user input via the display device 106 .
- the first user input may be associated with a selection of the ROI 116 within the input image 112 of the site 114 .
- the selection of the ROI 116 may correspond to a selection of a specific area in the input image 112 of the site 114 .
- the input image 112 may correspond to a medical image and may be an MRI of brain of a patient. It should be noted that the aforesaid example of the input image 112 should not be construed to be limiting and other examples of the first image corresponding to other body parts of the patient may be possible, without deviation from the scope of the disclosure.
- the site 114 may correspond to a tumor in the input image 112 .
- the tumor may be visible as a white dot in the input image 112 .
- the ROI 116 in the input image 112 may correspond to a defined area around the tumor.
- the magnification system 102 may be configured to receive a second user input via the display device 106 .
- the second user input may be associated with a first magnification factor of the selected ROI 116 .
- the first magnification factor may correspond to a factor by which the ROI 116 within the input image 112 may be magnified (or enlarged).
- the first magnification factor may be associated with a magnification of the ROI 116 .
- the magnification of the selected ROI 116 may correspond to a magnified image that may be predicted in accordance with the first magnification factor so that details in the selected ROI 116 become more visible and clearer.
- the magnification system 102 may be further configured to select the first neural network model 104 A from the set of neural network models 104 based on the second user input.
- the first neural network model 104 A may be associated with the first magnification factor.
- the first neural network model 104 A may be trained to magnify the ROI 116 in accordance with the first magnification factor. Details about the training of each of the set of neural network models 104 is provided based on a training of an exemplary neural network model, for example, in FIGS. 3 B and 4 .
- the magnification system 102 may be further configured to apply the selected first neural network model 104 A on the ROI 116 .
- the magnification system 102 may be further configured to modify the ROI 116 based on an application of the selected first neural network model 104 A on the ROI 116 .
- the modification of the ROI 116 may correspond to a magnified image that may be predicted in accordance with the first magnification factor.
- the magnification system 102 may be configured to control the display device 106 to display the modified ROI 118 . Details about the modified ROI 118 are provided, for example, in FIG. 3 A .
- FIG. 2 is a block diagram that illustrates an exemplary magnification system for magnifying an image based on trained neural networks, in accordance with an embodiment of the disclosure.
- FIG. 2 is explained in conjunction with elements from FIG. 1 .
- a block diagram 200 of the magnification system 102 may include a hardware processor 202 , a memory 204 , an input/output (I/O) device 206 , a network interface 208 , an inference accelerator 210 , and the set of neural network models 104 .
- I/O input/output
- the magnification system 102 may include a hardware processor 202 , a memory 204 , an input/output (I/O) device 206 , a network interface 208 , an inference accelerator 210 , and the set of neural network models 104 .
- I/O input/output
- the hardware processor 202 may be communicatively coupled to the memory 204 , the I/O device 206 , the network interface 208 , the inference accelerator 210 , and the set of neural network models 104 .
- the I/O device 206 may further include the display device 106 .
- the hardware processor 202 may include suitable logic, circuitry, and interfaces that may be configured to execute program instructions associated with different operations to be executed by the magnification system 102 . For example, some of the operations may include, but are not limited to, receiving the first user input, receiving the second user input, modifying the ROI 116 , and controlling the display device 106 to display the modified ROI 118 .
- the hardware processor 202 may include one or more specialized processing units, which may be implemented as an integrated processor or a cluster of processors that perform the functions of the one or more specialized processing units, collectively.
- the hardware processor 202 may be implemented based on a number of processor technologies known in the art.
- Examples of implementations of the hardware processor 202 may be an x86-based processor, a Graphics Processing Unit (GPU), a Reduced Instruction Set Computing (RISC) processor, an Application-Specific Integrated Circuit (ASIC) processor, a Complex Instruction Set Computing (CISC) processor, a microcontroller, a central processing unit (CPU), and/or other computing circuits.
- GPU Graphics Processing Unit
- RISC Reduced Instruction Set Computing
- ASIC Application-Specific Integrated Circuit
- CISC Complex Instruction Set Computing
- microcontroller a central processing unit (CPU), and/or other computing circuits.
- the memory 204 may include suitable logic, circuitry, interfaces, and/or code that may be configured to store the program instructions to be executed by the hardware processor 202 .
- the memory 204 may store various images, such as the input image 112 , a first image, a second image, a third image, a fourth image, and a fifth image.
- the memory 204 may also store the set of neural network models 104 .
- the memory 204 may be further configured to store a first loss function and a second loss function.
- Examples of implementation of the memory 204 may include, but are not limited to, Random Access Memory (RAM), Read Only Memory (ROM), Electrically Erasable Programmable Read-Only Memory (EEPROM), Hard Disk Drive (HDD), a Solid-State Drive (SSD), a CPU cache, and/or a Secure Digital (SD) card.
- RAM Random Access Memory
- ROM Read Only Memory
- EEPROM Electrically Erasable Programmable Read-Only Memory
- HDD Hard Disk Drive
- SSD Solid-State Drive
- CPU cache volatile and/or a Secure Digital (SD) card.
- SD Secure Digital
- the I/O device 206 may include suitable logic, circuitry, and interfaces that may be configured to receive one or more user inputs and provide an output.
- the magnification system 102 may receive the first user input, the second user input, and the third user input, via the I/O device 206 .
- the I/O device 206 may further display the modified ROI 118 .
- the I/O device 206 which includes various input and output devices, may be configured to communicate with the hardware processor 202 . Examples of the I/O device 206 may include, but are not limited to, a touch screen, a keyboard, a mouse, a joystick, a microphone, a display device (such as the display device 106 ), and a speaker.
- the network interface 208 may include suitable logic, circuitry, and interfaces that may be configured to facilitate a communication between the hardware processor 202 , the set of neural network models 104 , the display device 106 , and the server 108 , via the communication network 110 .
- the network interface 208 may be implemented by use of various known technologies to support wired or wireless communication of the magnification system 102 with the communication network 110 .
- the network interface 208 may include, for example, an antenna, a radio frequency (RF) transceiver, one or more amplifiers, a tuner, one or more oscillators, a digital signal processor, a coder-decoder (CODEC) chipset, a subscriber identity module (SIM) card, or a local buffer circuitry.
- RF radio frequency
- the network interface 208 may be configured to communicate via wireless communication with networks, such as the Internet, an Intranet or a wireless network, such as a cellular telephone network, a public switched telephonic network (PSTN), a radio access network (RAN), a wireless local area network (LAN), and a metropolitan area network (MAN).
- networks such as the Internet, an Intranet or a wireless network, such as a cellular telephone network, a public switched telephonic network (PSTN), a radio access network (RAN), a wireless local area network (LAN), and a metropolitan area network (MAN).
- networks such as the Internet, an Intranet or a wireless network, such as a cellular telephone network, a public switched telephonic network (PSTN), a radio access network (RAN), a wireless local area network (LAN), and a metropolitan area network (MAN).
- PSTN public switched telephonic network
- RAN radio access network
- LAN wireless local area network
- MAN metropolitan area network
- the wireless communication may use one or more of a plurality of communication standards, protocols and technologies, such as Global System for Mobile Communications (GSM), Enhanced Data GSM Environment (EDGE), wideband code division multiple access (W-CDMA), Long Term Evolution (LTE), code division multiple access (CDMA), time division multiple access (TDMA), Bluetooth, Wireless Fidelity (Wi-Fi) (such as IEEE 802.11a, IEEE 802.11b, IEEE 802.11g or IEEE 802.11n), voice over Internet Protocol (VoIP), light fidelity (Li-Fi), Worldwide Interoperability for Microwave Access (Wi-MAX), a protocol for email, instant messaging, and a Short Message Service (SMS).
- GSM Global System for Mobile Communications
- EDGE Enhanced Data GSM Environment
- W-CDMA wideband code division multiple access
- LTE Long Term Evolution
- CDMA code division multiple access
- TDMA time division multiple access
- Wi-Fi Wireless Fidelity
- the inference accelerator 210 may include suitable logic, circuitry, interfaces, and/or code that may be configured to operate as a co-processor for the hardware processor 202 to accelerate computations associated with the operations of the each of the set of neural network models 104 for the magnification task.
- An example of an accelerated computation may be generation of the modified ROI 118 in less time than what is typically incurred without the use of the inference accelerator 210 .
- the inference accelerator 210 may implement various acceleration techniques, such as parallelization of some or all of the operations of the corresponding neural network model.
- the inference accelerator 210 may be implemented as a software, a hardware, or a combination thereof.
- Example implementations of the inference accelerator 210 may include, but are not limited to, a GPU, a Tensor Processing Unit (TPU), a neuromorphic chip, a Vision Processing Unit (VPU), a field-programmable gate arrays (FGPA), a Reduced Instruction Set Computing (RISC) processor, an Application-Specific Integrated Circuit (ASIC) processor, a Complex Instruction Set Computing (CISC) processor, a microcontroller, and/or a combination thereof.
- TPU Tensor Processing Unit
- VPU Vision Processing Unit
- FGPA field-programmable gate arrays
- RISC Reduced Instruction Set Computing
- ASIC Application-Specific Integrated Circuit
- CISC Complex Instruction Set Computing
- the functions or operations executed by the magnification system 102 may be performed by the hardware processor 202 .
- Various operations executed by the hardware processor 202 are described in detail, for example, in FIGS. 3 A, 3 B, 4 , and 5 .
- a first user input may be received.
- the hardware processor 202 in conjunction with the I/O device 206 and the network interface 208 of the magnification system 102 , may be configured to receive the first user input from a user of the magnification system 102 .
- the user may be, for example, a technician or a physician in a medical environment.
- the first user input may be received for the input image 112 of the site 114 .
- the input image 112 may correspond to a medical image generated when a medical scan, such as radiology, pathology, or the like, is performed on a specific infected body part, referred to as the site 114 .
- the input image 112 may be generated when an MRI scan is performed on a patient for detection of brain cancer in the patient.
- the input image 112 may be loaded by the user into a view on the display device 106 .
- the user may provide a first user input by use of an input device from the I/O device 206 .
- the first user input may be provided by the user by moving the mouse to the target, i.e., the ROI 116 .
- the first user input may be provided by the user by manipulating a first user interface (UI) element rendered on a display screen of the display device 106 .
- UI user interface
- the ROI 116 may correspond to an area in the input image 112 that may be of interest to the user.
- the input image 112 is a medical image of the brain of the patient generated by an MRI scan
- the ROI 116 may correspond to an area around a tumor (usually visible as a white dot) in the medical image. In such manner, the ROI 116 may be selected within the input image 112 of the site 114 .
- a second user input may be received.
- the hardware processor 202 in conjunction with the I/O device 206 and the network interface 208 of the magnification system 102 , may be configured to receive the second user input from the user.
- the second user input may be associated with a magnification factor of the selected ROI 116 .
- the magnification factor may be associated with a degree of magnification required for the ROI 116 .
- the magnification factor may correspond to a factor by which the ROI 116 may be magnified, enlarged, or zoomed in.
- the magnification factor may be one of 2 ⁇ , 4 ⁇ , 8 ⁇ , 10 ⁇ , and so on.
- the hardware processor 202 may be configured to control the display device 106 to render a user interface (UI) element on a display screen of the display device 106 .
- the second user input associated with the first magnification factor may be received via the rendered UI element.
- the second user input may be provided by the user by using a physical or virtual keypad from the I/O device 206 .
- the second user input may be provided by the user by manipulating the UI element rendered on a display screen of the display device 106 .
- the UI element may include a magnify ruler, an icon designated to initiate a magnifier, or the like.
- the user may be able to change the magnification factor, based on which the ROI 116 may be modified in real-time in accordance with the value of the updated magnification factor.
- a first neural network model may be selected.
- the hardware processor 202 in conjunction with the memory 204 of the magnification system 102 and the server 108 , may be configured to select the first neural network model 104 A from the set of neural network models 104 based on the second user input.
- the hardware processor 202 of the magnification system 102 may be configured to select the first neural network model 104 A based on a value of the first magnification factor included in the received second user input.
- each neural network from the set of neural network models 104 may be trained to magnify the ROI 116 or the input image 112 in entirety in accordance with the magnification factor.
- the first neural network model 104 A from the set of neural network models 104 may be trained to magnify the ROI 116 in accordance with the first magnification factor.
- the second neural network model 104 B from the set of neural network models 104 may be trained to magnify the ROI 116 in accordance with a second magnification factor.
- the Nth neural network model 104 N from the set of neural network models 104 may be trained to magnify the ROI 116 in accordance with an Nth magnification factor.
- the magnified ROI 116 may be predicted in accordance with the magnification factor based on which the corresponding neural network model is trained. Details about training each of the set of neural network models 104 are provided, for example, in FIGS. 3 B and 4 .
- the selected first neural network model 104 A may be applied on the ROI 116 .
- the hardware processor 202 in conjunction with the memory 204 , the inference accelerator 210 of the magnification system 102 and the server 108 , may be configured to apply the selected first neural network model 104 A on the ROI 116 .
- the selected first neural network model 104 A may be applied on the ROI 116 to modify the ROI 116 .
- the modified ROI 118 may correspond to a magnified image that is predicted in accordance with the first magnification factor.
- the ROI 116 may be modified.
- the hardware processor 202 in conjunction with the selected first neural network model 104 A, may be configured to modify the ROI 116 .
- the modification of the ROI 116 corresponds to a magnified image that is predicted in accordance with the first magnification factor.
- the modified ROI 118 may correspond to a magnified image that is predicted in accordance with the first magnification factor.
- the modified ROI 118 may provide additional and clearer details in comparison to the ROI 116 visualized in the input image 112 .
- a boundary of the tumor is a key factor in determination of the cancer or a type of cancer.
- the modified ROI 118 may provide additional and clearer details within the boundary of the tumor as compared to the ROI 116 visualized in the input image 112 .
- the display device may be controlled to display the modified ROI 118 .
- the hardware processor 202 in conjunction with the memory 204 and the display device 106 , may be configured to control the display device 106 to display the modified ROI 118 .
- the modified ROI 118 may be displayed on the user interface rendered on the display device 106 .
- the magnification system 102 may be configured to receive a third user input from the user, in a manner similar to the second user input, as described above.
- the received third user input may be associated with a second magnification factor different from the magnification factor.
- the magnification system 102 may be configured to select the second neural network model 104 B from the set of neural network models 104 .
- the magnification system 102 may further apply the selected second neural network model 104 B model on the ROI 116 .
- the magnification system 102 may further modify the ROI 116 based on the application of the second neural network model 104 B on the ROI 116 .
- the ROI 116 may be magnified in accordance with the second magnification factor.
- the first magnification factor may be ‘2 ⁇ ’
- the magnification system 102 may be configured to select the first neural network model 104 A.
- the magnification system 102 may be further configured to modify the ROI 116 based on the application of the first neural network model 104 A on the ROI 116 .
- the magnification system 102 may further control the display device 106 to display the modified ROI 118 .
- the magnification system 102 may receive the third user input associated with the second magnification factor of ‘3 ⁇ ’.
- the magnification system 102 may be configured to select the second neural network model 104 B and modify the ROI 116 based on the application of the second neural network model 104 B on the ROI 116 .
- the magnification system 102 may further control the display device 106 to display the modified ROI 118 . Therefore, the disclosed magnification system 102 may be capable of seamlessly selecting different neural network models based on the received user inputs and modify the ROI 116 accordingly.
- the visualization provided to the user may be a smooth transit from one modification of the ROI 116 to another one.
- the value of the magnification factor may be a fractional value or a decimal value and may contain a natural number and a fractional part.
- the value of the magnification factor may be ‘x ⁇ y’. The natural number in such value is indicated as ‘x’ and the fractional part is indicated as ‘y’.
- the magnification system 102 may be configured to apply a nearest integer function on the value of the magnification factor based on a determination of the fractional part in the value of the magnification factor.
- the nearest integer function may round up the value to the nearest integer less than or equal to the given value.
- the magnification system 102 may be configured to calculate a new value of the magnification factor based on the application of the nearest integer function.
- the calculated new value may be equal to the nearest integer less than or equal to the given number. For example, if the value is ‘3.8’, then the calculated new value may be ‘4’ and if the value is ‘3.3’, then the calculated new value may be ‘3’.
- the magnification system 102 may be configured to select the third neural network model from the set of neural network models 104 based on the calculated new value. It should be noted that the visualization of the modified ROI 118 has a smooth transit in accordance with one magnification factor to another magnification factor.
- the magnification system 102 may display real-time and magnified result when the user moves a mouse over the ROI 116 displayed on the display device 106 .
- the magnification system 102 may update the result on the display device 106 in real-time when the mouse pointer is placed within or on the boundary of the ROI 116 and an event, such as scrolling the mid-button of the mouse, may trigger the update of the modified ROI 118 .
- FIG. 3 B is a flowchart that illustrates training of a neural network model from the set of neural network models for magnifying an image, in accordance with an embodiment of the disclosure.
- FIG. 3 B is explained in conjunction with elements from FIGS. 1 , 2 , 3 A, and 4 .
- FIG. 3 B there is shown a flowchart 300 B that illustrates exemplary operations from 320 to 334 , as described herein.
- the exemplary operations illustrated in the flowchart 300 B may start at 320 and may be performed by each of the set of neural network models 104 , such as an exemplary neural network model 402 (depicted in FIG.
- the exemplary neural network model 402 may be applied on a first image 404 (depicted in FIG. 4 ), or an ROI of the first image 404 .
- the hardware processor 202 may be configured to apply the exemplary neural network model 402 on the first image 404 or the ROI.
- the first image 404 may correspond to the medical image, such as the input image 112 .
- a second image 406 (depicted in FIG. 4 ) may be generated based on the application of the exemplary neural network model 402 on the first image or the ROI.
- the exemplary neural network model 402 may be configured to be applied on the first image 404 or the ROI to generate the second image 406 .
- the first image 404 may be modified in accordance with the first magnification factor based on the application of the exemplary neural network model 402 .
- the second image 406 may correspond to a modification of the first image 404 .
- a third image 408 (depicted in FIG. 4 ) may be generated based on an application of a downsampling operation on the second image 406 .
- the hardware processor 202 may be configured to generate the third image 408 based on an application of a downsampling operation on the second image 406 .
- the second image 406 may be downsampled in accordance with a resizing factor that may be associated with the first magnification factor.
- the resizing factor may be a reciprocal of the first magnification factor.
- a first loss function may be calculated based on a comparison of the first image 404 or the ROI with the generated third image 408 .
- the hardware processor 202 may be configured to calculate the first loss function based on a comparison of the first image 404 or the ROI with the generated third image 408 .
- the calculated first loss function may correspond to, but is not limited to, a pixel-wise loss function, or a perceptual loss function.
- a fourth image 410 (depicted in FIG. 4 ) may be generated based on the application of the downsampling operation on the first image 404 or the ROI.
- the hardware processor 202 may be configured to generate the fourth image 410 based on the application of the downsampling operation on the first image 404 or the ROI. Similar to the generation of the third image 408 , the hardware processor 202 may be configured to apply the downsampling operation on the first image 112 to generate the fourth image 410 .
- the first image 112 may be downsampled in accordance with the resizing factor that may be associated with the first magnification factor. In an embodiment, the resizing factor may be the reciprocal of the first magnification factor.
- a fifth image 412 (depicted in FIG. 4 ) may be generated based on the application of the exemplary neural network model 402 on the generated fourth image 410 .
- the exemplary neural network model 402 may be configured to be applied on the fourth image 410 to generate the fifth image 412 .
- a second loss function may be calculated based on the comparison of the first image 404 or the ROI with the generated fifth image 412 .
- the hardware processor 202 may be configured to calculate the second loss function based on the comparison of the first image 404 or the ROI with the generated fifth image 412 .
- the calculated second loss function may correspond to, but is not limited to, the pixel-wise loss function, or the perceptual loss function.
- the exemplary neural network model 402 may be trained based on the calculated first loss function and the calculated second loss function.
- the exemplary neural network model 402 may be trained based on the first image 404 or the ROI, the third image 408 , and the fifth image 412 .
- the hardware processor 202 may be configured to train the exemplary neural network model 402 based on the calculated first loss function and the calculated second loss function.
- FIG. 4 is an exemplary diagram that illustrates training of a neural network model from the set of neural network models for magnifying an image, in accordance with an embodiment of the disclosure.
- FIG. 4 is explained in conjunction with elements from FIGS. 1 , 2 , 3 A, and 3 B .
- FIG. 4 there is shown an exemplary diagram 400 that includes the exemplary neural network model 402 . It should be understood that the workflow of the exemplary diagram 400 may be applied to train the exemplary neural network model 402 that corresponds to each neural network model from the set of neural network models 104 .
- the workflow indicated in the exemplary diagram 400 corresponds to an unsupervised learning-based approach that may be used to train the exemplary neural network model 402 .
- the exemplary neural network model 402 may be configured to receive a medical image, such as the first image 404 of a body part of a patient.
- the first image 404 which is a low resolution (LR) image, may correspond to the ROI 116 or the medical image in entirety.
- the medical image may be generated using a magnetic resonance imaging (MRI) technique that is based on a magnetic field and computer-generated radio waves to create detailed images of the organs and tissues in a body part, which is the knee portion of the patient in the above embodiment.
- MRI magnetic resonance imaging
- the exemplary neural network model 402 may be configured to be applied on the first image 404 or a specific region, such as the ROI 116 , of the first image 404 . Based on the application of the exemplary neural network model 402 on the first image 404 or the specific region within the first image 404 , the exemplary neural network model 402 may be configured to generate the second image 406 .
- the second image 406 may correspond to a high resolution (HR) image with a magnification factor ‘N’.
- the first image 404 or the specific region within the first image 404 may correspond to a magnified image that is predicted in accordance with the first magnification factor, i.e., the magnification factor ‘N’.
- the third image 408 based may be generated based on an application of a downsampling operation on the second image 406 .
- the second image 406 may be downsampled to the third image 408 which may correspond to an LR image.
- the third image 408 may be generated by downsampling the second image 406 by a resizing factor, such as ‘1/N’, associated with the first magnification factor, i.e., the magnification factor ‘N’.
- the downsampling operation may reduce the size of the second image 406 based on the resizing factor to obtain the third image 408 .
- the third image 408 may include lesser pixels as compared to a number of pixels in the second image 406 .
- the resizing factor may be reciprocal of the first magnification factor.
- the modification of the first image 404 to the third image 408 using the exemplary neural network model 402 may correspond to an unsupervised learning with no HR ground truth as high-resolution ground truth images for generating ground truth images do not exist.
- the exemplary neural network model 402 may be configured to apply the downsampling operation on the first image 404 or the specific region within the first image 404 to generate the fourth image 410 .
- the fourth image 410 may be generated by downsampling the first image 404 or the specific region within the first image 404 by the resizing factor.
- the fourth image 410 may correspond to an LR image with a resizing factor ‘1/N’. Based on the application of the downsampling operation on the first image 404 or the specific region within the first image 404 , the exemplary neural network model 402 may generate the fourth image 410 .
- the fourth image 410 thus generated may include lesser pixels as compared to the number of pixels in the second image 406 .
- the downsampling operation applied on the first image 404 or the specific region within the first image 404 may reduce the size of the first image 404 or the specific region within the first image 404 based on the resizing factor to obtain the fourth image 410 .
- the fourth image 410 may include lesser pixels as compared to the number of pixels in the first image 404 or the specific region within the first image 404 .
- the fourth image 410 may be thereafter provided to the exemplary neural network model 402 .
- the exemplary neural network model 402 may be configured to generate the fifth image 412 , whose size may be same as the size of the first image 404 .
- the fifth image 412 may correspond to an LR image with a magnification factor ‘N’.
- the fifth image 412 may be magnified in accordance with the first magnification factor, i.e., the magnification factor ‘N’.
- the modification of the first image 404 to the fifth image 412 using the exemplary neural network model 402 may correspond to a self-supervised learning.
- a suitable logic, circuitry, interfaces, and/or code may be configured to calculate loss functions, such as a first loss function (L 1 ) and a second loss function (L 2 ) of the exemplary neural network model 402 . More specifically, the suitable logic, circuitry, interfaces, and/or code may calculate the first loss function (L 1 ) based on a comparison of the first image 404 with the generated third image 408 , and the second loss function (L 2 ) based on a comparison of the first image 404 with the generated fifth image 412 .
- the calculated first loss function (L 1 ) and the second loss function (L 2 ) may correspond to, but is not limited to, a pixel-wise loss function, or a perceptual loss function.
- the exemplary neural network model 402 may be trained and thereafter may be used to realize the magnification system 102 .
- the trained exemplary neural network model 402 may be used in real-time scenarios for magnification of any image (or ROI) by the magnification factor ‘N’.
- other similar instances of the exemplary neural network model 402 may be trained for different magnification factors, such as 2 ⁇ , 3 ⁇ , 4 ⁇ . . . , N ⁇ . Once trained, all such instances of the exemplary neural network model 402 may be stored in the memory 204 .
- magnification factor When the magnification factor is set based on the second user input at an exemplary magnifier realized by the magnification system 102 , a neural network model for corresponding magnification factor may be selected. Accordingly, the corresponding magnified image may be predicted by the magnification system 102 .
- FIG. 5 is a conceptual diagram illustrating an example of a hardware implementation for a magnification system used for magnifying an image based on trained neural networks, in accordance with an exemplary embodiment of the disclosure.
- the hardware implementation shown by a representation 500 for the network environment 100 employs a processing system 502 for magnifying an image based on trained neural networks, in accordance with an exemplary embodiment of the disclosure, as described herein.
- the processing system 502 may comprise a processor 504 , a non-transitory computer-readable medium 506 , a bus 508 , a bus interface 510 , and a transceiver 512 .
- the processor 504 such as the hardware processor 202 , may be configured to manage the bus 508 and general processing, including the execution of a set of instructions stored on the non-transitory computer-readable medium 506 .
- the set of instructions when executed by the processor 504 , causes the magnification system 102 to execute the various functions described herein for any particular apparatus.
- the processor 504 may be implemented, based on a number of processor technologies known in the art. Examples of the processor 504 may be RISC processor, ASIC processor, CISC processor, and/or other processors or control circuits.
- the non-transitory computer-readable medium 506 may be used for storing data that is manipulated by the processor 504 when executing the set of instructions.
- the data is stored for short periods or in the presence of power.
- the bus 508 may be configured to link together various circuits.
- the network environment 100 employing the processing system 502 and the non-transitory computer-readable medium 506 may be implemented with bus architecture, represented generally by bus 508 .
- the bus 508 may include any number of interconnecting buses and bridges depending on the specific implementation of the magnification system 102 and the overall design constraints.
- the bus interface 510 may be configured to provide an interface between the bus 508 and other circuits, such as the transceiver 512 , and external devices, such as the display device 106 , and the server 108 .
- the transceiver 512 may be configured to provide a communication of the magnification system 102 with various other apparatus, such as the display device 106 , via a network.
- the transceiver 512 may communicate via wireless communication with networks, such as the Internet, the Intranet and/or a wireless network, such as a cellular telephone network, a wireless local area network (WLAN) and/or a metropolitan area network (MAN).
- networks such as the Internet, the Intranet and/or a wireless network, such as a cellular telephone network, a wireless local area network (WLAN) and/or a metropolitan area network (MAN).
- WLAN wireless local area network
- MAN metropolitan area network
- the wireless communication may use any of a plurality of communication standards, protocols and technologies, such as 5th generation mobile network, Global System for Mobile Communications (GSM), Enhanced Data GSM Environment (EDGE), Long Term Evolution (LTE), wideband code division multiple access (W-CDMA), code division multiple access (CDMA), time division multiple access (TDMA), Bluetooth, Wireless Fidelity (Wi-Fi) (such as IEEE 802.11a, IEEE 802.11b, IEEE 802.11g and/or IEEE 802.11n), voice over Internet Protocol (VoIP), and/or Wi-MAX.
- GSM Global System for Mobile Communications
- EDGE Enhanced Data GSM Environment
- LTE Long Term Evolution
- W-CDMA wideband code division multiple access
- CDMA code division multiple access
- TDMA time division multiple access
- Wi-Fi Wireless Fidelity
- IEEE 802.11a, IEEE 802.11b, IEEE 802.11g and/or IEEE 802.11n voice over Internet Protocol (VoIP), and/or Wi-MAX.
- one or more components of FIG. 5 may include software whose corresponding code may be executed by at least one processor, for across multiple processing environments.
- the processor 504 may be configured or otherwise specially programmed to execute the operations or functionality of the hardware processor 202 , the memory 204 , the I/O device 206 , the network interface 208 , and the inference accelerator 210 or various other components described herein, as described with respect to FIGS. 1 to 5 .
- magnification system 102 that may be configured to magnify an image based on trained neural networks.
- the magnification system 102 may comprise, for example, the hardware processor 202 , the memory 204 , the I/O device 206 , and the network interface 208 , and/or the inference accelerator 210 .
- the hardware processor 202 of the magnification system 102 may be configured to receive a first user input associated with a selection of the ROI 116 within the input image 112 of the site 114 .
- the hardware processor 202 may be further configured to receive the second user input associated with the first magnification factor of the selected ROI 116 .
- the first magnification factor may be associated with the magnification of the ROI 116 .
- the hardware processor 202 of the magnification system 102 may be further configured to modify the ROI 116 based on an application of the first neural network model 104 A on the ROI 116 .
- the modification of the ROI 116 may correspond to a magnified image that may be predicted in accordance with the first magnification factor.
- the hardware processor 202 of the magnification system 102 may be further configured to control the display device 106 to display the modified ROI 118 .
- the magnification system 102 may execute further operations comprising receiving, by the hardware processor 202 , a first user input associated with a selection of a region of interest, such as the ROI 116 , within the input image 112 of the site 114 .
- the magnification system 102 may execute further operations comprising receiving a second user input associated with a first magnification factor of the selected ROI 116 .
- the first magnification factor may be associated with a magnification of the ROI 116 in the input image 112 .
- the magnification system 102 may execute further operations comprising modifying the ROI 116 based on an application of the first neural network model 104 A on the ROI 116 .
- the first neural network model 104 A may be trained to magnify the ROI 116 in accordance with the first magnification factor.
- a third user input may be received associated with a second magnification factor different from the first magnification factor.
- the magnification system 102 may execute further operations comprising selecting the second neural network model 104 B from a set of neural network models 104 based on the third user input.
- the magnification system 102 may execute further operations comprising applying the selected second neural network model 104 B on the ROI 116 .
- the magnification system 102 may execute further operations comprising modifying the ROI based on the application of the second neural network model 104 B on the ROI 116 .
- the ROI 116 may be magnified in accordance with the second magnification factor.
- the magnification system 102 may execute further operations comprising applying the exemplary neural network model 402 on the first image 404 or the ROI in the first image 404 .
- the magnification system 102 may execute further operations comprising generating the second image 406 based on the application of the exemplary neural network model 402 on the first image 404 or the ROI.
- the magnification system 102 may execute further operations comprising generating the third image 408 based on an application of a downsampling operation on the second image 406 .
- the magnification system 102 may execute further operations comprising generating the fourth image 410 based on the application of the downsampling operation on the first image 404 .
- the magnification system 102 may execute further operations comprising generating a fifth image 412 based on the application of the exemplary neural network model 402 on the generated fourth image 410 .
- the magnification system 102 may execute further operations comprising training the exemplary neural network model 402 based on the first image 404 , the third image 408 , and the fifth image 412 .
- the third image 408 may be generated by downsampling the second image 406 by a resizing factor associated with the first magnification factor.
- the fourth image 410 may be generated by downsampling the first image 404 or the ROI by the resizing factor associated with the first magnification factor.
- the resizing factor may be reciprocal of the first magnification factor.
- the magnification system 102 may execute further operations comprising calculating a first loss function based on a comparison of the first image 404 or the ROI with the generated third image 408 .
- the magnification system 102 may execute further operations comprising calculating a second loss function based on the comparison of the first image 404 or the ROI with the generated fifth image 412 .
- the exemplary neural network model 402 may be trained based on the calculated first loss function and the calculated second loss function.
- the magnification system 102 may execute further operations comprising determining a fractional part in a first value of the first magnification factor.
- the magnification system 102 may execute further operations comprising applying a nearest integer function on the first value of the first magnification factor based on a determination of the fractional part.
- the magnification system 102 may execute further operations comprising calculating a new value of the first magnification factor based on the application of the nearest integer function.
- the magnification system 102 may execute further operations comprising selecting a third neural network model from the set of neural network models 104 based on the calculated new value of the first magnification factor.
- circuitry is “operable” to perform a function whenever the circuitry comprises the necessary hardware and/or code (if any is necessary) to perform the function, regardless of whether performance of the function is disabled, or not enabled, by some user-configurable setting.
- Another embodiment of the disclosure may provide a non-transitory machine and/or computer-readable storage and/or media, having stored thereon, a machine code and/or a computer program having at least one code section executable by a machine and/or a computer, thereby causing the machine and/or computer to perform the steps as described herein for generating a novel molecular structure using a protein structure.
- the present disclosure may also be embedded in a computer program product, which comprises all the features enabling the implementation of the methods described herein, and which when loaded in a computer system is able to carry out these methods.
- Computer program in the present context means any expression, in any language, code or notation, either statically or dynamically defined, of a set of instructions intended to cause a system having an information processing capability to perform a particular function either directly or after either or both of the following: a) conversion to another language, code or notation; b) reproduction in a different material form.
- a software module may reside in RAM memory, flash memory, ROM memory, EPROM memory, EEPROM memory, registers, hard disk, physical and/or virtual disk, a removable disk, a CD-ROM, virtualized system or device such as a virtual server or container, or any other form of storage medium known in the art.
- An exemplary storage medium is communicatively coupled to the processor (including logic/code executing in the processor) such that the processor can read information from, and write information to, the storage medium. In the alternative, the storage medium may be integral to the processor.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Medical Informatics (AREA)
- General Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Optics & Photonics (AREA)
- Data Mining & Analysis (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Human Computer Interaction (AREA)
- High Energy & Nuclear Physics (AREA)
- Computational Linguistics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Computing Systems (AREA)
- Pathology (AREA)
- Radiology & Medical Imaging (AREA)
- Heart & Thoracic Surgery (AREA)
- Surgery (AREA)
- Animal Behavior & Ethology (AREA)
- Public Health (AREA)
- Veterinary Medicine (AREA)
- Apparatus For Radiation Diagnosis (AREA)
Abstract
Description
Claims (20)
Priority Applications (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US17/943,724 US12094080B2 (en) | 2022-09-13 | 2022-09-13 | System and method for magnifying an image based on trained neural networks |
| CN202311148430.3A CN117132471A (en) | 2022-09-13 | 2023-09-06 | Method and computer program product for magnifying an image |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US17/943,724 US12094080B2 (en) | 2022-09-13 | 2022-09-13 | System and method for magnifying an image based on trained neural networks |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20240087082A1 US20240087082A1 (en) | 2024-03-14 |
| US12094080B2 true US12094080B2 (en) | 2024-09-17 |
Family
ID=88854355
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US17/943,724 Active 2043-05-13 US12094080B2 (en) | 2022-09-13 | 2022-09-13 | System and method for magnifying an image based on trained neural networks |
Country Status (2)
| Country | Link |
|---|---|
| US (1) | US12094080B2 (en) |
| CN (1) | CN117132471A (en) |
Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20140301665A1 (en) * | 2011-12-26 | 2014-10-09 | Canon Kabushiki Kaisha | Image data generating apparatus, image data display system, and image data generating method |
| CN107133933A (en) | 2017-05-10 | 2017-09-05 | 广州海兆印丰信息科技有限公司 | Mammography X Enhancement Method based on convolutional neural networks |
| US20190042860A1 (en) * | 2017-08-04 | 2019-02-07 | Samsung Electronics Co., Ltd. | Method and apparatus of detecting object of interest |
| US20190333199A1 (en) * | 2018-04-26 | 2019-10-31 | The Regents Of The University Of California | Systems and methods for deep learning microscopy |
| US20200326526A1 (en) * | 2017-12-26 | 2020-10-15 | Aetherai Co. Ltd. | Control method for automated microscope system, microscope system and computer-readable storage medium |
| US11185294B2 (en) | 2016-02-26 | 2021-11-30 | The Trustees Of The University Of Pennsylvania | Super-resolution tomosynthesis imaging systems and methods |
-
2022
- 2022-09-13 US US17/943,724 patent/US12094080B2/en active Active
-
2023
- 2023-09-06 CN CN202311148430.3A patent/CN117132471A/en active Pending
Patent Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20140301665A1 (en) * | 2011-12-26 | 2014-10-09 | Canon Kabushiki Kaisha | Image data generating apparatus, image data display system, and image data generating method |
| US11185294B2 (en) | 2016-02-26 | 2021-11-30 | The Trustees Of The University Of Pennsylvania | Super-resolution tomosynthesis imaging systems and methods |
| CN107133933A (en) | 2017-05-10 | 2017-09-05 | 广州海兆印丰信息科技有限公司 | Mammography X Enhancement Method based on convolutional neural networks |
| US20190042860A1 (en) * | 2017-08-04 | 2019-02-07 | Samsung Electronics Co., Ltd. | Method and apparatus of detecting object of interest |
| US20200326526A1 (en) * | 2017-12-26 | 2020-10-15 | Aetherai Co. Ltd. | Control method for automated microscope system, microscope system and computer-readable storage medium |
| US20190333199A1 (en) * | 2018-04-26 | 2019-10-31 | The Regents Of The University Of California | Systems and methods for deep learning microscopy |
Also Published As
| Publication number | Publication date |
|---|---|
| CN117132471A (en) | 2023-11-28 |
| US20240087082A1 (en) | 2024-03-14 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| TWI832966B (en) | Method and apparatus for automated target and tissue segmentation using multi-modal imaging and ensemble machine learning models | |
| CN110060313B (en) | Image artifact correction method and system | |
| US10499857B1 (en) | Medical protocol change in real-time imaging | |
| JP7014910B2 (en) | Systems and methods for image processing | |
| KR20210140757A (en) | Network training method, image processing method and apparatus, electronic device and storage medium | |
| JP6700622B2 (en) | System and method for processing multimodal images | |
| US10275886B2 (en) | Tumor segmentation and tissue classification in 3D multi-contrast | |
| CN110163876B (en) | Left ventricular segmentation method, system, device and medium based on multi-feature fusion | |
| US20150131898A1 (en) | Blind image deblurring with cascade architecture | |
| KR20210153700A (en) | Image processing method and apparatus, electronic device, storage medium and computer program | |
| CN110378913B (en) | Image segmentation method, device, equipment and storage medium | |
| WO2021259391A2 (en) | Image processing method and apparatus, and electronic device and storage medium | |
| JP2015502120A (en) | Method and apparatus for non-local super-resolution reconstruction | |
| KR20220016212A (en) | Image processing method and apparatus, electronic device, storage medium and program product | |
| JP6448680B2 (en) | Image adjustment | |
| JP2021111388A (en) | Neural network architecture search system and method | |
| CN112396119A (en) | Image processing method and device, electronic equipment and storage medium | |
| CN112966592A (en) | Hand key point detection method, device, equipment and medium | |
| US12094080B2 (en) | System and method for magnifying an image based on trained neural networks | |
| CN114418931B (en) | Method and device for extracting residual lung lobes after operation, electronic equipment and storage medium | |
| KR20230156330A (en) | Efficient pose estimation through iterative refinement | |
| CN114723723A (en) | Medical image processing method, computer device and storage medium | |
| CN114022492A (en) | Blood vessel segmentation method, apparatus, computer equipment and storage medium | |
| CN114586065A (en) | Method and system for segmenting images | |
| Mohanty et al. | Towards synthetic generation of clinical rosacea images with GAN models |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: UII AMERICA, INC., MASSACHUSETTS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LIU, YIKANG;SUN, SHANHUI;CHEN, TERRENCE;SIGNING DATES FROM 20220906 TO 20220908;REEL/FRAME:061078/0883 Owner name: SHANGHAI UNITED IMAGING INTELLIGENCE CO., LTD., CHINA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:UII AMERICA, INC.;REEL/FRAME:061078/0992 Effective date: 20220912 |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| ZAAB | Notice of allowance mailed |
Free format text: ORIGINAL CODE: MN/=. |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |