WO2023053103A1 - Method and system for automatic scanning and focusing of uneven surfaces for identification and classification of particulates - Google Patents
Method and system for automatic scanning and focusing of uneven surfaces for identification and classification of particulates Download PDFInfo
- Publication number
- WO2023053103A1 WO2023053103A1 PCT/IB2022/059430 IB2022059430W WO2023053103A1 WO 2023053103 A1 WO2023053103 A1 WO 2023053103A1 IB 2022059430 W IB2022059430 W IB 2022059430W WO 2023053103 A1 WO2023053103 A1 WO 2023053103A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- focusing
- scanning
- classification
- identification
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 53
- 239000002245 particle Substances 0.000 claims description 34
- 238000012545 processing Methods 0.000 claims description 10
- 238000004458 analytical method Methods 0.000 claims description 9
- 230000003287 optical effect Effects 0.000 claims description 5
- 238000005054 agglomeration Methods 0.000 claims description 2
- 230000002776 aggregation Effects 0.000 claims description 2
- 238000009434 installation Methods 0.000 claims 1
- 230000002452 interceptive effect Effects 0.000 claims 1
- 238000007781 pre-processing Methods 0.000 claims 1
- 238000001514 detection method Methods 0.000 abstract description 5
- 230000008901 benefit Effects 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 4
- 230000010354 integration Effects 0.000 description 3
- 238000000399 optical microscopy Methods 0.000 description 3
- 238000011160 research Methods 0.000 description 3
- 230000001960 triggered effect Effects 0.000 description 3
- 230000001419 dependent effect Effects 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 238000003384 imaging method Methods 0.000 description 2
- 238000005259 measurement Methods 0.000 description 2
- 238000000386 microscopy Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000002360 preparation method Methods 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 239000000443 aerosol Substances 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 239000006185 dispersion Substances 0.000 description 1
- 238000002296 dynamic light scattering Methods 0.000 description 1
- 239000000839 emulsion Substances 0.000 description 1
- 238000012395 formulation development Methods 0.000 description 1
- 239000000499 gel Substances 0.000 description 1
- 238000011835 investigation Methods 0.000 description 1
- 235000015110 jellies Nutrition 0.000 description 1
- 239000008274 jelly Substances 0.000 description 1
- 239000007788 liquid Substances 0.000 description 1
- 238000012423 maintenance Methods 0.000 description 1
- QSHDDOUJBYECFT-UHFFFAOYSA-N mercury Chemical compound [Hg] QSHDDOUJBYECFT-UHFFFAOYSA-N 0.000 description 1
- 229910052753 mercury Inorganic materials 0.000 description 1
- 238000010603 microCT Methods 0.000 description 1
- 238000007431 microscopic evaluation Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000000877 morphologic effect Effects 0.000 description 1
- 239000002547 new drug Substances 0.000 description 1
- 238000011192 particle characterization Methods 0.000 description 1
- 230000008560 physiological behavior Effects 0.000 description 1
- 238000012805 post-processing Methods 0.000 description 1
- 239000000843 powder Substances 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 238000004626 scanning electron microscopy Methods 0.000 description 1
- 238000004062 sedimentation Methods 0.000 description 1
- 238000007873 sieving Methods 0.000 description 1
- 238000001179 sorption measurement Methods 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 239000000725 suspension Substances 0.000 description 1
- 238000004627 transmission electron microscopy Methods 0.000 description 1
- 238000012800 visualization Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/69—Microscopic objects, e.g. biological cells or cellular parts
- G06V20/693—Acquisition
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B21/00—Microscopes
- G02B21/36—Microscopes arranged for photographic purposes or projection purposes or digital imaging or video purposes including associated control and data processing arrangements
- G02B21/365—Control or image processing arrangements for digital or video microscopes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/69—Microscopic objects, e.g. biological cells or cellular parts
- G06V20/698—Matching; Classification
Definitions
- This invention relates generally to the field of image processing and particularly to applications thereof for qualitative and quantitative analyses.
- An isolated embodiment of the present invention is disclosed in this paper, which relates specifically to a method, and its implementing system, whereby detection, classification and identification of objects of interest (namely, particulates) can be conveniently and rapidly undertaken, if any present and seen, in one or more photographic images of a sample being analyzed.
- Image processing generally refers to digitization of optical images, and performing operation(s) on the so-converted data to augment and/or extract further meaningful information, preferably in an automated manner.
- Signal dispensation of source data, approach for processing said input source data and interpretation of post-processing output are major areas of interdisciplinary research in field of the present invention wherein image visualization, restoration, retrieval, measurement and recognition are prime loci of progressive investigation.
- Particle analysis and particle characterization are major areas of research in new drug or formulation development in pharmaceutical industry. A proper analysis of particle size and shape reduces development time to a great extent. However, most of the current microscopic analysis is done manually which requires more time besides being prone to subjective interpretation and requires an expert to take the decision.
- microphotographic images in above parlance, is found to be employed variably in state- of-art technologies for study of microscopic particles wherein identifying indicia among their physical, chemical, compositional, morphological attributes and/ or physiological behaviors are utilized for qualitative and/ or quantitative determinations including identification and size distribution of the particles under study.
- implements are presently limited to non-visual light microscopy applications such as X-ray microtomography (pCT), transmission electron microscopy (TEM), scanning electron microscopy (SEM) and the like. Therefore, it would be advantageous to have some means for availing advantages of image processing technology for visual light I optical microscopy, particularly particle analysis applications.
- the art therefore requires a particle identification and classification technology that is capable of plug- and-play integration in existing optical microscopy application environments with minimal bias on capital, integration and operative expenses and at the same time, being of a nature that allows accurate and precise implementation by any person even ordinarily skilled in the art.
- Ability to succinctly discern despite strong variability among objects of interest, low contrast, and/or high incidence of agglomerates and background noise are additional characters desirable in said particle identification and classification technology presently lacking in state-of-art.
- the method so provided is fully automated via fast and optimized computational logic with low processing time, low demands on processor resources, and effective use of available computer memory stores. It is another objective further to the aforesaid objective(s) that the method so provided is error-free and lends itself to accurate implementation even at hands of a user of average skill in the art.
- FIG. 1 is a flowchart describing general logic for implementation of the present invention substantially according to the disclosures hereof.
- FIG. 2 is a flowchart describing logic for the AutoScanner feature included in the logic presented at FIG. 1 .
- the present invention propounds a fast and resource-optimized computer-implemented automated methodology for automatic scanning and focusing of uneven surfaces for identification and classification of particulates using a microscope having a motorized stage which is fitted with an imaging system such as a camera.
- the disclosures herein are directed towards establishment of a method, and its implementing system, whereby detection, classification and identification of objects of interest (namely, particulates) can be conveniently and rapidly undertaken, if any present and seen, in one or more photographic images of a sample being analyzed.
- images referred are ones obtained from a microscope having a motorized stage which is fitted with an imaging system such as a camera.
- a sample to be analyzed is processed using standard microscopy sample preparation and taken on stage of microscope for microphotography.
- resolution of the present invention is correlated with optics of the microscope, and not the camera or computing system involved.
- Camera fitments for optical microscopes are inexpensive and commonly available. Assemblage and operations of these components requires no particular skill or collateral knowledge.
- the present invention is free of constraints entailing otherwise from capital, operation and maintenance costs besides negating the requirement of trained skilled operators for implementation of the present invention.
- step (01) the user initializes I starts the application of the present invention (named “ipvPAuto” and referred so throughout this document).
- step (02) the user is prompted (via suitable user interface) to create I select method to set particle range, magnification selection etc.
- step (03) is caused to be executed, wherein the analysis area is scanned (by a routine named “AutoScanner”), and images captured are saved with names I identification of the scan position.
- a Scanning info text file is created via step (04) by AutoScanner specifying therein number of rows, number of columns, and total fields. Thereafter, a it is determined via query at step (05), as to whether the image is captured by AutoScanner and ready in a shared folder. If determination is negative, the execution is paused at step (06) till this is achieved, else, the logic is programmed to terminate (after suitable threshold I benchmark) via step (07). If determination is positive, execution logic is directed, via step (08) to read the image and image position (row and column in scan area from name).
- step (09) the image data is preprocessed via step (09) to smoothen said image by removing noise. Thereafter, contours in said image are identified, via step (10) and contours of same gray value variation (gradient) are mapped out.
- step (11) for identification of objects via a sub-process including forming contour groups, and finding best contour from group from user criteria selection (Sharpness, Bounding box, Circularity, and Perimeter).
- pre-arranged I pre-programmed filters are applied at step (13) to remove artifacts.
- Filters applied are selected among group including a) size filter - Filter Particles not in defined range; b) Sharpness Filter - Filter blur particles (less than defined sharpness); and c) Agglomeration Filter - Filter non isolated particles identified on shape features.
- the logic hereof is programmed to determine, at step (14), as to whether the identified object I particle on left or top of the image boundary. If this determination is positive, sub-process is triggered at step (15), whereby boundary particle identification is achieved by steps of a) searching overlapping particle in left I top image; b) Cropping image parts to create new cropped image; and c) Finding particle in the cropped image. If this determination is negative, another sub-process is triggered at step (16), whereby particle classification is achieved via steps of a) Adding particles in particle list; b) sorting particles in defined particle range; and c) Adding image in scanned area list.
- step (17) it is sought to be determined whether or not it is the last image of total fields. If this determination is positive, computation of result statistics is triggered at step (18). Else, if this determination is negative, the logic is programmed to lead via step (19) to terminate (after suitable threshold I benchmark) via step (07). Else, the logic is deemed to execute in intended manner, and culminates via step (20).
- step (21) a scanning map is created.
- scanning map from input magnification is created, and then scanning information (for example, rows, cols, fields) is saved in file in shared folder with ipvPAuto at step (22).
- microscope stage is automatically moved in step (23) to the center position of the scanning area.
- the AutoScanner is initialized at step (24) by a sub-process comprising a) Moving the stage automatically to find optimum focus position; b) Computing brightness, sharpness and focusing range; and c) Computing texture value.
- step (25) to assess whether the texture value measures is high (than a known texture value of the filter paper used). If this determination is positive, a message is outputted to the user at step (26), that filter paper (sample is presumed to be held on this filter paper) is absent and scanning should be stopped. If determination here is positive however, stage is automatically moved, at step (27) to the start of the scanning.
- scanning is performed at step (28) by a) Moving stage in x direction by one step; b) When end field reached of the row, moving one step down; c) Moving stage in opposite x direction; d) When end field reached of the row, moving one step down; e) Continuing scan until end of field is reached.
- step (29) auto-focusing for other than boundary fields is undertaken at step (29), by sub-process including a) Auto focusing on computing the focus direction; b) Capturing and saving image in shared folder; c) If totally de-focused, outputting a message for focusing and wait to restart scanning. Thereafter, the logic is programmed to seek at step (30), whether or not the current field is imaged the last field. If yes, scanning is instructed to stop via step (31), else, the system loops to scanning as per step (28) described above.
- the present invention is able to process microphotographic images of samples including dry powder, liquid, gel, jelly, aerosols, emulsions, suspension, dispersion and so on and in practice, has been observed to provide results in few seconds.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Chemical & Material Sciences (AREA)
- Analytical Chemistry (AREA)
- Optics & Photonics (AREA)
- Image Analysis (AREA)
Abstract
Disclosed herein is a method, and its implementing system, whereby detection, classification and identification of objects of interest (namely, particulates) can be conveniently and rapidly undertaken, if any present and seen, in one or more photographic images of a sample being analyzed.
Description
Method and system for automatic scanning and focusing of uneven surfaces for identification and classification of particulates”
Cross references to related applications: This non-provisional patent application claims the benefit of US provisional application no. 63/251640 filed on 03 October 2021 , the contents of which are incorporated herein in their entirety by reference.
Statement Regarding Federally Sponsored Research or Development: None applicable
Reference to Sequence Listing, a Table, or a Computer Program Listing Compact Disc Appendix: None
Field of the invention
This invention relates generally to the field of image processing and particularly to applications thereof for qualitative and quantitative analyses. An isolated embodiment of the present invention is disclosed in this paper, which relates specifically to a method, and its implementing system, whereby detection, classification and identification of objects of interest (namely, particulates) can be conveniently and rapidly undertaken, if any present and seen, in one or more photographic images of a sample being analyzed.
Background of the invention and description of related art
Image processing generally refers to digitization of optical images, and performing operation(s) on the so-converted data to augment and/or extract further meaningful information, preferably in an automated manner. Signal dispensation of source data, approach for processing said input source data and interpretation of post-processing output are major areas of interdisciplinary research in field of the present invention wherein image visualization, restoration, retrieval, measurement and recognition are prime loci of progressive investigation.
Particle analysis and particle characterization are major areas of research in new drug or formulation development in pharmaceutical industry. A proper analysis of particle size and shape reduces development time to a great extent. However, most of the current microscopic analysis is done manually which requires more time besides being prone to subjective interpretation and requires an expert to take the decision.
Processing of microphotographic images, in above parlance, is found to be employed variably in state- of-art technologies for study of microscopic particles wherein identifying indicia among their physical, chemical, compositional, morphological attributes and/ or physiological behaviors are utilized for qualitative and/ or quantitative determinations including identification and size distribution of the
particles under study. However, such implements are presently limited to non-visual light microscopy applications such as X-ray microtomography (pCT), transmission electron microscopy (TEM), scanning electron microscopy (SEM) and the like. Therefore, it would be advantageous to have some means for availing advantages of image processing technology for visual light I optical microscopy, particularly particle analysis applications.
Conventionally, detection and classification of particles has been practiced via sieving, sedimentation, dynamic light scattering, electrozone sensing, optical particle counting, XRD line profile analysis, adsorption techniques and mercury intrusion or further indirect methods such as surface area measurements. However, resolution of these techniques leave a lot to be desired, besides relying on availability of expensive equipment and collateral prior expertise of skilled operators for arriving at the determination intended. Such analysis, as will be obvious to the reader, tends to be less reproducible due to unavoidable personal biases and therefore inaccurate for faultless determinations. There is hence a need for some way that makes possible the integration of image analytics for particle classification in optical microscopy applications.
The art therefore requires a particle identification and classification technology that is capable of plug- and-play integration in existing optical microscopy application environments with minimal bias on capital, integration and operative expenses and at the same time, being of a nature that allows accurate and precise implementation by any person even ordinarily skilled in the art. Ability to succinctly discern despite strong variability among objects of interest, low contrast, and/or high incidence of agglomerates and background noise are additional characters desirable in said particle identification and classification technology presently lacking in state-of-art.
A better understanding of the objects, advantages, features, properties and relationships of the present invention will be obtained from the underlying specification, which sets forth the best mode contemplated by the inventor of carrying out the present invention.
Objectives of the present invention
The present invention is identified in addressing at least all major deficiencies of art discussed in the foregoing section by effectively addressing the objectives stated under, of which:
It is a primary objective to provide an effective method for automatic scanning and focusing of uneven surfaces for identification and classification of particulates.
It is another objective further to the aforesaid objective(s) that the method so provided is fully automated via fast and optimized computational logic with low processing time, low demands on processor resources, and effective use of available computer memory stores.
It is another objective further to the aforesaid objective(s) that the method so provided is error-free and lends itself to accurate implementation even at hands of a user of average skill in the art.
It is another objective further to the aforesaid objective(s) that implementation of the method so provided does not involve any complicated or overtly expensive hardware.
It is another objective further to the aforesaid objective(s) that implementation of the method is possible via a remote server, in a software-as-a-service (SaaS) model.
The manner in which the above objectives are achieved, together with other objects and advantages which will become subsequently apparent, reside in the detailed description set forth below in reference to the accompanying drawings and furthermore specifically outlined in the independent claims. Other advantageous embodiments of the invention are specified in the dependent claims.
Brief description of drawings
The present invention is explained herein under with reference to the following drawings, in which,
FIG. 1 is a flowchart describing general logic for implementation of the present invention substantially according to the disclosures hereof.
FIG. 2 is a flowchart describing logic for the AutoScanner feature included in the logic presented at FIG. 1 .
The above drawings are illustrative of particular examples of the present invention but are not intended to limit the scope thereof. Though numbering has been introduced to demarcate reference to specific components in relation to such references being made in different sections of this specification, all components are not shown or numbered in each drawing to avoid obscuring the invention proposed.
Attention of the reader is now requested to the detailed description to follow which narrates a preferred embodiment of the present invention and such other ways in which principles of the invention may be employed without parting from the essence of the invention claimed herein.
Summary of the present invention
The present invention propounds a fast and resource-optimized computer-implemented automated methodology for automatic scanning and focusing of uneven surfaces for identification and classification of particulates using a microscope having a motorized stage which is fitted with an imaging system such as a camera.
Detailed description
Principally, general purpose of the present invention is to assess disabilities and shortcomings inherent to known systems comprising state of the art and develop new systems incorporating all available advantages of known art and none of its disadvantages. Accordingly, the disclosures herein are directed towards establishment of a method, and its implementing system, whereby detection, classification and identification of objects of interest (namely, particulates) can be conveniently and rapidly undertaken, if any present and seen, in one or more photographic images of a sample being analyzed.
In the embodiment recited herein, the reader shall presume that images referred are ones obtained from a microscope having a motorized stage which is fitted with an imaging system such as a camera. For this, a sample to be analyzed is processed using standard microscopy sample preparation and taken on stage of microscope for microphotography. As will be realised further, resolution of the present invention is correlated with optics of the microscope, and not the camera or computing system involved. Camera fitments for optical microscopes are inexpensive and commonly available. Assemblage and operations of these components requires no particular skill or collateral knowledge. Hence, the present invention is free of constraints entailing otherwise from capital, operation and maintenance costs besides negating the requirement of trained skilled operators for implementation of the present invention.
Reference is now made to the accompanying FIG. 1 , which is a flowchart describing general logic for implementation of the present invention. As seen here, execution of the present invention begins at step (01) where the user initializes I starts the application of the present invention (named “ipvPAuto” and referred so throughout this document). This triggers step (02) in which the user is prompted (via suitable user interface) to create I select method to set particle range, magnification selection etc. Once this is done, step (03) is caused to be executed, wherein the analysis area is scanned (by a routine named “AutoScanner”), and images captured are saved with names I identification of the scan position.
With continued reference to the accompanying FIG. 1 , it can be seen that once image data is available as per foregoing narration, a Scanning info text file is created via step (04) by AutoScanner specifying therein number of rows, number of columns, and total fields. Thereafter, a it is determined via query at step (05), as to whether the image is captured by AutoScanner and ready in a shared folder. If determination is negative, the execution is paused at step (06) till this is achieved, else, the logic is programmed to terminate (after suitable threshold I benchmark) via step (07). If determination is positive, execution logic is directed, via step (08) to read the image and image position (row and column in scan area from name).
With continued reference to the accompanying FIG. 1 , it can be seen that once image is read, the image data is preprocessed via step (09) to smoothen said image by removing noise. Thereafter, contours in said image are identified, via step (10) and contours of same gray value variation (gradient) are mapped out. This allows, in step (11) for identification of objects via a sub-process including forming contour
groups, and finding best contour from group from user criteria selection (Sharpness, Bounding box, Circularity, and Perimeter).
With continued reference to the accompanying FIG. 1 , it can be seen that once objects are identified, feature computation is undertaken via step (12) on basis of size, shape, color, and texture. Thereafter, pre-arranged I pre-programmed filters are applied at step (13) to remove artifacts. Filters applied are selected among group including a) size filter - Filter Particles not in defined range; b) Sharpness Filter - Filter blur particles (less than defined sharpness); and c) Agglomeration Filter - Filter non isolated particles identified on shape features.
With continued reference to the accompanying FIG. 1 , it can be seen that after the determinations described above, the logic hereof is programmed to determine, at step (14), as to whether the identified object I particle on left or top of the image boundary. If this determination is positive, sub-process is triggered at step (15), whereby boundary particle identification is achieved by steps of a) searching overlapping particle in left I top image; b) Cropping image parts to create new cropped image; and c) Finding particle in the cropped image. If this determination is negative, another sub-process is triggered at step (16), whereby particle classification is achieved via steps of a) Adding particles in particle list; b) sorting particles in defined particle range; and c) Adding image in scanned area list.
With continued reference to the accompanying FIG. 1 , the reader shall appreciate another determination being posed at step (17), whereby it is sought to be determined whether or not it is the last image of total fields. If this determination is positive, computation of result statistics is triggered at step (18). Else, if this determination is negative, the logic is programmed to lead via step (19) to terminate (after suitable threshold I benchmark) via step (07). Else, the logic is deemed to execute in intended manner, and culminates via step (20).
According to a related aspect of the present invention explained with reference to the accompanying FIG. 2, execution of the AutoScanner feature introduced above can be seen in that it initiates, via step (21) in which a scanning map is created. For this, scanning map from input magnification is created, and then scanning information (for example, rows, cols, fields) is saved in file in shared folder with ipvPAuto at step (22). Thereafter, microscope stage is automatically moved in step (23) to the center position of the scanning area. Once this is arrived at, the AutoScanner is initialized at step (24) by a sub-process comprising a) Moving the stage automatically to find optimum focus position; b) Computing brightness, sharpness and focusing range; and c) Computing texture value.
With continued reference to the accompanying FIG. 2, it shall be seen that once texture value is computed, it is subjected to a determination via step (25) to assess whether the texture value measures is high (than a known texture value of the filter paper used). If this determination is positive, a message is outputted to the user at step (26), that filter paper (sample is presumed to be held on this filter paper) is absent and scanning should be stopped. If determination here is positive however, stage is
automatically moved, at step (27) to the start of the scanning. Thereafter, scanning is performed at step (28) by a) Moving stage in x direction by one step; b) When end field reached of the row, moving one step down; c) Moving stage in opposite x direction; d) When end field reached of the row, moving one step down; e) Continuing scan until end of field is reached.
With continued reference to the accompanying FIG. 2, auto-focusing for other than boundary fields is undertaken at step (29), by sub-process including a) Auto focusing on computing the focus direction; b) Capturing and saving image in shared folder; c) If totally de-focused, outputting a message for focusing and wait to restart scanning. Thereafter, the logic is programmed to seek at step (30), whether or not the current field is imaged the last field. If yes, scanning is instructed to stop via step (31), else, the system loops to scanning as per step (28) described above.
Via implementation logic disclosed, it shall be appreciated how detection, classification and identification of objects of interest (namely, particulates) is brought about by the present invention. As will be generally realized, applicability and/ or performance of the present invention is not designed to be dependent on any particular sample composition and/ or preparation techniques. Accordingly, the present invention is able to process microphotographic images of samples including dry powder, liquid, gel, jelly, aerosols, emulsions, suspension, dispersion and so on and in practice, has been observed to provide results in few seconds.
As will be realized further, the present invention is capable of various other embodiments and that its several components and related details are capable of various alterations, all without departing from the basic concept of the present invention. Accordingly, the foregoing description will be regarded as illustrative in nature and not as restrictive in any form whatsoever. Modifications and variations of the system and apparatus described herein will be obvious to those skilled in the art. Such modifications and variations are intended to come within ambit of the present invention, which is limited only by the appended claims.
Claims
7
We claim,
1 ) A method for automatic scanning and focusing of uneven surfaces for identification and classification of particulates, the method comprising- a) Constituting an application environment by communicatively associating an optical microscope having a motorized stage to a computer, wherein-
■ the method for automatic scanning and focusing of uneven surfaces for identification and classification of particulates is provisioned for execution, as an executable software, on said computer; and
■ the optical microscope is outfitted with a digital camera for capturing images from the field of view of said microscope and relaying said captured images in real time to said computer for processing by the executable software provisioned on said computer. b) Defining, via a computer user interface of the executable software, a set of scanning parameters being opted at instance of the user among particle size, analysis area, scan position, and magnification; c) In accordance with the defined set of scanning parameters, causing at least one image to be captured from the microscope corresponding to the analysis area selected, therein saving the at least one image to a ready shared folder in memory of the computer with its filename corresponding to the scan position selected; d) In accordance with logic of the executable software-
■ creating a text file specifying therein the number of rows, number of columns, and total fields;
■ Via an interactive user interface, allowing the user to set a set of predefined parameters for object determination, said parameters being sharpness, bounding box, circularity, and perimeter;
■ If saved to the ready shared folder, reading the at least one image and its filename, and preprocessing it by smoothening said at least one image for removal of noise;
■ Identifying contours in said at least one image, therein selecting contours of same gray value variation to form contour groups and determining, among said contour groups, the best contours on basis of the predefined parameters for object determination;
■ Once objects are identified, computing feature data corresponding to said objects on basis of their size, shape, color, and texture;
■ Applying at least one filter on basis of size, sharpness, and agglomeration to the feature data generated to remove artifacts and result in filtered object data;
■ Determining whether the filtered object data corresponds to objects present on the left and top boundaries of the image under processing, and based on this determination, causing the execution of a suitable sub-process for resulting in
8 either between boundary particle identification and particle classification respectively;
■ Determining whether the image being processed is the last image of total fields, and based on this determination, causing the execution of a suitable sub-process for either between termination with computation of result statistics and termination of the execution of the executable software upon reaching threshold of the user- defined parameter values respectively. e) In accordance with flow of execution at step d), causing execution of the AutoScanner sub-process included in the executable software, said sub-process consisting of-
■ Creating a scanning map from input magnification to generate scanning information consisting of the rows, columns and fields is saved in the ready shared folder;
■ Causing the microscope stage to move automatically, to the center position of the scanning area;
■ Once center position of the scanning area is arrived at, moving the microscope stage to find an optimum focus position and therein computing the brightness, sharpness, focusing range and texture value;
■ Determining whether the computed texture value measures is higher than the known texture value of the filter paper used to hold the sample and based on this determination, causing the execution of a suitable sub-process for resulting in stoppage or commencement of scanning respectively. f) Auto-focusing for other than boundary fields, therein determining whether the current field imaged is the last field and based on this determination, causing the execution of a suitable sub-process for resulting in stoppage or continued scanning by looping the aforesaid steps respectively. ) The method for automatic scanning and focusing of uneven surfaces for identification and classification of particulates according to claim 1 , wherein, if the filtered object data corresponds to objects present on the left and top boundaries of the image under processing, the suitable sub-process caused to be executed is for boundary particle identification, said sub-process consisting of- a) Searching overlapping particle in the left and top image; b) Cropping image parts to create new cropped image; and c) Finding boundary particle in the cropped image. ) The method for automatic scanning and focusing of uneven surfaces for identification and classification of particulates according to claim 1 , wherein, the filtered object data does not corresponds to objects present on the left and top boundaries of the image under processing, the suitable sub-process caused to be executed is for particle classification, said sub-process consisting of-
9 a) Adding particles in particle list; b) Sorting particles in the user-defined particle range; and c) Adding image in scanned area list. ) The method for automatic scanning and focusing of uneven surfaces for identification and classification of particulates according to claim 1 , wherein, if the image being processed is the last image of total fields, the suitable sub-process caused to be executed is for computation of result statistics. ) The method for automatic scanning and focusing of uneven surfaces for identification and classification of particulates according to claim 1 , wherein, if the image being processed is not the last image of total fields, the suitable sub-process caused to be executed is of termination of execution of the executable software. ) The method for automatic scanning and focusing of uneven surfaces for identification and classification of particulates according to claim 1 , wherein, if the computed texture value is higher than the prior known texture value, the suitable sub-process caused to be executed is of outputting a message to the user, via user interface of the executable software, that the filter paper presumed to hold the sample being processed is absent and scanning should be stopped. ) The method for automatic scanning and focusing of uneven surfaces for identification and classification of particulates according to claim 1 , wherein, if the computed texture value is not higher than the prior known texture value, the suitable sub-process caused to be executed is of scanning using with conventional means of automation of the stage of the microscope, said sub-process comprising- a) Moving stage in one direction by one step; b) When end field reached of the row in said direction, moving the stage one step down; c) Moving stage in a direction opposite to that opted for in step a); d) When end field reached of the row, moving one step down; and e) Looping steps a) to d) until end of field is reached. ) The method for automatic scanning and focusing of uneven surfaces for identification and classification of particulates according to claim 1 , wherein auto-focusing for other than boundary fields is undertaken by a sub-process comprising- a) Auto focusing on computing the focus direction; b) Capturing and saving image in shared folder; and c) If totally de-focused, outputting a message for focusing and wait to restart scanning. ) The method for automatic scanning and focusing of uneven surfaces for identification and classification of particulates according to claim 1 , wherein if the current field imaged is the last
10 field, the sub-process caused to be executed is of cessation of execution of the executable software. ) The method for automatic scanning and focusing of uneven surfaces for identification and classification of particulates according to claim 1 , wherein if the current field imaged is not the last field, the sub-process caused to be executed is of continuing execution of the executable software until the last field is reached. ) The method for automatic scanning and focusing of uneven surfaces for identification and classification of particulates according to claim 1 , wherein the executable software is provisioned for execution on the computer by either between a standalone installation and online access from a cloud server in a software-as-a-service model.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US202163251640P | 2021-10-03 | 2021-10-03 | |
US63/251,640 | 2021-10-03 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2023053103A1 true WO2023053103A1 (en) | 2023-04-06 |
Family
ID=85780515
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IB2022/059430 WO2023053103A1 (en) | 2021-10-03 | 2022-10-03 | Method and system for automatic scanning and focusing of uneven surfaces for identification and classification of particulates |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO2023053103A1 (en) |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040208352A1 (en) * | 2003-04-21 | 2004-10-21 | Damian Neuberger | Determination of particle size by image analysis |
US20040254738A1 (en) * | 2003-06-12 | 2004-12-16 | Cytyc Corporation | Method and system for organizing multiple objects of interest in field of interest |
-
2022
- 2022-10-03 WO PCT/IB2022/059430 patent/WO2023053103A1/en active Application Filing
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040208352A1 (en) * | 2003-04-21 | 2004-10-21 | Damian Neuberger | Determination of particle size by image analysis |
US20040254738A1 (en) * | 2003-06-12 | 2004-12-16 | Cytyc Corporation | Method and system for organizing multiple objects of interest in field of interest |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10783348B2 (en) | Method and system for detection and classification of particles based on processing of microphotographic images | |
CN108288027B (en) | Image quality detection method, device and equipment | |
US6529612B1 (en) | Method for acquiring, storing and analyzing crystal images | |
US8600143B1 (en) | Method and system for hierarchical tissue analysis and classification | |
JP3822242B2 (en) | Method and apparatus for evaluating slide and sample preparation quality | |
US11062168B2 (en) | Systems and methods of unmixing images with varying acquisition properties | |
Viles et al. | Measurement of marine picoplankton cell size by using a cooled, charge-coupled device camera with image-analyzed fluorescence microscopy | |
Bollmann et al. | Automated particle analysis: calcareous microfossils | |
WO2017150194A1 (en) | Image processing device, image processing method, and program | |
JP7418639B2 (en) | Particle analysis data generation method, particle analysis data generation program, and particle analysis data generation device | |
JPH10318904A (en) | Apparatus for analyzing particle image and recording medium recording analysis program therefor | |
US20120249770A1 (en) | Method for automatically focusing a microscope on a predetermined object and microscope for automatic focusing | |
JP2023542619A (en) | Computer-implemented method for quality control of digital images of specimens | |
US20240112362A1 (en) | Method and system for automatic scanning and focusing of uneven surfaces for identification and classification of particulates | |
WO2023053103A1 (en) | Method and system for automatic scanning and focusing of uneven surfaces for identification and classification of particulates | |
US20240203141A1 (en) | Photomicrographic image-processing method for automatic scanning, detection and classification of particles | |
Allen et al. | Machine vision for automated optical recognition and classification of pollen grains or other singulated microscopic objects | |
WO2023112002A1 (en) | Photomicrographic image-processing method for automatic scanning, detection and classification of particles | |
US20230196539A1 (en) | Artificial intelligence based method for detection and analysis of image quality and particles viewed through a microscope | |
WO2023053104A1 (en) | Method and system for tracking and analysis of particles due to thermal variations | |
JP4344862B2 (en) | Method and apparatus for automatic detection of observation object | |
ATICI et al. | Determination of blood group by image processing using digital images | |
Vega-Alvarado et al. | Images analysis method for the detection of Chagas parasite in blood image | |
JPH1090163A (en) | Particle analyzer | |
WO2002097409A1 (en) | Method for the automated recognition, spectroscopic analysis and identification of particles |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 22875313 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 22875313 Country of ref document: EP Kind code of ref document: A1 |