WO2012119191A1 - Method and software for analysing microbial growth - Google Patents
Method and software for analysing microbial growth Download PDFInfo
- Publication number
- WO2012119191A1 WO2012119191A1 PCT/AU2012/000211 AU2012000211W WO2012119191A1 WO 2012119191 A1 WO2012119191 A1 WO 2012119191A1 AU 2012000211 W AU2012000211 W AU 2012000211W WO 2012119191 A1 WO2012119191 A1 WO 2012119191A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- classifier
- pixel
- pixels
- image
- image data
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/42—Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
- G06V10/422—Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation for representing the structure of the pattern or shape of an object therefor
- G06V10/426—Graphical representations
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/69—Microscopic objects, e.g. biological cells or cellular parts
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/69—Microscopic objects, e.g. biological cells or cellular parts
- G06V20/698—Matching; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30024—Cell structures in vitro; Tissue sections in vitro
Definitions
- the present invention relates to a method and software for analysing microbial growth on a solid culture medium, particularly microbial growth in the form of isolated bacterial colonies grown following the inoculation and incubation of the solid culture medium with a microbiological sample. It is envisaged that the method and software of the present invention will find use in a microbiological laboratory.
- microbiological samples are first dispensed upon the surface of a solid culture medium followed by the spreading of the microbiological sample across the surface of the medium (called "streaking"). Typically, multiple streaks are made of increasing dilution of the inoculum across the solid culture medium.
- the streaks of increasing dilution tend to provide, generally towards the tail of the streaks, a number of single cells that allow for the growth of isolated microbiological colonies after incubation.
- These isolated colonies may then be analysed for various physical features e.g. colony morphology, and may undergo staining and other procedures which may be necessary for determining, for example, the genus, the species and/or the strain of the previously unidentified organism in the microbiological sample.
- this analysis has been carried out visually in a microbiological laboratory by skilled technologists, resulting in the technologist making a microbiological assessment.
- This assessment may be based on the detection of the presence or absence of bacterial colonies, the detection of colour(s) of each colony type, the mapping of colour distribution to determine the presence of variations in colour which could be attributed to fermentation or haemolysis, the differentiation between confluent and isolated colony growth, the measurement of colony texture or viscosity, and the determination of two-dimensional and three-dimensional shape, and/or enumeration of the different types of colonies.
- the solid culture medium is progressed to the next step of the laboratory workflow and becomes the subject of further confirmatory identification and antibiotic susceptibility testing, in line with current regulatory requirements.
- the present invention provides a method for analysing microbial growth on a solid culture medium, the method including:
- a microbiological sample hereinafter often referred to simply as a "sample”
- a sample can be dispensed either upon the surface of the medium or within the medium so that microbial growth can be grown following the inoculation and incubation of the medium with the sample. That is, the microbial growth on the medium results from the inoculation and incubation of the sample, such as a urine sample, an enteric sample, a blood sample, a lymph sample, a tissue sample, a water sample, a food sample or other relevant sample, on the medium.
- the sample such as a urine sample, an enteric sample, a blood sample, a lymph sample, a tissue sample, a water sample, a food sample or other relevant sample, on the medium.
- the medium will usually be, for example, agar and will usually be contained in a container such as a plate, and, in a more specific example, a Petri dish, which may have a lid.
- a container such as a plate, and, in a more specific example, a Petri dish, which may have a lid.
- the combination of the medium and the plate is hereinafter referred to throughout the specification as a "culture plate” which might sometimes be referred to in the art as an "agar plate".
- the present invention uses the classifier for classifying pixels in an image of the medium with or without microbial growth in order to provide the microbiological assessment. Little or no human operator intervention is required. This may enable a saving of the time to sort culture plates and may allow more effective utilisation of skilled human resources, which can be diverted to other tasks.
- the method includes classifying the microbial growth on the medium based on the results of the pixel classifications.
- the method further includes enumerating one or more colonies of the microbial growth based on the results of the pixel classifications.
- the method can then analyse the results of the pixel classifications using, for example, a connected components algorithm to determine a series of regions indicative of predicted colonies of the microbial growth.
- the method further includes the step of enumerating the colonies of the microbial growth from the series of regions based on a size and/or shape of each of said regions.
- the microbial growth may include, for example, one or more bacterial growths, fungal growths, viral plaques or protist growths and the growth may take the form of a colony, mycelium, hypha, plaque or other visible microbial structure.
- each microbial growth may be growth that originates from a single microbe (such as where a sample is applied to a medium in a diluted manner such that individual microbes are separated).
- the medium may include any medium which supports the growth of a microbe.
- a medium may contain one or more microbial nutrients including, for example, a carbon source, a nitrogen source, essential elements and/or essential vitamins.
- the medium will also typically contain a gelling agent including, for example, gelatine, agar, gellan gum, agarose or agargel.
- the medium may include a selective culture medium, which is generally known in the art to include media with restricted nutrient compositions, media containing an antibiotic, or the like, that allow for the growth of some microorganisms while preventing growth of other microorganisms.
- the sample for analysis is disposed on more than one culture plate or a split culture plate having more than one segment.
- the analysis of microbial growth is performed by taking images across the different culture plates, or segments, and the results of the pixel classifications are analysed as above so that a microbiological assessment can be made based on a comparison of microbial growth on the different culture plates or segments.
- microbiological samples may not result in microbial growth on a particular medium.
- the microbiological sample may not include a microbe capable of growth on a selective medium (e.g. a medium containing a specific antibiotic or nutrient composition that restricts growth of some microbes) or the microbiological sample may be sterile (e.g. not include any microbes).
- the image data obtained for the medium may not include a microbial growth on the medium (e.g. a negative result).
- the microbiological assessment may be used to identify whether the culture plate falls into one of up to 4 categories - 1. waste (e.g. the culture plate showed no pathogens), 2. reincubation required, 3.
- the assessment includes the detection of microbial growth colonies in the medium and can include, for example, the classification of these colonies, such as bacterial colonies, and the mapping of medium colour distribution to determine the presence of variations in colour which could be attributed to fermentation or haemolysis, the differentiation between confluent and isolated colony growth, the measurement of colony texture or viscosity, and the determination of two- dimensional and three-dimensional shape, and/or enumeration of the different types of colony on the medium.
- the image data for the medium and any microbial growth thereon may include one or more images taken using an image capture device, such as a high resolution colour digital camera, a laser rangefinder or any other appropriate device.
- the image data is obtained from an existing image which could have, for example, been taken for another purpose.
- the image data is obtained from an image taken using an image capture device dedicated to the present method.
- the images may be processed to produce the image data, for example a plurality of images including the same microbial growth may be averaged, demosaiced and colour calibrated.
- the image data may include a series of such processed images, for example taken using different lighting configurations. Further details of image capture and processing will be given below.
- the classifier may be a boosted classifier.
- the classifier may be a decision tree, random forest, or may use Linear Discriminant Analysis (LDA) or any other technique to classify the pixels.
- LDA Linear Discriminant Analysis
- a boosted classifier combines the output of a set of "weak" classifiers (with low, but better than random, classification accuracy) to produce a single "strong" classifier with high accuracy.
- the result of each weak classifier is preferably weighted according to the confidence in the correctness of the weak classifier.
- An example of an appropriate boosting algorithm for creating the boosted classifier is Discrete AdaBoost, which will be described in more detail below.
- Discrete AdaBoost is a variant of AdaBoost, described in Yoav Freund and Robert E.
- Discrete AdaBoost adaptively improves the performance of the classifier by giving greater weight to examples misclassified by a weak classifier when training the next weak classifier.
- Other boosting algorithms may alternatively be used, such as Real AdaBoost, Gentle AdaBoost or FilterBoost.
- the variants use different methods of weighting and training the weak classifiers.
- the classifier used to classify the pixels is a boosted decision tree classifier.
- the weak classifiers used are decision trees.
- Decision trees are binary trees where each node contains a rule, each branch represents a result of the rule, and hence a decision, and each leaf representing a classification.
- the tree is traversed from the root node to classify a single feature vector.
- a common arrangement is that if a test passes the left branch of that node is traversed, otherwise the right branch is traversed instead.
- the boosted decision tree classifier may be a multi-class classifier. For example, the AdaBoost.
- MH Multi-Label Hamming
- the AdaBoost. MH algorithm is described in Jerome Friedman, Trevor Hastie, and Robert Tibshirani. "Additive logistic regression: a statistical view of boosting. " Annals of Statistics, 28:2000, 1998, the contents of which are herein incorporated by reference.
- classification may be performed as a two stage process.
- the method may include using a first classifier to initially classify each pixel as one of a first plurality of classes, and subsequently using a second classifier to classify each pixel in one or more of the first plurality of classes as one of a second plurality of classes.
- the result is a chain of dependent classifiers, where the output of the first classifier is used to either augment the feature vectors or restrict the applicable pixels when training the second classifier.
- the first classifier may be a binary cascade classifier to initially classify each pixel as background or non-background
- the second classifier may be a multi-class boosted decision tree classifier to classify each non-background pixel as one of the second plurality of classes.
- the initial coarse classification of background pixels reduces the number of pixels that need to be more accurately classified into one of the colony types.
- the initial binary classification may have a high false positive rate (background pixels classified as non- background) as misclassified pixels can be correctly classified in the next stage.
- a suitable initial binary classifier may follow the cascade classifier method described in Paul Viola and Michael Jones "Robust real-time object detection", Second International Workshop on Statistical and Computational Theories of Vision, July 13 2001, the contents of which are herein incorporated by reference.
- the classes may represent organism genus, species or strains, agar types, background, non-background.
- the classes are flexible, to allow for different agar types and organism types that may be of interest to different end-users.
- the method may further include assigning a confidence value to each pixel classification.
- the confidence value represents the probability of correct classification of the pixel.
- the method may further include applying a post processing algorithm to improve the results of the pixel classifications.
- a post processing algorithm may include morphological operations such as dilation or erosion or, alternatively, a graph cut.
- An example of a suitable graph cut algorithm is given in Yuri Boykov, Olga Veksler, and Ramin Zabih "Fast approximate energy minimization via graph cuts" IEEE Trans. Pattern Anal. Mach. Intel/., 23:1222-1239, November 2001., the contents of which are herein incorporated by reference.
- Graph cut algorithms compute an optimal partitioning of a graph into two or more sets. The application of the graph cut algorithm may lead to the reclassification of some of the pixels in the image data. Applying a graph cut algorithm may include:
- Applying a graph cut algorithm may improve the classification results for pixel classifications that have been given low confidence by the classifier.
- the graph cut algorithm may take into account the confidence value of the pixel classification for the pixel corresponding to that node and/or the confidence value of the pixel classification for neighbouring pixels. This spatially smoothes the results, reducing noise in the final classifications.
- the high confidence annotations will normally be kept, while low confidence annotations may be replaced to obtain a more homogeneous output. For example, if a low confidence pixel is surrounded by neighbouring pixels classified into the same class with high confidence, there can be more confidence in the low confidence pixel's classification. By the same token, if the neighbouring pixels are classified into a different class with high confidence, there is a strong case for changing the low confidence pixel's classification to that class.
- the graph cut algorithm may be a multi-class graph cut algorithm, which partitions the graph into three or more sets. This is a more complicated algorithm than a binary graph cut, but allows a more meaningful assessment to be provided as the pixels may be classified more specifically.
- Such an algorithm may be an alpha expansion procedure. In alpha expansion, a series of graph cuts are performed, each time segmenting between the current label for each node, and a candidate label from the set of possible labels. This procedure is repeated, iterating through each possible label, until convergence. In constructing the graph, auxiliary nodes are added between adjacent nodes with different labels, to include the cost of this labelling in the cut.
- the feature vector may be composed of values taken from feature images, which are created by applying one or more filters to the image data for the sample. Obtaining feature images is an intermediate step, and it will be appreciated that alternatively the feature vectors could be composed of values taken by applying the filters on a pixel- by-pixel basis.
- the term "filter” is to be understood to encompass any image processing or other algorithm that converts part or all of the image data into one or more values that may be input into the classifier.
- the one or more filters may be chosen from a group including: o RGB filter, which creates up to three feature images, one for the red channel, one for the green channel and one for the blue channel,
- o LAB filter which creates up to three or more feature images in the LAB colour space, where L represents lightness and A and B contain chromatic colour information.
- o Oriented Gaussian Filter which applies a 1 D Gaussian kernel in a variety of directions to the image. It is also possible to utilise first and second order derivatives of the Gaussian kernel,
- Position filter which for example creates a feature image showing the distance of each pixel from the edge of the culture plate.
- Other position filters may show the distance of each pixel from a particular label, if the filter is applied when the output of a previous classifier is available,
- o Colour quantisation filter where the number of colours in the image is reduced, usually via a histogram, to help locate areas of similar colour
- o Texture filter which creates one or more feature images measuring texture around each pixel.
- the filters may enhance the differences between colonies and the medium and the plate itself.
- Some of the filters include a feedback mechanism, whereby the output from a previous classifier can be used as an input.
- the feature vectors may be constructed using any number of filters.
- the feature vector for a pixel may include 9 values: Xi , x 2 and x 3 corresponding to the R, G and B channel values for that pixel, x 4 , x 5 and x 6 , corresponding to the L, A and B values for that pixel, x 7 corresponding to the distance of the pixel from the edge of the plate and x 8 and x 9 corresponding to image gradients in the x and y directions.
- any number of feature values may be used and feature values may be taken from different image data for the same growth (e.g. lit by a top light or a bottom light).
- applying one or more filters to the image data may include measuring texture of the image data by
- the gradient values may be extracted by convolving in x and y directions of the image data using a Sobel kernel.
- the use of a Sobel kernel smooths the image at the same time as extracting gradient information.
- Computing the trace of the covariance matrix of the image gradients measures the variability of the gradients in the region and hence the degree of texture, providing a "smoothness" metric.
- the computation required during training may be reduced by sparsely sampling the plurality of pixels from the image data. This is especially useful for large images.
- Two algorithms that may be used include dense or sparse sampling. Dense sampling takes every 1/N pixels, whereas the sparse sampling method takes N pixels at random from the available pool of labelled pixels in the image data. In training, the sparse sampling method constructs a separate pool for each class being trained, leading to N pixels being available for every class. The dense method constructs a single pool for all of the classes, so the number of pixels in each class will depend on the distribution of that class in the images.
- the method may also identify the medium as a negative control.
- the classifier may classify pixels as background, or non-growth.
- the method may determine whether the number of pixels classified as background exceeds a predefined minimum number for the image capture and analysis to be valid.
- the predetermined minimum number of pixels that must be classified as background e.g. bare agar
- the predetermined minimum number of pixels that must be classified as background may be derived from a table of values that depend on the type of agar and the other bacteria that are identified on the culture plate. Note that the maximum is simply the number of pixels that lie on the culture plate in the image.
- the confidence of the classifications may be examined. If any regions of sufficient size exist with consistently low confidence, it is likely that an unknown item is on the culture plate. In this case, the operator can be alerted, or the culture plate marked for inspection.
- the results of the pixel classifications are analysed. This may involve processing the pixel classifications to extract additional metadata, such as colony counts, colony sizes, the range of organisms present on the culture plate, growth patterns, haemolysis, artefacts, contaminants, defects in the agar, etc, as described above.
- additional metadata such as colony counts, colony sizes, the range of organisms present on the culture plate, growth patterns, haemolysis, artefacts, contaminants, defects in the agar, etc, as described above.
- the number of each type of colony is counted (enumerated) using a process known as quantitation. It will be appreciated by those persons skilled in the art that counting the colonies is a difficult task, since many colonies may not be isolated from each other. Three approaches are therefore proposed:
- Density based counting where the distance from the edge of each colony is included in the counting algorithm. This is based on the observation that large areas of growth should be counted as several distinct colonies. Hence interior pixels in the centre of a colony are assigned higher weight to achieve such an outcome.
- a table is formed from measuring many colony counts manually, and tabulating the results against this metric. The table may then be used to obtain a colony count from the metric when evaluated on an unseen culture plate.
- Image processing-based counting where the image containing the labels is analysed first by a connected components algorithm. Each component detected by this algorithm is examined in turn. If a particular colony is deemed too large by a preset table of thresholds, then the method proposed as number one is applied. If the colony is deemed too small, it is counted as a single colony. Otherwise, the shape of the component may be considered. If it appears circular, it counts as a single colony. Otherwise, a watershed algorithm is applied to divide the component into individual colonies, each of which increments the colony count.
- an algorithm will compute an appropriate enumeration of each colony type detected on the culture plate.
- Such an enumeration is meaningful to a domain expert, and may take the form of a quadrant specification, such as ⁇ +', or an estimate of the number of colonies per millilitre of solution, such as 10 3 cfu/ml.
- the enumerations of one or more culture plates are then combined to derive the assessment of the sample.
- the method may also use clinical rules to interpret the classifications generated by the classifier, and/or the quantitation (enumeration) generated, as described above.
- clinical rules may assess the age or clinical condition of the patient, the type of patient sample analysed, the type of bacteria identified, local laboratory practice or a range of other custom user-defined conditions that will override the normal decisions of the classifier.
- the clinical rules provide a bridge from the low-level assessment provided by the software, such as exact colony counts and sizes, to a typical high-level assessment in a format appropriate for a laboratory technologist, where only salient information is reported.
- image data for the medium may be obtained by any suitable arrangement of one or more image capture devices (e.g. cameras) with one or more lighting devices.
- the image data may include the image of the surface of the medium, the plate containing the medium and of the area about the culture plate.
- the image data will be stored in a database along with relevant plate data, including the plate barcode data, the image acquisition parameters, a time/date stamp, and image capture device specification data.
- Steps can be taken to process the raw images from the image capture device into a suitable form. Processing the captured images may include averaging a series of captured images, followed by demosiacing, white balance and colour correction to obtain the image data suitable for analysis.
- the apparatus and the method can provide a complete image acquisition and analysis function, and can be provided to a laboratory as a complete package.
- the image data may be sourced from another supplier, location or country, remote from where the method steps of the present invention are conducted, and so any image acquisition apparatus used to obtain the image data may also be located away from where the analysis of the image data takes place.
- the present invention extends to software for use with a computer including a processor and memory for storing the software, the software including a series of instructions executable by the processor to carry out the method in accordance with any one of the embodiments described above.
- the present invention also extends to a computer readable media including the software, and apparatus including a processor, a memory and software resident in memory accessible to the processor, the software executable by the processor to carry out the method in accordance with any one of the embodiments described above.
- the present invention provides a method for training a classifier for use in analysing microbial growth on a solid culture medium, the method including: obtaining image data for a plurality of samples of microbial growth on the solid culture medium;
- results of pixel classifications of each said pixel are analysed to derive a microbiological assessment of the microbial growth on the solid culture medium.
- Fig. 1 is a schematic diagram of an apparatus for use in analysing microbial growth on a solid culture medium in a culture plate.
- Fig. 2(a) is a screen shot of a GUI for a user to capture images
- Figs. 2(b) to 2(e) are close up views of details on the left side of the screen shot of Fig. 2(a).
- Fig. 3 is a series of screen shots showing a GUI for entering the position of a culture plate in an image.
- Fig. 4 is a flow chart showing a method for capturing an image of a solid culture medium and any microbial growth thereon.
- Fig. 5 is a flow chart showing a method for processing the captured image.
- Fig. 6 is a flow chart illustrating a method for estimating a mapping from apparent colour to actual colour.
- Fig. 7 is a graph showing an example curve for correcting the brightness of a captured image.
- Fig. 8 is a diagram of a HSV space showing a plurality of control points within a circular hue space.
- Fig. 9 is a graphical representation of (a) the mapping across the hue/saturation space, (b) the isolated hue mapping function and (c) the isolated saturation mapping function for a particular image capture device.
- Fig. 10 is a graph showing the accuracy of the mapping.
- Fig. 1 1 is two photos of a colour chart (a) before colour calibration and (b) after colour calibration.
- Fig. 12 is a flow chart illustrating a method for mapping an apparent colour to an actual colour.
- Fig. 13 is a flow diagram showing the image capture pipeline run on a split culture plate.
- Fig. 14 is a flow chart showing a method for training a classifier for use in analysing microbial growth on a solid culture medium.
- Fig. 15 is a series of representations of (a) image data for microbial growth on a solid culture medium, (b) a red channel feature image, (c) a green channel feature image and (d) a blue channel feature image.
- Fig. 16 is a series of representations of (a) a L feature image, (b) an A feature image, (c) a B feature image, (d) a variance of L in an 1 1 x 1 1 window feature image, (e) a variance of A in an 1 1 x 1 1 window feature image, (f) a variance of B in an 1 1 x 1 1 window feature image, (g) a mean of L in an 1 1 x 1 1 window feature image, (h) a mean of A in an 1 1 x 1 1 window feature image and (i) a mean of B in an 1 1 x 1 window feature image.
- Fig. 17 is a representation of a position feature image for a split culture plate.
- Fig. 18 is a series of representations of (a) a feature image of gradient values in the x direction, (b) a feature image of gradient values in the y direction, (c) a smoothness feature image for a 5 x 5 window, (d) a smoothness feature image for a 13 x 13 window and (e) a smoothness feature image for a 21 x 21 window.
- Fig. 19 is an example of a decision tree.
- Fig. 20 is a flow chart of a method for analysing a series of images of microbial growth on a solid culture medium.
- Fig. 21 is a flow chart of a method for applying a graph cut algorithm to improve classification results of the boosted classifier.
- Fig. 22 shows (a) an initial graph and (b) a partitioned graph showing the labelling after a graph cut algorithm has been applied.
- Fig. 23 shows a graph constructed with an auxiliary node for use in the alpha expansion procedure.
- Fig. 24 is a set of graphs showing possible cuts that can be made in the alpha expansion procedure.
- GUI program run on a desktop computer is used to capture and process images, and to generate reports for unknown culture plates, and
- GUI program run on a desktop computer to provide training labels, and to manipulate metadata.
- Extensible Markup Language (XML) files are used to store metadata and the Portable Network Graphics (PNG) format is used to store images. All files may be stored directly on disk or in a database. Metadata is stored in three main XML file types: an image metadata file, a plate metadata file, and a main metadata file for storing additional metadata and linking the image and plate metadata files together.
- the library may be linked to a Laboratory Information Management System (LIMS) to access information about the patient and sample.
- LIMS Laboratory Information Management System
- Images data is stored in four ways: an image metadata file, the unprocessed image (raw values read from the camera), the processed image (that has been demosaiced and colour corrected) and a preview image (having a smaller size).
- the image metadata file captures a range of information.
- An example image metadata file is shown below:
- the metadata stored includes the exposure time, date of capture, light information, colour transform used to correct the image and camera data.
- a set of properties in the file can be used to locate and verify the particular image.
- the first property is the path, which provides the location of the image file, relative to the location of the main metadata file, discussed below.
- the image width and height are then stored, along with the size of the image on disk, in bytes.
- an MD5 checksum is stored, to ensure the data has not been modified since it was captured. These properties may be checked on image load.
- the XML files are cryptographically signed to prevent modification.
- a dish may have several associated plate metadata files.
- the metadata stored includes the type of agar, growth information appropriate to the sample type it contains, a list of organism types, which are defined by either the operator or the automatic analysis process.
- the organism list associates the name of the organism with a label number, providing a way for the labelled images to be interpreted.
- a mask for the culture plate is provided, to allow software to read the appropriate section of the labelled image.
- a single labelled image is used for a single Petri dish, but as a dish may contain multiple agars and therefore multiple plate metadata files, masks are used to determine which labels are associated with which agar.
- the metadata stored includes an internal ID of the culture plate, the type of culture plate it is, either whole or split, a patient ID (which is intended to be read from a barcode on the culture plate), the date of capture, the name of the operator who was using the software when the image was captured, the sample type that was placed on the culture plate (e.g. urine or enteric), and a list of references to plate metadata files.
- a diagnosis for the culture plate is then recorded, whose source can either be the operator or the automatic analysis software. Allowed values for the diagnosis are positive, negative, incubate or inspect.
- the operator may also enter arbitrary text, which is stored in the notes tag.
- the location of the culture plate is stored in a relative format. To obtain the pixel location of the culture plate, the x and y coordinates must be multiplied by the image width, w, and height, h, respectively. The radius must be multiplied by Each of the images that are attached to the culture plate
- a biology database may also be used, that lists acceptable values for the agar, sample and organism types of a particular metadata instance. The precise contents of the database are dependent on the laboratory where the system is deployed.
- the biology database is organised first by sample type, then by agar type. Each agar type then contains a list of valid organisms. Every entry in the database contains at least three fields: A name, a short description and a description. The name is used whenever a machine-readable name for the entity is required and acts as a key into the database. The short description is intended for use within a GUI in places where such text is appropriate, such as in menus or other user interface elements. The description can be arbitrarily long and is the full, technical name for a particular entity.
- Fig. 1 shows an embodiment of an apparatus 100 for use in analysing microbial growth on a medium in a culture plate 102 in the form of an agar plate.
- the apparatus 100 includes the following components:
- An image capture device 104 in the form of a high resolution digital camera 106 of machine vision quality with an appropriate fixed focal length lens, 108.
- the camera 106 is positioned about 200 mm above a ring light 1 10.
- the ring light 1 10 has a large diameter relative to the diameter of the culture plate 102. In this example, the ring light has a diameter of 180mm.
- the ring light 1 10 contains several hundred white LEDs arranged in a circular array and a diffuser. This light provides low angle, diffused side lighting to enable the culture plate to be uniformly illuminated.
- the ring light 1 10 is positioned around 40 mm above an opaque cover 1 12 that forms part of the frame 1 18, and thus about 30 mm above the culture plate 102. The positioning of the ring light 1 10 so that light from the white LEDs impinge on the surface of the culture plate 102 at a low angle prevents a specular reflection of the LEDs from a central surface of the medium being captured by the image capture device 104.
- a lighting device 1 14 in the form of a flat panel light based on an array of white LEDs behind a diffuser.
- the lighting device 1 14 is located about 150 mm below the opaque cover 1 12. This distance is chosen so that light from the ring light 1 10 falls on the baffles rather than the light 1 14, to reduce rear illumination of the culture plate 102.
- the support 1 16 is a transparent glass stage that is 3 mm thick. The glass may be replaced if it becomes scratched over time.
- the support 1 16 includes two or more triangle shaped transparent positioning elements for positioning the culture plate 102 on the support.
- a frame 1 18 positions the image capture device 104, support 1 16, ring light 1 10 and lighting device 1 14 relative to each other.
- the frame 1 18 is made of an opaque material, such as sheet metal or plastic, which reduces the amount of light entering the apparatus 100.
- the internal surfaces of the apparatus 100 are blackened where possible to reduce reflection of light from the internal surfaces into the lens 108.
- the frame 1 18 includes a door 120 providing an access patch for a human operator to place the culture plate 102 on the support 1 16.
- a robotic plate-handling device may use the access path to place the culture plate 102 precisely on the support 1 16 for imaging, and then to remove the culture plate to a designated output channel/slide.
- the culture plate may be placed in an output channel representing one of the up to four categories described above.
- the opaque cover 1 12 is an aluminium plate that extends across the width of the frame 1 18 and effectively splits the frame 1 18 into a top enclosure 122 and bottom enclosure 124.
- the opaque cover 1 12 includes a hole 126 to allow light from the lighting device 1 14 to transmit through to the culture plate 102.
- the width of the hole 126 is just slightly larger than the width of the culture plate 102 (which is 90mm in this example and is typically between 88 and 100 mm) and is less than the diameter of the ring light 1 10. This prevents light emitted from the ring light
- the frame 1 18 also includes light baffles 130 positioned below the opaque cover 112.
- Means 131 for changing the position of the ring light 1 10 relative to the support 1 16 are also provided in the form of a rack and pinion assembly.
- the frame 1 18, opaque cover 1 12 and light baffles 130 define a cavity 132 such that the support 1 16 supports the culture plate 102 between the image capture device 104 and the cavity 132.
- the support (glass stage) 1 16 seals the cavity 132 and prevents unwanted material from falling into the cavity 132.
- the opaque cover 1 12 prevents light from the ring light 1 10 from illuminating visible areas of the cavity 132. In this configuration, the cavity 132 looks like a black background.
- a side angle light 134 is used to illuminate the culture plate 102 from an angle to highlight any surface topography on the agar, such as dimples or a granular texture.
- An alternative to the side angle light 134 is to activate only some of the LEDs in the ring light 1 10, such that the culture plate 102 is illuminated from one direction only.
- a processing means such as a computer 136 is connected to the image capture device 104, the ring light 1 10 and the lighting device 1 14 via a physical or wireless interface.
- the computer 136 may include a processor 138 and memory 140 storing software 142 for activating the different components, capturing raw data and processing the data.
- a library of images, metadata and other information may be stored at the computer 136, or may be accessible at the computer 136 via a network. Similarly, the LIMS may be accessed via the computer 136.
- the distance between components and position of components may be adjusted.
- the camera 106 and lens 108 are shown inside the frame 1 18, in another example, they could be positioned outside the frame 1 18, with the lens 108 protruding through a hole in the top surface of the frame 1 18.
- the width of the frame 1 18 could be decreased to reduce the overall size of the apparatus 100.
- This process may be suitable for obtaining images for use in classifying microbial growth on the culture plate 102 using a trained machine learning classifier, or in training such a classifier.
- a manual process will be described, where many steps are performed by a human operator, but it will be appreciated that many of the steps of the process may be automated and performed in software or by a robotic device.
- an inoculated and incubated culture plate 102 is positioned on the support 1 16, within the triangular stops, by a user.
- Culture plates 102 are generally stored within a laboratory with the agar facing down (to prevent condensation on the lid from falling on to, and damaging, the agar surface), so positioning the culture plate 102 on the support may include removing a lid of the culture plate 102 and rotating the culture plate so that the agar is facing upwards.
- the software 142 is activated to begin the image capture process.
- the software 142 requires the user to scan a barcode on the culture plate 102, or enter a number manually.
- the barcode links to a sample ID, which links the culture plate to a particular sample and, via a LIM system, to a particular patient.
- a live video preview of the camera output appears in the window shown in Fig. 2(a).
- the user may adjust the location of the culture plate 102 or the focus or aperture of the lens 108, based on the feedback provided by the live video stream.
- the user is then required to select the culture plate type (e.g. split or whole), sample type (e.g. urinary, enteric, blood or tissue) and agar type(s) (e.g. blood or chromogenic).
- sample type e.g. urinary, enteric, blood or tissue
- agar type(s) e.g. blood or chromogenic
- An example of a plate data selection options is shown in Fig. 2(b) and at the top left of Fig. 2(a).
- this information could be extracted from the LIM system, based on the barcode entered.
- the culture plate position is entered.
- This information can be entered by the user. The first is through a traditional set of sliders, which allow entry of the x, y location of the dish, along with its radius. If the dish is a split dish, the user must also enter the location of the split. An example of this is shown in Fig. 2(c) and the bottom left of Fig. 2(a).
- the second method is through manipulating a set of markers over the live video preview. The user places the circle around the agar via three points, which the user is able to move interactively. These three points uniquely define a circle.
- a fourth control point allows the user to specify the location of the central split, if appropriate. Such control points are quick and easy to place very accurately.
- Fig. 3 shows an example of how such an interaction might occur. In a system with a robotic culture plate placement, the location of the culture plate will be known, and the location of the split can be computed.
- the exposure and light settings may be drawn from a database on the basis of the culture plate, sample and agar type(s) selected.
- Fig. 2(d) shows radio buttons that may be used to select either the top light or bottom light to activate. Although not shown, the user may alternatively choose the side angle light to activate.
- the exposure settings may be adjustable by the user, for example using a range of automatic exposure algorithms that may be selected using a drop down menu such as shown in Fig. 3(e).
- Each of the automatic exposure algorithms has the same core. Given a target brightness, and a measured brightness, the new exposure,
- the difference between the various automatic exposure algorithms is in how they compute the brightness of the current image.
- Different algorithms that may be used include: Mean - In this mode, the mean brightness of all pixels in the image is computed.
- Centre-Weighted Mean computes the mean brightness of all pixels, but gives those pixels within of the centre additional weight (each centre sample is counted 8 times).
- Median - This method computes the median brightness across the image via a histogram calculation.
- Each bin of the histogram is 4 levels wide, hence if the input images are 16-bits the histogram has 16384 bins.
- the true bit-depth of the images could also be read from the camera and provided to the algorithm.
- Green - This method is the same as Mean, but uses the green channel alone to compute the brightness. This is advantageous for two reasons. First, there are twice as many true green pixels in a Bayer mosaiced image as red or blue. Second, the human eye is sensitive to green, so using it to adjust the images should provide a conceptually high quality image.
- the software 142 may further include processes for assisting the user to select good exposure settings, for example, by pulsating pixels between their normal colour and red. This helps to identify overexposed, or saturated, pixels, which can be rectified by reducing the target image brightness.
- Initial exposure settings are loaded from the database and adjusted prior to each capture using one of the algorithms described previously.
- image capture is performed using a method shown in Fig. 4.
- capture is initiated, at step 166 the lighting configuration (e.g. the ring light 1 10) is activated.
- the image is captured at step 168 by image capture device 104 using this lighting configuration.
- Image capture is repeated a number of times for each culture plate, each time having the same exposure (step 170). Each of these images is averaged together during the processing stage to reduce image noise. A sequence of five images is sufficient.
- the process is repeated for the other side of the culture plate in the case of a split culture plate (step 172) and for other lighting configurations, e.g.
- step 174 the captured images are handed off for data processing and then image capture is complete (step 178).
- image capture is complete (step 178).
- the images may be handed off to a separate thread of execution for asynchronous processing.
- the capture window then reverts to waiting for a barcode to be entered.
- the image, along with any salient metadata is saved in the library.
- a robotic arm may place the culture plate on the support before image capture and remove it after image capture.
- the software could automatically detect the position of the culture plate in the image and whether it is split or whole.
- a robotic system may be used to read an identifier or barcode on the culture plate.
- the barcode or other identifier may enable access to information such as culture plate type, sample type and agar type so that this information does not need to be entered by the user.
- the user would not adjust the exposure or target brightness settings.
- the raw data from the image capture device is processed to obtain a more accurate image for the sample that can be used to generate feature images for use in the boosting algorithm for generating the boosted classifier. It will be appreciated that although this embodiment describes the use of a boosted algorithm, the classifier could alternatively be generated using a different technique.
- the image processing starts at step 163, and includes at step 165 averaging the images taken by the image capture device, at step 167 demosaicing the averaged images, at step 169 colour correcting the resulting image, at step 171 computing the feature vectors and at step 173 stopping the processing.
- the images are averaged to combine the five (or other number) of images taken at the same exposures into a single overall image.
- the use of five averaged images reduces the effects of noise.
- Demosaicing is performed because most cameras have sensors that can measure only one colour at each physical location, for example using a Bayer arrangement of colour filters. This means the raw data from the camera is a mosaiced pattern of individual red, green and blue pixels.
- Suitable methods of demosaicing include: o Nearest Neighbour - Nearest neighbour is a simple method that produces fast, low quality results. In this method, each 2 x 2 block is processed together. The red and blue values are copied from their original location to the three other pixels missing that respective value. The green values can then be copied either vertically or horizontally. This results in each 2 x 2 block having all channels filled in at all locations.
- each 2 x 2 block is collapsed into a single pixel.
- the two green values are averaged and the red and blue values are simply assigned. This method requires no interpolation and hence is the most faithful to the original pixel data.
- AHD Adaptive-Homogeneity-Directed demosaicing
- Colour correction of the image involves correcting for white balance and brightness, and using a pre-calculated mapping to map apparent colour of the averaged image to true colour.
- White balance and colour correction are beneficial for ensuring accurate determination of agar and colony colour(s), which is important for speciation and the detection of phenomenon such as fermentation and haemolysis. Hence high quality colour calibration is important.
- the white balance and colour correction parameters used to correct the image are specific to the image capture device and lighting arrangement and are pre-calculated using a colour chart.
- a colour chart With reference to Fig. 6(a), at step 180, an image of a reference colour chart is captured.
- the chart may be a standard "gretagmacbeth ColorCheckerTMColor Rendition Chart", which has 24 colour patches, each of which has an associated reference red, green and blue channel value.
- the image may be captured by activating the camera 106 manually or via electronic signals sent using the software 142.
- the data captured by the camera sensors is transmitted via the physical or wireless connection to the computer 36, with colour calibration disabled.
- the user indicates the location of each colour patch in the reference colour chart.
- the software could cause a GUI to display a preview of the image from the camera 106, and the user may click in the four corner squares of the chart in a specific order to enable the locations of the colour patches to be determined.
- the location of the patches may be determined using a different user interface, machine vision techniques or by the placing of the reference colour chart in a specific known location that is a fixed distance from the camera.
- the average colour of each location is captured in a 21 x 21 pixel window and used for the patch colour at step 184.
- different sized pixel windows could be used to compute the average colour.
- the white balance is computed from a single patch - namely the third darkest grey patch in the reference colour chart. It is assumed that the R, G and B values are equal in this neutral grey patch.
- the white balance is represented as:
- w g is implicitly defined as one.
- the weightings w r and w are calculated to produce r' and b' values equal to the g' value.
- the computed white balance values will be used to correct the white balance of images captured by the camera.
- a curve C is computed which maps the apparent grey colour values together as well as possible. This provides a brightness mapping for the camera.
- the form of the curve is: where g is the input colour, g' is the output and a, b and c are the parameters of the curve C that are computed.
- Fig. 7 shows an example curve obtained from this process. Once estimated, the curve may be applied independently per colour channel.
- the final stage of the colour mapping function (step 190) aims to produce a mapping for the hue and saturation of each pixel.
- the apparent hue and saturation of a plurality of the colour patches are measured, based on the white balance and colour corrected image and a hue and saturation mapping is computed at step 190.
- the hue mapping is provided using a series of control points to map the apparent hue values to actual hue values. It uses cubic interpolation (or another higher order interpolation method) to achieve this.
- Each control point consists of a position in a circular hue space, which is fixed, and an angular offset, which is computed during optimisation.
- the offset is cubically interpolated to provide an offset for a particular hue value that will map the particular hue value to an actual hue.
- the saturation mapping function uses a series of control points to scale saturation values. These control points may be the same or different from the control points used to map the hue values.
- the saturation mapping function may use the same interpolation scheme as the hue mapping function. Note that saturation must be in the range [0, 1] and hence the output value from this function is clamped to ensure this range is enforced. In an alternative, the saturation mapping may be done via a curve similar to that applied for brightness correction.
- the control points used in the hue and saturation mapping functions may be spread evenly around the edge of the circular hue space as shown in Fig. 8.
- This circular hue space includes 16 control points, represented by coloured circles (for example, control point 192).
- the coloured circle at hue of 0 degrees and saturation of 1 may be red
- at hue of 90 degrees and saturation of 1 may be green
- at hue of 180 degrees and saturation of 1 may be blue and at hue of 270 degrees and saturation of 1 may be purple.
- the direction of angular offset 194 and scale 196 is also shown in Fig. 4.
- the angular position of the control points are defined by their hue value and the radial position of the control points are defined by their saturation.
- Two rings of control points are shown, an inner ring having saturation of 0.5 and an outer ring having saturation of 1. In the case of a single ring of control points, only a single value is required to select the appropriate point, the value corresponding to hue. Hue is expressed as an angle and the function must take account of this when selecting control points.
- the maximum number of control points is defined by the number of colour samples available. Each control point has two parameters, and each sample provides three residuals. Hence, n colour samples can have at most 3n/2 control points. In practice, however, the measurement process induces errors in the colour measurements. Selecting the control points may include computing an error for a series of candidate mappings using different numbers of control points, and selecting a mapping from the series based on the minimal number of points yielding an acceptable error. . It has been found that 12 control points are sufficient for a chart with 24 patches.
- the angular offsets and scales are optimised to best map the apparent colours of patches from the image onto the actual colours of the equivalent colour patches in the reference colour chart. This is done by calculating a mapping f (p), such that the distance between the true and mapped colours is minimised:
- the colour may be represented in any colour space, for example RGB, YUV, LAB or XYZ.
- the cost function includes terms which ensure the scales maintain a value around one and the offsets around zero to prevent very large changes (such as offsets larger than 360°). Given the following error measure: where is the offset for the control point and s, is the scale for control point. It is sought to select and s i to minimise ie A solution to this equation may be obtained using the Levenberg-Marquardt algorithm, though any other unconstrained minimiser would also suffice.
- the cost function tries to minimise the error in mapping the colours together, while keeping the hue offsets near zero and the saturation scales near one.
- an angular offset 194 and scale 196 is determined for each control point so that the sum of squares of the distance between each transformed apparent colour (e.g. RGB) and the reference colour (e.g. RGB) for each patch is minimised whilst holding the angular offset near zero and the scale close to one (minimises 1 minus the scale).
- Fig. 9 shows a graphical representation of (a) the mapping across the hue/saturation space, (b) the isolated hue mapping function and (c) the isolated saturation mapping function for a particular camera.
- Fig. 10 provides some details on the accuracy of the mapping.
- Each position on the y-axis represents one of the 24 colour patches and the x-axis represents the colour values, where each can be between zero and one.
- An example of colour correcting an image using the mapping is shown in Fig. 1 1.
- the first image (a) shows the input image, as captured by the camera, after the user has selected the position of the colour patches. Note that in this example, the white balance is already roughly correct.
- the second image (b) shows the corrected colours. The small coloured squares visible over the patches show the true colour of each patch.
- a method for mapping an apparent colour to an actual colour includes at step 181 capturing an image using the camera.
- the apparent RGB colour values for each pixel may be determined and the colour values may be white balance and brightness corrected using the weights/curve previously computed.
- the method then involves at step 183 for one or more pixels in the image determining the apparent hue and saturation of the pixel (step 185), interpolating between the angular offsets of two or more control points in the circular hue space to determine an angular offset associated with the apparent hue (step 187), interpolating between the scales of the two or more control points in the circular hue space to determine a scale associated with the apparent saturation (step 189), applying the angular offset to the apparent hue (step 191 ), and applying the scale to the apparent saturation (step 193).
- the HSV space colour can be transformed back into the RGB space, or to any other output space as desired.
- a linear method may also be substituted for the HSV method, where the colours are mapped via:
- T is a 3x3 transformation matrix and the remaining variables are scalars.
- Fig. 13 shows the overall image capture pipeline for a split culture plate having half blood agar and half chromogenic agar.
- the pipeline runs twice, once for each lighting configuration (top light illuminated and bottom light illuminated).
- the pipeline also has two branches, to capture images of each side of the culture plate separately, each branch being optimised for the type of agar.
- the relevant halves of the demosaiced images are stitched together to form a composite image 214 which is colour corrected 216.
- the raw images may be composited together, and the composited image demosaiced, reducing the computation required.
- the result is image data for a medium.
- the same process of capturing and processing images may be used to obtain image data for training the classifier, and image data for classifying a culture plate.
- the image data is labelled by an experienced microbiologist. This may be done off-site and away from the imaging and lighting apparatus. Labelling is the process of providing correct labels for the image data for a particular culture plate. The choice of labels may be different for different agar and bacteria types of interest. At the time of labelling, the operator may also check the metadata for the culture plate to ensure that it is correct. The labelling may be done via simple drawing tools, typical of any painting software. The labelling program may load metadata associated with a particular image, and use that to determine what organisms are applicable to a particular culture plate. The user may then select which side of the culture plate they wish to label, if it is a split culture plate, which may change the set of available organisms.
- the user has three main tools available for painting: A brush tool, a fill tool and an eraser.
- the brush tool is a variable sized circular brush.
- the fill tool performs flood fill from the point under the cursor. It fills contiguous regions of background or a particular label colour, depending on what was under the cursor when clicked.
- the eraser functions the same as the brush tool, but writes the 'unknown' label rather than an organism label.
- the eye dropper is used to change the currently selected organism to the one under the cursor. It also sets the window status to the name of the organism under the cursor, when moving the mouse around on the screen.
- the inspector sets the window status to the current colour under the cursor and position in pixels of the cursor on the image. It also highlights any organisms under the cursor and provides their size in pixels.
- the plate location tool is the same as the tool in the capture window, and is used to adjust the position of the culture plate in the image.
- each pixel within the selected area is assigned that label.
- the labelled image data is stored in the library, together with metadata extracted from the labels.
- a boosted decision tree classifier is trained to distinguish between the plate itself and bacterial growth, and between different types of growth.
- a method for training the classifier includes at step 220, obtaining labelled image data for a plurality of samples of microbial growth, at step 222, for each labelled image data: for a plurality of pixels in the image data, generating a feature vector of values obtained by applying one or more filters to the image data (at step 224), and at step 226, using the feature vectors to train a boosted classifier to classify each pixel in the plurality of pixels according to a label associated with the pixel in the labelled image data.
- the one or more filters are applied to transform the image data and extract information which can be used to train the classifier, allowing it to learn the distinctive features of each type of bacterial growth.
- the filters extract information on the colour and texture of the culture plate at and around each pixel.
- the output of the filters may be a set of feature images.
- RGB filter - Some growth types can be easily distinguished by their colour.
- the RGB filter is the simplest filter type, and copies the RGB channels of the image data into the feature vector, giving R, G, and B values for each pixel.
- Fig. 15 shows the original image data (a), the red channel feature image (b), the green channel feature image (c) and the blue channel feature image (d).
- LAB filter - Additional colour information is provided by the LAB filter. This filter primarily converts RGB image data into the LAB colour space (where the L channel represents lightness, and the A and B channels contain colour information). This space allows for easier discrimination between the colours of some bacteria types.
- this filter may be used to measure the mean and variance of L, A, and B in different sized windows.
- 16 shows a L feature image (a), an A feature image (b), a B feature image (c), a variance of L in an 1 1 x 1 1 window feature image (d), a variance of A in an 1 1 x 1 1 window feature image (e), a variance of B in an 1 1 x 1 1 window feature image (f), a mean of L in an 1 1 x 1 1 window feature image (g), a mean of A in an 1 1 x 1 1 window feature image (h) and a mean of B in an 1 1 x 1 1 window feature image (i).
- Oriented Gaussian Filter which applies a 1 D Gaussian kernel in a variety of directions to the image. It is also possible to utilise first and second order derivatives of the Gaussian kernel.
- Position filter provides the classifier with the distance of each pixel from the edge of the culture plate. This feature was motivated by the observation that very difficult cases often occur on the edge of the culture plate either on the meniscus of the agar, or above the plastic feet of the dish. The presence of the position information allows the classifier to adapt for the varying appearance of the culture plate at different distances from the edge. The distance of each pixel from the edge is obtained by performing a distance transform on the mask image. An example of a position feature image is shown in shown in Fig. 17.
- Colour quantisation filter where the number of colours in the image is reduced, usually via a histogram, to help locate areas of similar colour.
- Texture filter aids the classifier in distinguishing between types with similar colour, but different surface textures.
- the image is first converted to greyscale. To reduce noise, the image is then smoothed by convolution with a Gaussian kernel.
- the 2D Gaussian kernel is defined as:
- the texture filter then extracts edge information, returning the greyscale x and y image gradients at each pixel. Gradient values for the pixels are extracted by convolution with the Sobel kernels,
- FIG. 18(a) A feature image of gradient values in the x direction is shown in Fig. 18(a), and a feature image of gradient values in the y direction is shown in Fig. 18(b).
- the filter also computes a 'smoothness' metric in 5 x 5, 13 x 13 and 21 x 21 pixel windows around the pixel, measuring how textured the culture plate is in that region.
- the smoothness metric used is the trace of the covariance matrix of the gradient values in each window, giving a measure of the total variance of the gradient in that window. Examples of smoothness feature images for a 5 x 5, 13 x 13 and 21 x 21 window are shown in Fig 17.(c), (d) and (e) respectively.
- a vector of feature values x (x-i , x n ) for each of a plurality of pixels is generated from these feature images, and passed to the classifier. If there is more than one image for a culture plate (for example a set of first images taken using a top lighting configuration, and a set of second images taken using a bottom lighting configuration), then the feature vector may include values from feature images created using either or both sets of images. A label y associated with each pixel is also passed to the classifier.
- a feature vector may be generated for every pixel in the image data, or only in relation to a sample of pixels. Sampling the pixels reduces the computation required and therefore reduces the time taken to train the classifier.
- a random sampling procedure may be used to select the set of examples used in training from the complete set of labelled pixels. This may be a sparse sampling or dense sampling procedure. The sparse sampling method constructs a separate pool for each class being trained, from which N samples are selected. If the number of labelled pixels for a class is less than this specified number of samples, pixels are sampled with replacement. Training on equally sized sets of examples for each class avoids biasing the classifier towards better performance on more commonly seen types. The dense sampling method retrieves one in every N samples. In training the background classifier, the set of negative examples includes examples of each growth type. For N negative samples and J growth types, N/J samples of each type are selected.
- a classifier To classify each pixel, and hence either label it as plate background or one of a set of growth types, a classifier must be trained.
- a boosted classifier is trained, using decision trees as weak classifiers.
- a decision tree is built recursively, beginning with a root node, by finding the test at each node which gives the optimal split of the input data into positive and negative samples, then dividing the data according to the test and repeating the process for each child node on its subset of the data.
- a child node will be split unless the tree has reached a maximum specified depth, all samples in a node belong to the same class, no split can be found giving a better than random separation of the data, or a node has too few samples to perform a meaningful split.
- An example decision tree is shown in Fig. 19.
- the boosting procedure used in this embodiment is Discrete AdaBoost. Pseudocode for Discrete AdaBoost is given below:
- Discrete AdaBoost iteratively trains a series of T weak classifiers h t (x), adapting to improve training performance with each additional classifier.
- Each classifier is trained on the set of N labelled training examples (x, y), with a weight w , associated with each example indicating that example's relative importance.
- classifier is optimised to minimise the error which is the sum of the weights of the examples classified incorrectly by that classifier.
- the weights are updated to place a higher importance on examples which are misclassified by the current set of classifiers, to improve performance on these examples for the next classifier. This is done by increasing the weight for a misclassified example based on the error of the classifier, and normalising the weights across all of the examples.
- the AdaBoost. MH procedure may be used to classify for more than two types using a binary classifier.
- an additional feature value is added to the feature vector for each pixel, having an integer value (1 , J) indicating the class of the training example.
- each training example is repeated J times with the label y for the example set to +1 if J is the class of the sample, or -1 otherwise.
- the pixel is assigned the class with the strongest classifier result.
- a confidence value may be set for that pixel depending on the relative strength of the strongest classifier.
- the confidence value in assigning the label /,- to feature vector x, computed with the formula for probability of correct classification in a multiclass classifier with J classes may be derived using the following equation:
- classifiers may be run independently in parallel to obtain results for particular colony types, media types, background clutter, additional colony properties or other distinguishing features.
- classification may be performed as a two stage process.
- a binary classifier distinguishing between background pixels and all other types, quickly classifies the majority of background pixels.
- a multiclass classifier then classifies the remaining pixels.
- the initial binary classifier may follow the cascade classifier method.
- This classifier is intended to rapidly classify much of the background, with a more accurate classification performed in the next stage. As such, a relatively high false positive rate (background pixels classified as foreground) is acceptable, as misclassified pixels can be correctly classified in the next stage. However, the false negative rate should be very low, as this cannot be corrected in the next stage.
- the number of trees used by the classifier is kept as small as possible.
- An initial classifier is trained, and its classification performance is tested against a specified maximum false positive rate. If it is greater than this threshold, a new classifier is trained, and this continues until the combination of classifiers has the desired false positive rate. As each classifier is added, the threshold for classification as foreground is adjusted to achieve the desired false negative rate for that classifier stage.
- f is the maximum acceptable false positive rate per stage
- d is the minimum acceptable true positive rate per layer
- F is the false positive rate for stage / '
- D is the true positive rate for stage /
- F target is the maximum overall false positive rate
- P is the set of positive examples
- N is the set of negative examples
- i is the classification threshold for each layer, above which a prediction by the classifier at that stage is considered positive.
- the output from the training stage is a boosted classifier that can be used to classify pixels in image data for new specimens.
- labels in the labelled image data may be edited, principally to correct classifier errors, and the classifier re-trained, creating a feedback loop. This may be repeated any number of times, but, in practice, two or three times is usually sufficient.
- the results from a classifier may also be used as input into a filter, to create feature vectors for training a new classifier as described above.
- the classifier may be trained on a server, with a library of training image data.
- a training program may be executed, and upon completion a classifier written out, along with various statistics and results from the training run.
- the classifier can then be used to automatically label pixels on a new culture plate.
- Fig. 20 A method for analysing a new culture plate having unknown microbial growth is shown in Fig. 20.
- the metadata for the culture plate is set.
- the metadata may be entered by a user or obtained from a LIM System as described above. Options for metadata values may be chosen from a biology database 252.
- an image of the culture plate is captured (or images of multiple culture plates are captured if the culture plates are being bulk processed). The image capture may be performed by an operator via a GUI program as described above.
- the software then takes the image, demosaics it and detects the culture plate location at step 256, at step 258 runs a set of image filters upon the demosaiced image and then at step 260 feeds the results into the appropriate classifier.
- the filtering and classification steps are iterated if necessary.
- the classifier then generates a per-pixel estimate of the bacteria on the culture plate, which includes a predicted label and the confidence in the label.
- a graph cut is run on the estimate and the output of the graph cut is used to generate a report at step 264.
- Graph cut algorithms compute an optimal partition of a graph into two or more sets.
- a method for improving pixel classification results is shown in Fig. 21.
- a graph is constructed with each node corresponding to a pixel (or one of a sample of pixels) in the image data. Labels are also added, each label corresponding to a class.
- edges are added between nodes corresponding to adjacent or neighbouring pixels, and at step 274 an edge is added between each node and each label.
- the graph cut algorithm is used to cut edges to a node and partition the graph into classes, the graph cut algorithm using as input the pixel classification results for the pixel corresponding to that node and neighbouring pixels.
- Fig. 22 shows an example of an initial graph (a) and the partitioned graph (b) showing the labelling after the cut.
- a source node S and sink node T correspond to the possible node labels.
- the graph cut determines a label for each node, cutting the edge between each node and the sink or source, and cutting edges between adjacent nodes with different labels.
- Each time a particular link is cut a cost is incurred.
- the graph cut procedure minimizes an energy function of the form:
- N is the set of nodes
- M is the set of all neighbouring nodes
- x is the feature data
- U(l,; x,) assigns a cost to each node for each possible label assignment (implemented as a cost on cutting the edge to the source or sink)
- V(l p , l q ) assigns a cost to each edge for assigning different labels to the attached nodes.
- pl i (X i ) is the confidence in assigning label I i to feature vector x i , computed with the formula for probability of correct classification in a multiclass classifier with J classes where h(x,k) is the classifier result for class k.
- Other inputs into the graph cut algorithm may include image colour and edge data.
- the expected colony size for a labelling could be included in the U (l i ; x i ) cost assignment.
- Another factor that may be taken into account is a predetermined rule from an expert system, such as definitions of colony types that would not appear next to each other or are likely to appear together. This may be incorporated into the smoothness term V(l p , l q ). If the number of possible labels is greater than 2, an alpha expansion procedure is used to perform a multiple class labelling from a binary labelling. Psuedocode for an alpha expansion procedure is given on page 2 of Yuri Boykov, Olga Veksler, and Ramin Zabih "Fast approximate energy minimization via graph cuts" IEEE Trans.
- the procedure iterates over steps 3.1 and 3.2 for each of the possible classes of, for example, bacteria type or background, checking whether any labels in different classes could be switched to that class without increasing the total energy E of the graph. Once the procedure has been repeated for each class and no further energy reduction is possible, the optimal labelling is returned.
- auxiliary nodes are added between adjacent nodes with different labels, to include the cost of this labelling in the cut.
- Fig. 23 shows the addition of an auxiliary node a between nodes p and q, and gives the edge weights for this graph.
- the cost of a cut is equal to the sum of its edge weights.
- a cut must sever one of three groupings of edges, as shown in Fig. 24. This diagram was taken from page 6 of Yuri Boykov, Olga Veksler, and Ramin Zabih's paper.
- the alpha expansion procedure determines which of the three groupings has the minimum edge weight and cuts these edges.
- a pixel is assigned the label alpha if the cut separates the pixel from the label alpha.
- the resulting classifications of the pixels may be analysed to output an assessment. This may involve counting how many pixels are labelled in each class to determine the types and quantities of growth on the culture plate. It may involve assessing whether the number of pixels of a particular class is greater than a predetermined number. Three approaches for counting colonies are proposed:
- Basic counting where the colony count is estimated by determining an average colony size for each organism type on each type of agar. The number of pixels of each organism detected on a particular culture plate can be divided by this number to yield the number of colonies.
- Density based counting where the distance from the edge of each colony is included in the counting algorithm. This is based on the observation that large areas of growth should be counted as several distinct colonies. Hence interior pixels in the centre of a colony are assigned higher weight to achieve such an outcome.
- a table is formed from measuring many colony counts manually, and tabulating the results against this metric. The table may then be used to obtain a colony count from the metric when evaluated on an unseen culture plate.
- Image processing-based counting where the image containing the labels is analysed first by a connected components algorithm. Each component detected by this algorithm is examined in turn. If a particular colony is deemed too large by a preset table of thresholds, then the method proposed as number one is applied. If the colony is deemed too small, it is counted as a single colony. Otherwise, the shape of the component may be considered. If it appears circular, it counts as a single colony.
- the counts are then quantified to produce a meaningful measure of growth for a microbiology expert.
- the assessment may include an identification of the medium as a negative control.
- This method determines an expected value for the area of bare agar (e.g. a predefined minimum number of pixels classified as background) and this must exceed a certain minimum area for the image capture and analysis to be valid.
- the precise area of bare agar that must be visible is derived from a table of values that depend on the type of agar and the other bacteria that are identified on the culture plate.
- Clinical rules are used to analyse the result of the pixel classifications and/or quantitation and generate the assessment.
- the clinical rules may involve a table incorporating the type of patient sample, the type of agar, the mix of bacteria that have been identified, national standards, local laboratory practice and some user- customisable rules.
- the main metadata and plate metadata files may be updated to reflect the assessment and to add further information extracted from the classifications.
- each culture plate may then, in an embodiment, be sent to one of four output plate stacks of the apparatus 100 (not shown in the Figures): (1 ) waste, (2) re-incubate, (3) identification of potentially pathogenic bacteria, or (4) human review required. It is envisaged that in some forms some of these stacks may be merged into a single stack. In any event, an operator can then remove culture plates from the output plate stacks for further action. Once all culture plates have been processed, the whole workflow can be repeated or the system can be shut down.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- Quality & Reliability (AREA)
- Radiology & Medical Imaging (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Medical Informatics (AREA)
- Data Mining & Analysis (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Apparatus Associated With Microorganisms And Enzymes (AREA)
- Measuring Or Testing Involving Enzymes Or Micro-Organisms (AREA)
- Image Analysis (AREA)
- Investigating Or Analysing Biological Materials (AREA)
- Investigating Or Analysing Materials By Optical Means (AREA)
Abstract
Description
Claims
Priority Applications (7)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201280021787.9A CN103518224B (en) | 2011-03-04 | 2012-03-02 | Method for analysing microbial growth |
KR1020137026188A KR101989202B1 (en) | 2011-03-04 | 2012-03-02 | Method and software for analysing microbial growth |
JP2013555707A JP5997185B2 (en) | 2011-03-04 | 2012-03-02 | Method and software for analyzing microbial growth |
EP12754708.1A EP2681715B1 (en) | 2011-03-04 | 2012-03-02 | Method and software for analysing microbial growth |
ES12754708T ES2928897T3 (en) | 2011-03-04 | 2012-03-02 | Method and software to analyze microbial growth |
AU2012225196A AU2012225196B2 (en) | 2011-03-04 | 2012-03-02 | Method and software for analysing microbial growth |
US14/002,722 US9292729B2 (en) | 2011-03-04 | 2012-03-02 | Method and software for analysing microbial growth |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
AU2011900786 | 2011-03-04 | ||
AU2011900786A AU2011900786A0 (en) | 2011-03-04 | Method and software for analysing microbial growth |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012119191A1 true WO2012119191A1 (en) | 2012-09-13 |
Family
ID=46797325
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/AU2012/000211 WO2012119191A1 (en) | 2011-03-04 | 2012-03-02 | Method and software for analysing microbial growth |
Country Status (8)
Country | Link |
---|---|
US (1) | US9292729B2 (en) |
EP (1) | EP2681715B1 (en) |
JP (1) | JP5997185B2 (en) |
KR (1) | KR101989202B1 (en) |
CN (1) | CN103518224B (en) |
AU (1) | AU2012225196B2 (en) |
ES (1) | ES2928897T3 (en) |
WO (1) | WO2012119191A1 (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2824921A3 (en) * | 2013-07-08 | 2015-04-15 | Samsung Display Co., Ltd. | Image and video in mosaic formats |
EP3147893A1 (en) * | 2015-09-24 | 2017-03-29 | Dolby Laboratories Licensing Corp. | Light field simulation techniques for dual modulation |
CN107580715A (en) * | 2015-04-23 | 2018-01-12 | Bd科斯特公司 | Method and system for automatic counting microbial colonies |
EP3848472A3 (en) * | 2020-01-13 | 2021-12-15 | Airamatrix Private Limited | Methods and systems for automated counting and classifying microorganisms |
US11341648B2 (en) | 2015-04-23 | 2022-05-24 | Timothy M. Wiles | Colony contrast gathering |
Families Citing this family (30)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9292927B2 (en) * | 2012-12-27 | 2016-03-22 | Intel Corporation | Adaptive support windows for stereoscopic image correlation |
JP6174894B2 (en) * | 2013-04-17 | 2017-08-02 | キヤノン株式会社 | Image processing apparatus and image processing method |
AU2015212758B2 (en) | 2014-01-30 | 2019-11-21 | Bd Kiestra B.V. | A system and method for image acquisition using supervised high quality imaging |
EP3102664A4 (en) * | 2014-02-04 | 2017-11-01 | Molecular Devices, LLC | System for identifying and picking spectrally distinct colonies |
CN103871031A (en) * | 2014-03-07 | 2014-06-18 | 西安电子科技大学 | Kernel regression-based SAR image coherent speckle restraining method |
CN105095901B (en) * | 2014-04-30 | 2019-04-12 | 西门子医疗保健诊断公司 | Method and apparatus for handling the block to be processed of sediment urinalysis image |
WO2016013885A1 (en) * | 2014-07-25 | 2016-01-28 | Samsung Electronics Co., Ltd. | Method for retrieving image and electronic device thereof |
FR3026745B1 (en) * | 2014-10-06 | 2018-03-02 | Fondation Mediterranee Infection | METHOD OF ANALYSIS AND AUTOMATED INTERPRETATION OF ANTIBIOGRAM |
WO2016120463A1 (en) * | 2015-01-31 | 2016-08-04 | Ventana Medical Systems, Inc. | Systems and methods for area-of-interest detection using slide thumbnail images |
CN106097231A (en) * | 2016-08-15 | 2016-11-09 | 黄星 | A kind of microorganism monitoring statistical management method for clean area |
US10585121B2 (en) * | 2016-09-12 | 2020-03-10 | Tektronix, Inc. | Recommending measurements based on detected waveform type |
EP3535382A4 (en) * | 2016-11-04 | 2020-06-10 | Becton, Dickinson and Company | System and method for selecting colonies |
US11057542B2 (en) | 2017-04-26 | 2021-07-06 | Hewlett-Packard Development Company, L.P. | Color calibration |
US11727699B2 (en) * | 2017-10-05 | 2023-08-15 | Becton, Dickinson And Company | Application development environment for biological sample assessment processing |
CN107833187A (en) * | 2017-10-27 | 2018-03-23 | 中国农业大学 | A kind of facilities vegetable disease method for early warning and system based on spore count |
US11982627B2 (en) | 2018-01-31 | 2024-05-14 | Monsanto Technology Llc | Systems and methods for quantitative phenotype analysis |
CN112424612A (en) * | 2018-07-23 | 2021-02-26 | 株式会社岛津制作所 | Microfluid device observation apparatus and microfluid device observation method |
EP3620984B1 (en) * | 2018-09-06 | 2024-04-10 | Accenture Global Solutions Limited | Digital quality control using computer visioning with deep learning |
WO2020050687A1 (en) * | 2018-09-07 | 2020-03-12 | 주식회사 더웨이브톡 | Microorganism information providing device and method |
BR112021005978A2 (en) * | 2018-09-28 | 2021-06-29 | Dow Global Technologies Llc | system for training a hybrid machine learning classifier, computer implemented method, article of manufacture, and computing device. |
WO2020079933A1 (en) * | 2018-10-18 | 2020-04-23 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | Information processing device, information processing method, and program |
JP7311310B2 (en) * | 2018-10-18 | 2023-07-19 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | Information processing device, information processing method and program |
KR102140385B1 (en) * | 2018-12-31 | 2020-07-31 | (주)유아이엠디 | Cell-zone labeling apparatus and cell-zone detecting system including the same apparatus |
US20220139527A1 (en) * | 2019-03-01 | 2022-05-05 | Agricam Aktiebolag | Method, device and system for detection of micro organisms |
WO2020261555A1 (en) * | 2019-06-28 | 2020-12-30 | オリンパス株式会社 | Image generating system and image generating method |
US20230351602A1 (en) * | 2020-08-17 | 2023-11-02 | Thrive Bioscience, Inc. | Cell segmentation image processing methods |
CN112651305B (en) * | 2020-12-10 | 2022-08-30 | 哈尔滨师范大学 | Method for identifying microorganism species |
EP4341373A2 (en) * | 2021-05-20 | 2024-03-27 | Clara Foods Co. | Systems for end-to-end optimization of precision fermentation-produced animal proteins in food applications |
WO2024091807A1 (en) * | 2022-10-24 | 2024-05-02 | Charles River Laboratories, Inc. | Automated enumeration system |
CN117253555A (en) * | 2023-11-17 | 2023-12-19 | 山东阜丰发酵有限公司 | Method for improving xanthan gum fermentation process and operation system thereof |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6690817B1 (en) * | 1993-08-18 | 2004-02-10 | Applied Spectral Imaging Ltd. | Spectral bio-imaging data for cell classification using internal reference |
US20040253660A1 (en) * | 2003-06-12 | 2004-12-16 | Giles Scientific, Inc. | Automated microbiological testing apparatus and method |
US7711174B2 (en) * | 2004-05-13 | 2010-05-04 | The Charles Stark Draper Laboratory, Inc. | Methods and systems for imaging cells |
US7958063B2 (en) * | 2004-11-11 | 2011-06-07 | Trustees Of Columbia University In The City Of New York | Methods and systems for identifying and localizing objects based on features of the objects that are mapped to a vector |
CN101916359B (en) * | 2005-01-27 | 2012-06-20 | 剑桥研究和仪器设备股份有限公司 | Methods and apparatus for classifying different parts of a sample into respective classes |
JP4999163B2 (en) * | 2006-04-17 | 2012-08-15 | 富士フイルム株式会社 | Image processing method, apparatus, and program |
US8278057B2 (en) * | 2007-09-14 | 2012-10-02 | Nestec S.A. | Addressable antibody arrays and methods of use |
JP4891197B2 (en) * | 2007-11-01 | 2012-03-07 | キヤノン株式会社 | Image processing apparatus and image processing method |
CN103080331B (en) * | 2010-06-23 | 2014-07-09 | 株式会社N-Tech | Method for detecting microorganisms, device for detecting microorganisms and program |
-
2012
- 2012-03-02 WO PCT/AU2012/000211 patent/WO2012119191A1/en active Application Filing
- 2012-03-02 EP EP12754708.1A patent/EP2681715B1/en active Active
- 2012-03-02 KR KR1020137026188A patent/KR101989202B1/en active IP Right Grant
- 2012-03-02 AU AU2012225196A patent/AU2012225196B2/en active Active
- 2012-03-02 ES ES12754708T patent/ES2928897T3/en active Active
- 2012-03-02 US US14/002,722 patent/US9292729B2/en active Active
- 2012-03-02 CN CN201280021787.9A patent/CN103518224B/en active Active
- 2012-03-02 JP JP2013555707A patent/JP5997185B2/en active Active
Non-Patent Citations (8)
Title |
---|
CIESIELSKI ET AL.: "A Comparison of Image Processing Techniques and Neural Networks for an Automated Visual Inspection Problem", PROCEEDINGS OF THE 5TH JOINT AUSTRALIAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 1992, HOBART, TASMANIA, pages 147 - 152, XP000566278 * |
JEROME FRIEDMANTREVOR HASTIEROBERT TIBSHIRANI: "Additive logistic regression: a statistical view of boosting", ANNALS OF STATISTICS, vol. 28, 1998, pages 2000 |
K. HIRAKAWAT. W. PARKS: "Adaptive homogeneity-directed demosaicing algorithm", ICIP, 2003 |
PAUL VIOLAMICHAEL JONES: "Robust real-time object detection", SECOND INTERNATIONAL WORKSHOP ON STATISTICAL AND COMPUTATIONAL THEORIES OF VISION, 13 July 2001 (2001-07-13) |
See also references of EP2681715A4 |
YOAV FREUNDROBERT E. SCHAPIRE: "A decision-theoretic generalization of on-line learning and an application to boosting", J. COMPUT. SYST. SCI., vol. 55, August 1997 (1997-08-01), pages 119 - 139 |
YURI BOYKOVOLGA VEKSLERRAMIN ZABIH: "Fast approximate energy minimization via graph cuts", IEEE TRANS, vol. 23, November 2001 (2001-11-01), pages 1222 - 1239 |
YURI BOYKOVOLGA VEKSLERRAMIN ZABIH: "Fast approximate energy minimization via graph cuts", IEEE TRANS. PATTERN ANAL. MACH. INTELL., vol. 23, November 2001 (2001-11-01), pages 1222 - 1239 |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2824921A3 (en) * | 2013-07-08 | 2015-04-15 | Samsung Display Co., Ltd. | Image and video in mosaic formats |
US9549161B2 (en) | 2013-07-08 | 2017-01-17 | Samsung Display Co., Ltd. | Image and video in mosaic formats |
CN107580715A (en) * | 2015-04-23 | 2018-01-12 | Bd科斯特公司 | Method and system for automatic counting microbial colonies |
US11341648B2 (en) | 2015-04-23 | 2022-05-24 | Timothy M. Wiles | Colony contrast gathering |
US11669971B2 (en) | 2015-04-23 | 2023-06-06 | Bd Kiestra B.V. | Colony contrast gathering |
US11674116B2 (en) | 2015-04-23 | 2023-06-13 | Bd Kiestra B.V. | Method and system for automated microbial colony counting from streaked sample on plated media |
EP3147893A1 (en) * | 2015-09-24 | 2017-03-29 | Dolby Laboratories Licensing Corp. | Light field simulation techniques for dual modulation |
EP3848472A3 (en) * | 2020-01-13 | 2021-12-15 | Airamatrix Private Limited | Methods and systems for automated counting and classifying microorganisms |
Also Published As
Publication number | Publication date |
---|---|
EP2681715A1 (en) | 2014-01-08 |
JP5997185B2 (en) | 2016-09-28 |
AU2012225196A1 (en) | 2013-03-21 |
US20140219538A1 (en) | 2014-08-07 |
CN103518224A (en) | 2014-01-15 |
ES2928897T3 (en) | 2022-11-23 |
JP2014515128A (en) | 2014-06-26 |
CN103518224B (en) | 2017-05-17 |
US9292729B2 (en) | 2016-03-22 |
KR101989202B1 (en) | 2019-07-05 |
KR20140045923A (en) | 2014-04-17 |
AU2012225196B2 (en) | 2015-05-14 |
EP2681715B1 (en) | 2022-07-13 |
EP2681715A4 (en) | 2014-10-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2681715B1 (en) | Method and software for analysing microbial growth | |
US11341648B2 (en) | Colony contrast gathering | |
AU2021202750B2 (en) | Method and system for automatically counting microbial colonies | |
US12039795B2 (en) | System and method for monitoring bacterial growth of bacterial colonies and predicting colony biomass |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 12754708 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2012225196 Country of ref document: AU Date of ref document: 20120302 Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2013555707 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 20137026188 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012754708 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14002722 Country of ref document: US |