US12175725B2 - Image processing device, image viewing system, image processing program, and image processing method - Google Patents
Image processing device, image viewing system, image processing program, and image processing method Download PDFInfo
- Publication number
- US12175725B2 US12175725B2 US17/326,853 US202117326853A US12175725B2 US 12175725 B2 US12175725 B2 US 12175725B2 US 202117326853 A US202117326853 A US 202117326853A US 12175725 B2 US12175725 B2 US 12175725B2
- Authority
- US
- United States
- Prior art keywords
- color
- image data
- tag
- value
- extracted
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/56—Extraction of image or video features relating to colour
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/51—Indexing; Data structures therefor; Storage structures
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/53—Querying
- G06F16/538—Presentation of query results
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/583—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/5838—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using colour
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/5866—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, manually generated location and time information
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/90—Determination of colour characteristics
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/50—Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/19—Recognition using electronic means
- G06V30/19007—Matching; Proximity measures
- G06V30/19073—Comparing statistics of pixel or of feature values, e.g. histogram matching
Definitions
- the present disclosure relates to an image processing device that extracts colors used in pixels in image data, and particularly to improvements in posting image data to a social networking service (SNS).
- SNS social networking service
- An SNS is used for various purposes from personal interests to corporate marketing and government public relations activities.
- Service users who use the SNS make a text search and an image search in search for image data they want to see.
- the number of times of image data display on the service user browser due to such search is referred to as “impressions”.
- the number of times of positive actions taken by the service users, such as gazing and enlarging image data, accessing the contributor's page, and clicking the “Like!” icon is referred to as “engagement”.
- engagement Increasing the impressions and engagement is one of the keys to information dissemination.
- a contributor who wants to disseminate information creates a text string expressing the characteristics of image to be posted, and uploads the image data together with this text string to the server of the SNS service to improve the impressions and engagement.
- Patent Literature 1 JP 2001-166908 A
- Patent Literature 2 JP 2019-159526 A
- a color selected from the palette is specified as a search condition.
- the text string indicating the correct color name listed in the color samples needs to be uploaded together with the image data to the service of the SNS service.
- Traditional Japanese and Western colors are classified variously. Thus, there is no guarantee that ordinary service users who have no knowledge of clothing, design, and art will list the color name in such classifications. If a service user creates a text string indicating a color name and uploads image data to the SNS server with incorrect knowledge, there is no improvement in the impressions and engagement. In addition, there is no guarantee that a service user who performs searching selects a color correctly in accordance with the classification of traditional Japanese and Western colors. As a result, there is no improvement in the impressions and engagement for the posted image data due to the synergistic effect with the mistake by the person who posted the image data and the mistake by the service user who performed searching.
- An object of the present disclosure is to provide an image processing device capable of allowing many service users to view image data in search for the image data with a search condition based on color selection from a palette in an image viewing system.
- an image processing device that assists searching for image data in an image viewing service
- the image processing device reflecting one aspect of the present invention comprises: a recognizer that recognizes a part representing a target object from the image data; an extractor that extracts a color used in pixels in the part recognized; a determiner that determines a color that is not extracted from the part but is likely to be used in searching by a service user, based on a search tendency in searching by the service user; and an adder that creates a first tag indicating the color extracted by the extractor and a second tag indicating the color determined by the determiner and adds the first tag and the second tag as search tags to the image data.
- FIG. 1 illustrates an image viewing system including an image processing device according to the present disclosure
- FIGS. 2 A and 2 B exemplarily illustrate a search screen that enables color selection by clicking a pull-down menu and pulling out a palette
- FIG. 3 exemplarily illustrates a search result screen displayed on a terminal
- FIG. 4 illustrates the control system of the image processing device
- FIG. 5 exemplarily illustrates color sample data in which the KGB values of an image in association with a color name and a color code
- FIG. 6 exemplarily illustrates pieces of training data corresponding one-to-one to a plurality of types for a subject
- FIG. 7 illustrates histogram of oriented gradients (HOG) features
- FIG. 8 is a scatter diagram with a plurality of features plotted
- FIG. 9 exemplarily illustrates a setting screen displayed on a display device in the process of training a type for a subject
- FIG. 10 exemplarily illustrates an extended color table
- FIG. 11 is a flowchart illustrating the main routine of a tag creation procedure
- FIG. 12 A is a flowchart of a procedure for object separation
- FIG. 12 B is a flowchart of a procedure for extended color determination
- FIG. 13 A illustrates the process of edge extraction
- FIG. 13 B illustrates the process of hue separation
- FIG. 14 illustrates the process of determination of an extracted color and an extended color based on an object
- FIG. 15 is a flowchart of a procedure for extended color determination
- FIG. 16 illustrates the process of determination of an extended color based on a sphere
- FIG. 17 is a flowchart of a procedure for extended color determination
- FIG. 18 illustrates the process of determination of an extended color based on a cylinder
- FIG. 19 is a flowchart of a procedure for extended color determination
- FIG. 20 is illustrates the process of determination an extended color with emphasized value and saturation
- FIG. 21 is a flowchart of a procedure for tag generation according to the distance between color coordinates
- FIGS. 22 A and 22 B illustrate the process of switching between creation of a single tag and creation of a plurality of tags according to the distance between color coordinates
- FIGS. 23 A to 23 C illustrate an expression of color space coordinates with serial numbers.
- FIG. 1 illustrates an image viewing system 1 including an image processing device 2000 according to the present disclosure.
- the image viewing system 1 includes a multifunction peripheral (MFP) 1000 ; terminals 1011 , 1012 , 1013 , and 1014 ; the image processing device 2000 ; and a social networking service (SNS) server 3000 (including a storage 3001 and a search engine 3002 ).
- MFP multifunction peripheral
- SNS social networking service
- the MFP 1000 draws out documents one by one from the bundle of documents placed on the placement table of an automatic document feeder 1006 integrally formed with a document retaining cover 1005 , and reads the documents with a scanner 1004 in response to an operation on a touch panel display 1001 to acquire image data corresponding to each document.
- Image data represents an image with an RGB color mixing system.
- Image data is transferred and stored in a file format such as JPEG, TIFF, PDF, or compact PDF.
- RGB values with 24-bit information are assigned to each pixel.
- the RGB values represent various colors in additive mixing of a value of R in 256 gradation levels (0 to 255), a value of G in 256 gradation levels (0 to 255), and a value of B in 256 gradation levels (0 to 255).
- Such a value of R, a value of B, and a value of G represent three-dimensional coordinates in the RGB color space in the RGB color mixing system (typically referred to as RGB color space).
- the image processing device 2000 takes in the image data read by the MFP 1000 and the image data stored in a file server 1100 , adds an appropriate tag to each piece of image data, and uploads the tagged image data to the storage of the SNS server 3000 .
- a tag is a text string that is associated with an image data file, and transferred and stored together with the image data file.
- the tag has various types such as a tag included in the attribute information or the file management information of the image data file, a tag stored in the link information of the image data file, and a tag stored in the same directory as the image data file.
- the number of characters is limited for posting a text string at one time (for example, within 200 characters or less), and a person who is to post an image (hereinafter referred to as contributor) within the limitation range can add two or more tags to the image data.
- the SNS server 3000 receives posting of image data through a public network 3010 , and provides a viewing service for the posted image data to terminals 1011 , 1012 , 1013 , and 1014 .
- screen data including various screens such as a login screen, a start screen, a search-condition setting screen, and a search-result display screen is created and transmitted to the terminals 1011 to 1014 operated by the SNS service users.
- FIG. 2 A exemplarily illustrates a search-condition setting screen displayed in the browser.
- clicking the pull-down menu and pulling out a palette 1011 P enables color selection.
- at least one text string (momo-iro in the figure) indicating the color name selected from the palette is input into a search field 1011 B.
- a search command with “momo-iro” as a keyword is output to the SNS server 3000 .
- the storage 3001 includes a plurality of network drives under the authority of the administrator of the image viewing system 1 .
- the storage 3001 has a directory for each account. Each time a contributor who has a corresponding account makes a post, a subdirectory is created in the directory for the account.
- the subdirectory is a posting area corresponding to the first posting by the contributor, and the image data and the text string uploaded by the contributor are collectively stored in this posting area.
- the search engine 3002 is one of the service providing programs launched by the SNS server 3000 .
- a search command including a search condition is transmitted from any of the browsers of terminals 1011 to 1014 , the search engine 3002 searches the storage 3001 for any image data with a tag matching the search condition.
- an AND condition or an OR condition for at least two colors can be specified, and under the AND condition, image data posted with at least two tags each indicating a different color is hit. Under the OR condition, image data posted with any of at least two tags each indicating a different color is hit.
- FIG. 3 exemplarily illustrates a search result screen 1011 F.
- the search result screen has a plurality of reduced images 1021 , 1022 , 1023 , 1024 , 1025 , and . . . of the image data matching a search condition disposed and listed.
- the clicked reduced image is enlarged and displayed.
- the search engine 3002 also performs a listing function in response to clicking a hashtag.
- the hashtag is a character string with a hash mark “#” prepended to a sentence.
- the SNS server 3000 reads a plurality of pieces of image data with the same hashtag from the storage 3001 and causes the browsers of terminals 1011 to 1014 to display the plurality of pieced of image data.
- FIG. 4 illustrates the control system of the image processing device 2000 .
- the image processing device 2000 includes a storage 201 , a random access memory (RAM) 202 , a central processing unit (CPU) 203 , a non-volatile memory 204 , a network interface card (NIC) 205 , and a peripheral port 206 .
- the image processing device 2000 performs input and output of image data through the NIC.
- the storage 201 serves as a hard disk drive (HDD) or a solid state drive (SSD) on which an operating system (OS) 201 S and a tagging application 201 A are installed.
- the OS 201 S and the tagging application 201 A installed in the storage 201 are loaded on the RAM 202 .
- HDD hard disk drive
- SSD solid state drive
- the CPU 203 reads an execution code included in the OS 201 S or the tagging application 201 A and decodes the execution code for execution.
- the non-volatile memory 204 stores various settings for operation of the tagging application 201 A.
- the NIC 205 performs input and output of image data between the NIC 205 , the MFP 1000 , and the SNS server 3000 . In the input and output, image data acquired by reading with the automatic document feeder 1006 of the MFP 1000 is taken in and written into the storage 201 .
- the CPU 203 also uploads the image data written in the storage 201 to the SNS server 3000 .
- Color sample data is stored in the non-volatile memory 204 .
- color sample data is data in which the RGB values of image are associated with the corresponding color name and color code.
- a color name is a text string defined as a color name in a color sample of a type such as primary color, Japanese color, Western color, web216, pastel color, vivid color, monotone, or metro color.
- a color code is a text string indicating a 3-byte hexadecimal value. A tag added to image data is created with the text string and the color code listed in this color sample.
- the tagging application 201 A installed in the storage 201 performs image processing on the image data read in the RAM 202 , and creates a color tag to be added to the image data, with a color name or color code listed in the color sample data.
- a color tag is a hash tag that enables search for image data by color selection from the palette, and has a structure in which a hash mark is prepended to a text string indicating a color name and a color code.
- Pre-processing to be performed prior to tag creation includes supervised learning of a subject shown in image data.
- Such supervised training is achieved by interactive operation with a display device 2001 , a character device 2002 , and a pointing device 2003 .
- the recognition dictionary includes pieces of training data corresponding one-to-one to a plurality of types for a subject.
- the example in FIG. 6 has a format in which pieces of data of discriminant functions F 1 , F 2 , F 3 , F 4 , F 5 , and . . . that define the hyperplane of a hard margin support vector machine (SVM) are in one-to-one association with types of a real object such as rose, pomegranate, azalea, or otome tsubaki.
- SVM hard margin support vector machine
- HOG histogram of oriented gradients
- the HOG features in the figure are obtained by aggregation of the number of line segments shown in 8 ⁇ 8 pixels for nine gradient directions (0°, 20°, 40°, 60°, 80°, 100°, 120°, 140°, and 160°). Extraction of features from a large number samples of image data gives the scatter diagram in FIG. 8 In the scatter diagram in FIG. 8 , features 2011 , 2012 , 2013 , 2014 , and 2015 acquired from the plurality of samples of the image data and features 2021 , 2022 , 2023 , 2024 , and 2025 are plotted on a two-dimensional plane.
- the discriminant functions in the training data illustrated in FIG. 6 define a hyperplane 2030 that separates the features plotted in the scatter diagram as above into two point clouds 2010 and 2020 .
- the hyperplane 2030 in FIG. 8 defined by the discriminant functions separates the features extracted from the image data read by the automatic document feeder 1006 and the scanner 1004 into the point could 2010 including the features 2011 , 2012 , 2013 , 2014 , and 2015 acquired from the plurality of samples of image data and the point group 2020 including the features 2021 , 2022 , 2023 , 2024 , and 2025 that do not represent the corresponding types. Such separation enables recognition what type the image data read by the scanner 1004 has.
- FIG. 9 exemplarily illustrates a setting screen displayed on the display device 2001 in the process of training a type for a subject.
- a sample of image data to be feature-extracted is arranged in a window 400 .
- An input field 401 for reception of input of characters included in a text string is arranged adjacent to the window 400 .
- the text string input in the input field 401 is regarded as a type of the target object appearing in the sample arranged on the left and the text string is stored in the non-volatile memory 204 .
- the setting screen in FIG. 9 is characteristic in that a palette 402 is arranged below the input field 401 and any of a plurality of colors listed on the palette 402 can be specified.
- a table in which the color as an extended color is in association with a type for a subject is written into the non-volatile memory 204 .
- the extended color table has a format in which extended colors are in association with types of a real object such as rose, pomegranate, azalea, or otome tsubaki. Most of the ordinary persons without insight into, design, art, and the like remember the color of a subject in correlation to the impression of the subject.
- the memory with such association is due to a biased view, for example, the setting sun is red or the apple is red (pseudo-memory).
- search relying on such pseudo memory no image with the tag indicating an extracted color is hit at all.
- the apple or the setting sun is perceived as “red,” they are not drawn in red pixels in images acquired by practically shooting the apple or the sunset. This is because no red pixel is extracted from the images of the apple or the setting sun and even if red pixels are extracted, the number is small.
- an extended color table is provided and a color associated with pseudo memory is selected as a tagging target.
- the tagging application 201 A creates a tag for each piece of image data read by the automatic document feeder 1006 and the scanner 1004 of the MFP 1000 .
- the flowchart of FIG. 11 illustrates this creation procedure.
- the variable max1 indicates the number of pieces of image data read by the automatic document feeder 1006 .
- the variable i indicates each of a plurality of pieces of image data read by the automatic document feeder 1006 , and the value varies from 1 to max1.
- the variable max2 indicates the total number of objects separated from the i-th image data.
- the variable j indicates each of a plurality of objects extracted from the i-th image data, and the value varies from 1 to max2.
- step S 101 the variable i is set to 1, and in step S 102 , at least one object is separated from the image data i.
- an object is a group of pixels representing a real object, and is a pixel group having some kind of unity, such as a plurality of pixels that is collected to achieve gradation in light and darkness, or that is unified with similar colors.
- the variable j is set to 1 (step S 103 ) to create the saturation histogram of the j-th object (object j) (step S 104 ).
- the RGB values of each pixel included in the object are converted into HSV values (consisting of a value of H as hue, a value of S as saturation, and a value of V as value).
- the number of appearances of pixels having the same saturation is counted.
- the saturation histogram indicating the appearance frequency of each saturation can be acquired.
- step S 106 performed is determination regarding whether the spread of the saturation histogram exceeds a subject reference.
- the subject reference here is a value empirically set for discrimination regarding whether an object is the subject acquired by shooting a real object, and is set to about 4 to 8 colors.
- the subject acquired by shooting a real object has complicated color tone and gradation due to the irradiation and reflection of ambient light.
- charts, illustrations, and the like created with a document creation application have no complicated color tone and gradation.
- the spread of the saturation histogram is compared with the subject reference to perform discrimination regarding whether the object is the subject acquired by shooting a real object.
- An object that is not a subject acquired by shooting and is similar to the subject (such as an object in a painting depicted by a natural painting method or the like) is also determined similarly to the subject. This is because any object represented by such a depiction also needs to be treated the same as the subject.
- step S 106 If the spread of the distribution exceeds the subject reference and a gradation is present in the saturation histogram (Yes in step S 106 ), an extended color is determined (step S 107 ). Then, a first color tag indicating an extracted color and a second color tag indicating the extended color are added to the image data i (step S 108 ).
- step S 109 a color tag indicating an extracted color is added to the image data i (step S 109 ).
- Step S 110 is determination regarding whether the loop continuation condition is satisfied. If the variable j indicating the object falls below the total number of objects max2 (Yes in step S 110 ), the variable j is incremented (step S 111 ), and the flow returns to step S 104 . As long as the variable j falls below the total number of objects max2, the repetition of steps S 104 to S 111 continues. As a result, the individual objects on the page j are subjected to processing.
- Step S 112 is determination regarding whether the loop continuation conditions of steps S 102 to S 111 are satisfied. If the variable i indicating the image data falls below the total number of pieces of image data max1 (Yes in step S 112 ), the variable i is incremented, and the flow returns to step S 102 (step S 113 ). As long as the variable i falls below the total number of pieces of image data max1, the repetition of steps S 102 to S 113 continues. As a result, the individual pieces of image data are subjected to processing. When the variable i reaches the total number of pieces of image data max1 and the processing on all the pieces of image data is completed, the result of the determination in step S 112 defining the end condition becomes No, and the processing of this flowchart ends.
- Step S 102 is a subroutine.
- FIG. 12 A is a flowchart of a procedure for object separation.
- a differential filter is applied to extract an edge in step S 121 .
- the image (differential image) acquired by applying the differential filter has a pattern in which fragmentary bright lines are arranged on a black background.
- the fragmented bright lines indicate portions where the gradation and hue are discontinuous and the RGB values vary rapidly.
- a pixel sequence appearing as a bright line in the differential image is referred to as an “edge” and represents the contour of a target object in fragments.
- step S 122 an attempt is made to separate a target object with the edge as a clue. Specifically, in step S 122 , among the pixels included in the edge, pixels adjacent to each other in eight directions including up, down, left, and right are connected, and a closed loop is extracted. This process will be described in detail.
- Any point on the edge is stored as a starting point, and performed is determination regarding whether or not the edge is adjacent to another edge in any of the eight directions. Even if a pixel-free area is present between the edges, when the pixel-free portion is 10% or less of the size of the edges, the pixel-free portion is ignored. The purpose is to give priority to the reproduction of the target object. A plurality of edges is traced while ignoring the gaps between the pixels, and performed is determination regarding whether or not the trajectories of the edges reach the starting point.
- the features are extracted from the object (step S 124 ). Then, the extracted features are applied to the training data acquired by machine learning to recognize a type for a subject appearing in the object (step S 125 ). Thereafter, a serial number is assigned to the object, and the object is registered in association with the coordinates in the coordinate system (step S 126 ). Then, at least one registered object is returned.
- FIG. 12 B illustrates the details of the procedure for extended color determination in step S 108 .
- an extended color is determined in consideration of the search tendency of SNS users. Specifically, service users who are SNS users tend to select, as a search condition, a color recognized as the color of a subject. Thus, performed is determination regarding whether an extended color corresponding to the recognized type is present in the extended color table (step S 131 ). If the extended color is present (Yes in step S 131 ), the extended color listed in the extended color table is returned to the main routine (step S 132 ). If no corresponding extended color is present in the extended color table (No in step S 131 ), the fact that no extended color is determined is returned to the main routine (step S 133 ).
- edges 511 , 512 , 513 , and 514 are extracted, and the area 1 of the areas 1 , 2 , 3 , 4 , and 5 surrounded by these continuous edges is defined as the object j.
- a histogram is created for the object j, it is shown as the histogram 521 in FIG. 14 .
- the color with the highest frequency 522 is the extracted color.
- the extended color table 314 indicates the colors of real objects such as rose, pomegranate, azalea, otome tsubaki, and sunset in association with one-to-one, as extended colors, the colors of the real objects recognized by the ordinary persons.
- a recognition dictionary if one (otome tsubaki) of the types listed in the extended color table is acquired, mono-iro is acquired as the extended color corresponding to the otome tsubaki.
- a hash mark “#” is prepended to each of the color names of these extracted color and extended color to create color tags. Together with the color tag, the image data is uploaded to the storage 3001 of the SNS server 3000 and provided for viewing by the SNS users of terminals 1011 to 1014 .
- an object as a pixel group representing a subject noticed by an SNS user is recognized with training data acquired in the process of machine learning, and a first color tag indicating a color used in the pixels in the recognized object is added to image data to be uploaded.
- a first color tag indicating a color used in the pixels in the recognized object is added to image data to be uploaded.
- an extended color is determined on the basis of a type of a separated object, and a second color tag to be added to the image data is created on the basis of the extended color.
- the extended color corresponding to the type is read from the extended color table to determine the extended color of the object.
- an extended color is determined on the basis of the search tendency of SNS users that a color close to the color in the gradation of a subject is selected as a search condition.
- FIG. 15 is a flowchart of the procedure for extended color determination, which is performed instead of the procedure in FIG. 12 B .
- the extended color determined in step S 141 of FIG. 15 is as shown in FIG. 16 . That is, in the RGB color space, the extended color has the coordinates on the surface of the sphere 601 with a radius r around the coordinates (R E , G E , B E ) of the extracted color C E , or the coordinates inside the sphere 601 .
- the color set located on the surface of or inside the sphere is likely to contain the color of the pixels included in the object.
- the processing of step S 142 is performed. Specifically, a difference set is obtained from excluding the set elements of the color set included in the object from the color set that satisfies Mathematical Expression 1 and that is located on the surface of or inside the sphere. Then, the extended color is determined on the basis of the difference set.
- the difference set can be created with the difference method of a set type and a frozenset type in the programming language Python.
- the extended color obtained as above is returned to the main routine (step S 143 ).
- the spread D of the saturation histogram indicates the number of gradation levels of the real object as the subject.
- the number of gradation levels is regarded as the diameter of a sphere and an extended color is determined on the basis of the color set located on the surface of or inside the sphere.
- an extended color is determined on the basis of the spread of the gradation shown in the saturation histogram and an extracted color.
- an extended color is determined on the basis of the search tendency of SNS users that a color close to a plurality of colors with a large frequency in a part representing a subject is selected as a search condition.
- FIG. 17 is a flowchart of the procedure for extended color determination, which is performed instead of the procedure in FIG. 12 B .
- a first convergent color and a second convergent color are obtained as colors each having the color frequency as a dominant reference (step S 151 ).
- a dominant reference value is a value of a predetermined ratio (for example, 10%) to the maximum frequency shown in a saturation histogram.
- a pixel color with the frequency of top 10% as above is a color of a dominant color scheme in an object, and is used to clarify the color scheme of a subject. The frequencies shown in the saturation histogram converge to the dominant reference value, so that colors with a small frequency are ignored in the extended color determination.
- the extended color determined in step S 153 above is as shown in FIG. 18 .
- the color set located on the surface of the cylinder 702 with the convergent straight line 701 as the central axis and a radius ⁇ d is likely to include a color of pixels of the object.
- step S 154 the processing of step S 154 is performed. Specifically, a difference set is obtained by excluding the set elements of the color set included in the object from the color set that satisfies i) and ii) and that is located on the surface of the cylinder, the extended color is determined on the basis of the difference set, and then the determined extended color is returned to the main routine (step S 155 ).
- the radius ⁇ d has a value to some extent that separates two to five colors in the RGB color space, and the color set occupying the surface of the cylinder that occupies the diameter is determined as the extended color.
- a tag indicating the color almost close to the color with a high frequency in the saturation histogram can be added to image data.
- the surface of the cylinder with the convergent straight line as the central axis is defined as the coordinate of the extended color.
- a three-dimensional straight line is defined on the basis of the color space coordinates of a color that converges to a certain frequency, and an extended color is determined on the basis of the range occupied by the surface of a cylinder with the three-dimensional straight line as an axis.
- a color close to a dominant color can be selected as the extended color.
- an extended color is determined on the basis of the distance from a straight line passing through the coordinates of a color shown in a histogram in the color space.
- an extended color is determined on the basis of the search tendency of SNS users that a color higher in saturation and/or value than a color used in a part representing a subject is selected as a search condition.
- FIG. 19 is a flowchart of the procedure for extended color determination, which is performed instead of the procedure in FIG. 12 B .
- the colors (RGB values) of a separated object are converted into HSV values and the HSV values are plotted into the HSV color space (step S 161 ).
- a rectangular solid passing though the coordinates of one with the largest value of H, the coordinates of one with the largest value of S, and the coordinates of one with the largest value of V among the HSV values plotted in the HSV color space is defined inside the HSV color space (step S 162 ).
- the defined rectangular solid is as illustrated in FIG. 20 . It is assumed that the RGB values of the pixels of the object are plotted at the coordinates 801 , 802 , 803 , and 804 to 807 illustrated in FIG. 20 in the HSV color space.
- the value of H at the coordinates 807 , the value of S at the coordinates 806 , and the value of V at the coordinates 805 are the largest, and thus the rectangular solid 810 passing through these coordinates is defined.
- the rectangular solid defined as above indicates where in the HSV color space the colors of the separated object are distributed, that is, the color distribution range of the pixels of the object in the HSV color space.
- faces of the rectangular solid indicate a boundary face of value and a boundary face of saturation.
- an extended color is determined on the basis of the color set located outside the rectangular solid (step S 163 ).
- the extended color is determined on the basis of the color set spaced apart from the faces of the rectangular solid (value boundary face, saturation boundary surface) by two to five colors.
- a second color tag representing a color that emphasizes the color saturation and value of a real object as the subject is added to image data.
- determined are the extended color 811 located outside the boundary line of the value and the extended colors 812 and 813 located outside the boundary lines of the value and the saturation.
- an extended color and an extracted color are represented with different color tags.
- performed is determination regarding whether an extracted color and an extended color are represented with different color tags or the extracted color and the extended color represented with a single color tag, in accordance with the distance between the coordinates of the extended color and the coordinates of the extracted color in the RGB color space.
- FIG. 21 illustrates the changed part of the main routine.
- steps S 171 and S 172 illustrated in this figure are added.
- Step S 171 is determination regarding whether the distance between the coordinates of an extracted color and the coordinates of an extended color in the RGB color space falls below a predetermined color vision threshold D t .
- the color vision threshold D t is the interval between at least two color space coordinates recognized as colors that are similar colors. Even if the color space coordinates are different, the human eyes may recognize a plurality of colors as the same colors or similar colors. In the color vision characteristics of the human eyes, a boundary value recognized as the same color or a similar color is set in advance as a color vision threshold, and the distance between the coordinates of the extracted color and the coordinates of the extended color is compared with the color vision threshold.
- step S 171 When the intercolor distance is not less than the color vision threshold D T , the result of the determination in step S 171 becomes No, and step S 108 in FIG. 11 is performed.
- step S 172 both the extracted color and the extended color are represented with a single color tag.
- a single color tag includes one indicating an intermediate color between the extracted color and the extended color, and one indicating a color system including the extracted color and the extended color.
- one indicating the extracted color or one indicating the extended color may be used.
- step S 171 a first color tag indicating the extracted color and a second color tag indicating the extended color of the above pair 911 are created (step S 108 ).
- the pair 912 of an extracted color and an extended color in FIG. 22 B is a target.
- the distance D (C E , C X ) between the coordinates C E of the extracted color and the coordinates C X of the extended color is the color vision threshold D T or less, and thus the result of the determination in step S 171 becomes Yes.
- the coordinates (R E , G E , B E ) of the extracted color C E and the coordinates (R X , G X , B X ) of the extended color C X in the above pair 912 are added together and divided by 2 to calculate the RGB values of the intermediate color of the extracted color and the extended color (step S 172 ).
- the color name of the color corresponding to the RGB values calculated in such a manner is used as a color tag to be added to image data.
- the color of each pixel of an object is expressed with a value of R in 256 gradation levels, a value of G in 256 gradation levels, and a value of B in 256 gradation levels however, the expression is not limited to this example.
- the color of each pixel of an object may be expressed with a serial number of the subspace inside the RUB color space.
- an edge is extracted from image data and an object including a subject is separated, but the present invention is not limited to this example.
- An object may be acquired by separating an area occupied by continuous pixels having similar colors. Specifically, as illustrated in FIG. 13 B , the RGB values of the pixels of the object are converted into HSV values to specify a pixel group in which pixels having the same hue are continuous. Such hue separation results in image separation according to a rough classification such as reddish color and greenish color.
- the separated pixel group is a series of pixels having substantially the same hue and has similar colors. Separation of a pixel group having such a similar color can provide an object similar to that acquired by the above edge extraction.
- a color having the maximum frequency in a saturation histogram is determined as an extracted color, but this is merely an example.
- a color at not less than a predetermined rank in the ranking of the colors used in the pixels in an object such as an occupancy of 10% or more in the entirety, may be determined as an extracted color. This is because that although it is recognized that “apples are aka”, the pixels occupying a large area in an actual apple are sekkoku and not necessarily aka and it is likely that a characteristic color representing a target object is extracted if some colors at higher ranks are extracted.
- the image viewing system may also implement an image posting service, a moving-image posting service, a personal introduction service, an online shopping service, a short-sentence posting service, a bulletin board service, a store-information providing service, and a purchase-and-sale and exchange service for used goods.
- the number of characters is limited for a text string that can be posted with image data, but the limitation is not limited to this example. Instead of the limitation for the number of characters, the number of color tags may be limited in addition to the limitation of for the number of characters.
- image data acquired by scanning with the MFP 1000 is uploaded to the SNS server, but uploading is not limited to this example. Image data shot by a smartphone, tablet, or camera may be uploaded to the SNS server.
- the tagging application 201 A is a pre-installed type application pre-installed in the image processing device 2000 , but the tagging application 201 A is not limited to this example.
- the tagging application 201 A may be recorded on a portable recording medium and may be sold as a software package, or may be uploaded to an application distribution server and may be distributed by the application distribution server.
- the image processing device 2000 is achieved with a computer device installed in the premises of the business establishment.
- the image processing device 2000 may be achieved with a cloud server.
- the cloud server has an operating system (OS) and a tagging application 201 A installed thereon.
- OS operating system
- a guest OS starts up and the tagging application 201 A is launched on the guest OS.
- the image processing device 2000 is a device independent of the MFP 1000 , but the image processing device 2000 is not limited to this example.
- the MFP 1000 and the image processing device 2000 may be integrally formed with one apparatus.
- the image processing device 2000 performs tagging and uploads image data, but the image processing device 2000 is not limited to this example.
- the image processing device 2000 may perform tagging and another device may upload image data.
- a tagging application 201 A may be installed and used on the terminals 1011 to 1014 operated by SNS users. This is because SNS users typically post images. Furthermore, a tagging application 201 A may be provided on an SNS server.
- a difference set is obtained to determine an extended color on the basis of the difference set, but the determination is not limited this example. If a color used as the color of the pixels of an object and not extracted as an extracted color is present on the surface of or inside the sphere of the second embodiment, the color may be determined as the extended color. Similarly, if a color used as the color of the pixels of the object and not extracted as an extracted color is present on the surface of the cylinder of the third embodiment, the color may be determined as the extended color.
- a tag indicating an extended color determined on the basis of the type of a pixel object a tag indicating an extended color determined on the basis of the number of gradation levels of the pixel object, a tag indicating an extended color determined on the basis of a first convergent color and a second convergent color, or a tag indicating an extended color determined on the basis of a saturated value of saturation or value may be added to image data.
- Such tagging enables broad-ranging color expression for the pixel object appearing in image data, resulting in an increase in impressions for uploading image data read by the seamier 1004 to the SNS server 3000 .
- HSV values may be described in the field of each color in the color sample data. Such description advantageously facilitates conversion of the RGB values to the HSV values.
- a conversion formula may be stored in advance in the non-volatile memory 204 and the RGB values may be converted into the HSV values on the basis of the conversion formula.
- a color name and a color code are used as color tags, but a tag is not limited to this example.
- a tag with a hash mark “#” added to a text string indicating a noun phrase, an adjective phrase, or an adverbial phrase including a color name may be created.
- text strings such as “#akamiwoobita”, “#akai”, “#akaku”, and “#akappoi” may be listed in advance in the color sample data and the extended color table, and may be added the hash tag including these text strings and the hash marks to image data.
- the color names for the color tags are Japanese, but the language is not limited to this example.
- the color names of a plurality of languages such as English, Chinese, Russian, French, Spanish, and Arabic may be listed in the color sample data and the extended color table, and an extracted color and an extended color may be expressed with the color tags of the plurality of languages.
- a language selection may be received from a distributor and the color tag of the selected language may be added to image data.
- the present disclosure can be used in the industrial fields of various industries that transmit information with SNS servers, and is likely to be used the industrial fields of various industries such as office automation (OA) equipment and information equipment, and the industrial fields of various industries such as retailing, antique dealers, leasing, real estate, advertising, transportation, and publishing.
- OA office automation
- retailing antique dealers, leasing, real estate, advertising, transportation, and publishing.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Library & Information Science (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Software Systems (AREA)
- Probability & Statistics with Applications (AREA)
- Image Analysis (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Information Transfer Between Computers (AREA)
- Processing Or Creating Images (AREA)
Abstract
Description
|{right arrow over (OC x)}−{right arrow over (OC E)}|≤r [Mathematical Expression 1]
(a+pt−R X , b+qt−G X , c+rt−B X)·(p, q, r)=0 ii)
ii)
√{square root over ((a+pt−R X)2+(b+qt−G X)2+(c+rt−B X)2)}=Δd [Mathematical Expression 2]
Claims (18)
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2020101752A JP7532927B2 (en) | 2020-06-11 | 2020-06-11 | IMAGE PROCESSING APPARATUS, IMAGE VIEWING SYSTEM, IMAGE PROCESSING PROGRAM, AND IMAGE PROCESSING METHOD |
| JP2020-101752 | 2020-06-11 |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20220019617A1 US20220019617A1 (en) | 2022-01-20 |
| US12175725B2 true US12175725B2 (en) | 2024-12-24 |
Family
ID=79195599
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US17/326,853 Active 2043-06-10 US12175725B2 (en) | 2020-06-11 | 2021-05-21 | Image processing device, image viewing system, image processing program, and image processing method |
Country Status (2)
| Country | Link |
|---|---|
| US (1) | US12175725B2 (en) |
| JP (1) | JP7532927B2 (en) |
Families Citing this family (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| KR20230118219A (en) * | 2022-02-03 | 2023-08-11 | 삼성디스플레이 주식회사 | Method of detecting stain of display panel, method of compensating stain of display panel including the same and display apparatus performing the same |
Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2001166908A (en) | 1999-12-13 | 2001-06-22 | Canon Inc | Memory color processing device and print system |
| US20040267740A1 (en) * | 2000-10-30 | 2004-12-30 | Microsoft Corporation | Image retrieval systems and methods with semantic and feature based relevance feedback |
| JP2019159526A (en) | 2018-03-09 | 2019-09-19 | パイオニア株式会社 | Line detection device, line detection method, program and storage medium |
-
2020
- 2020-06-11 JP JP2020101752A patent/JP7532927B2/en active Active
-
2021
- 2021-05-21 US US17/326,853 patent/US12175725B2/en active Active
Patent Citations (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2001166908A (en) | 1999-12-13 | 2001-06-22 | Canon Inc | Memory color processing device and print system |
| US20040267740A1 (en) * | 2000-10-30 | 2004-12-30 | Microsoft Corporation | Image retrieval systems and methods with semantic and feature based relevance feedback |
| JP2019159526A (en) | 2018-03-09 | 2019-09-19 | パイオニア株式会社 | Line detection device, line detection method, program and storage medium |
Also Published As
| Publication number | Publication date |
|---|---|
| JP7532927B2 (en) | 2024-08-14 |
| US20220019617A1 (en) | 2022-01-20 |
| JP2021196761A (en) | 2021-12-27 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US11907322B2 (en) | Generating app or web pages via extracting interest from images | |
| US8988450B1 (en) | Color palette maps for color-aware search | |
| JP5259894B1 (en) | Color name determination device, color name determination method, information recording medium, and program | |
| JP5259893B1 (en) | Color determination device, color determination system, color determination method, information recording medium, and program | |
| US10043231B2 (en) | Methods and systems for detecting and recognizing text from images | |
| US8553045B2 (en) | System and method for image color transfer based on target concepts | |
| US8369616B2 (en) | Chromatic matching game | |
| US8532377B2 (en) | Image ranking based on abstract concepts | |
| US9082047B2 (en) | Learning beautiful and ugly visual attributes | |
| US8587604B1 (en) | Interactive color palettes for color-aware search | |
| Chen et al. | Tag-based image retrieval improved by augmented features and group-based refinement | |
| US7925650B2 (en) | Image management methods, image management systems, and articles of manufacture | |
| US20110078176A1 (en) | Image search apparatus and method | |
| CN106560809A (en) | Modifying At Least One Attribute Of Image With At Least One Attribute Extracted From Another Image | |
| JP2007164648A (en) | Similar image search device, similar image search method, program, and information recording medium | |
| CN101620680A (en) | Recognition and translation method of character image and device | |
| JP2011154687A (en) | Method and apparatus for navigating image data set, and program | |
| CN114283422B (en) | Handwriting font generation method and device, electronic equipment and storage medium | |
| US12175725B2 (en) | Image processing device, image viewing system, image processing program, and image processing method | |
| JP2020057348A (en) | System and method for identifying images containing identification documents | |
| WO2021129289A9 (en) | Image processing method and apparatus, electronic device, and storage medium | |
| US11010978B2 (en) | Method and system for generating augmented reality interactive content | |
| US9767579B2 (en) | Information processing apparatus, information processing method, and non-transitory computer readable medium | |
| KR20150096552A (en) | System and method for providing online photo gallery service by using photo album or photo frame | |
| US20160162752A1 (en) | Retrieval apparatus, retrieval method, and computer program product |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: KONICA MINOLTA, INC., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:YAMANAKA, TOMOO;REEL/FRAME:056314/0185 Effective date: 20210427 |
|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |