KR101606760B1 - Apparatus and Method of Transforming Emotion of Image based on Object in Image - Google Patents
Apparatus and Method of Transforming Emotion of Image based on Object in Image Download PDFInfo
- Publication number
- KR101606760B1 KR101606760B1 KR1020150104116A KR20150104116A KR101606760B1 KR 101606760 B1 KR101606760 B1 KR 101606760B1 KR 1020150104116 A KR1020150104116 A KR 1020150104116A KR 20150104116 A KR20150104116 A KR 20150104116A KR 101606760 B1 KR101606760 B1 KR 101606760B1
- Authority
- KR
- South Korea
- Prior art keywords
- information
- image
- target
- candidate
- emotion
- Prior art date
Links
Images
Classifications
-
- G06T7/0044—
-
- G06K9/6261—
-
- G06T7/0059—
-
- G06T7/408—
-
- H04N5/232—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/64—Circuits for processing colour signals
- H04N9/74—Circuits for processing colour signals for obtaining special effects
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Image Analysis (AREA)
Abstract
The present invention relates to an apparatus and a method for converting emotions appearing in an image.
The object of the present invention is to provide an image processing apparatus and a method and a system for processing an image by using information about an object as well as color information included in a photograph based on the fact that the emotion felt by a person is most affected by an object included in the photograph, To convert the image to better express the desired emotion. To this end, the image emotion conversion apparatus according to the present invention receives an emotion conversion target image, extracts a target object to be emotion transformed from the target image, and includes information about a position or an area in the target image of the target object An object recognition unit for recognizing the target object and generating semantic information, which is information on the meaning of the target object, a target emotion information indicating a target emotion to be transformed with respect to the target image, A candidate selector for comparing at least one candidate object and the target object based on the semantic information and emotion information and selecting the candidate object according to the comparison result; Using the candidate object information including the information, And converting the image signal of the target image within the target image corresponding to the target image to generate an image converted so that the target image expresses emotion corresponding to the target emotion information.
Description
The present invention relates to an apparatus and a method for converting emotions appearing in an image.
There have been various image conversion technologies that process image signals from various smart devices such as a smart phone and a tablet or a computer and give specific effects to the images. Especially, as the use of social network services has been expanded, applications have been developed in which each user can take a picture by using a smart device, download a picture from the web, give a desired image effect, and store or retransmit the picture. have.
As the image processing technology that gives a specific effect to the image, there are generally used methods such as adjusting the saturation and brightness of the image or using a filter for blurring or sharpening the image, and furthermore, There are also methods for converting images using preset image filters or using color palettes to reflect them. For example, if the user wishes to change the mood of a specific image more pleasantly or more unpleasantly, there are image emotion conversion methods of converting the image so that the color, lightness, etc. are changed by reflecting such emotion information.
However, in the existing methods of converting emotions, a method of converting only the overall color information of an image is used in order to express effects according to specific emotion information. In this regard, But there is a limit in that it can not.
Korean Patent Laid-Open Publication No. 2014-0037373 (2014. 03. 27)
The problem to be solved by the present invention is that the emotion felt by a person is most influenced by an object included in the photograph, so that not only the color information included in the photograph but also information about the object are used together So as to better represent the target emotion desired by the user, and an apparatus associated therewith.
According to one aspect of the present invention, there is provided an image-emotion conversion apparatus that receives an image to be emotionally transformed, extracts a target object to be emotion-transformed from the target image, An object recognition unit for generating target object information including information on the target object, information on the target object, information on the target object, and generating semantic information, which is information on the meaning of the target object, A candidate selector for receiving target emotion information, comparing at least one candidate object and the target object based on the semantic information and emotion information, and selecting the candidate object according to the comparison result; And the candidate object information of the candidate object And converting the image signal of the corresponding region in the target image to generate an image in which the target image is transformed to express emotion corresponding to the target emotion information.
Here, the object recognition unit may analyze a video signal of pixels included in the target image area corresponding to the extracted target object to extract a predetermined feature.
Wherein the image empowerment conversion device extracts emotion information of the target object by inputting the semantic information and the characteristic of the extracted target object into a previously learned emotion information classifier and extracting emotion information of the target object from the emotion information of the target object, And an emotion information extracting unit that determines emotion information of the target image.
Here, the object recognition unit performs image segmentation on the target image to divide it into a plurality of areas, selects a part of the divided areas, extracts the target area corresponding to the target object, The target object information including information about a position or an area in the target image of the target object.
Here, the object recognizer may determine which class of the predetermined object class the object object corresponds to, using the image signal of the target image or the extracted feature, and generate the semantic information according to the class of the determined object have.
Here, the object recognition unit may include an image segmentation unit that divides the target image into a plurality of regions by performing image segmentation, extracts the target image region corresponding to the target object by selecting a portion of the divided regions, A feature extracting unit that extracts a predetermined feature by analyzing an image signal of pixels included in the target image within the target image corresponding to the target object; And a semantic information generator for generating the semantic information according to the determined class of the determined object.
Here, the candidate object is stored in a candidate database, and the semantic information, which is information on the meaning of the candidate object for each candidate object, and the emotion information previously set for the candidate object are stored in the candidate database, Compares the semantic information and the emotion information of the candidate object stored in the candidate database with the semantic information and the target emotion information of the target object generated by the object recognition unit, The candidate objects corresponding to the candidate objects are selected.
Here, the candidate selector may calculate the similarity between the semantic information and the emotion information of the candidate object stored in the candidate database, the semantic information of the target object generated by the object recognition unit, and the target emotion information, And the candidate objects are selected based on a similarity.
Wherein the object recognition unit extracts the target image area corresponding to the target object and uses the extracted target area information as the position information of the target object to generate the target object information including the position information of the target object Wherein the candidate object information including positional information of the candidate object in the candidate object is stored in the candidate database for each candidate object, and the candidate selecting unit selects the semantic information of the candidate object stored in the candidate database, The candidate object information, the semantic information of the target object generated by the object recognition unit, the target emotion information, and the target object information, and selects the candidate object according to the comparison result. can do.
Wherein the object recognition unit extracts the target image area corresponding to the target object, uses the texture information of the extracted area as texture information of the target object, and extracts the target object information including the texture information of the target object The candidate object information including the texture information of the candidate object for each of the candidate objects is stored in the candidate database, and the candidate selecting unit selects the semantic information, the emotion information, and the semantic information of the candidate object stored in the candidate database, The candidate object information is compared with the semantic information, the target emotion information, and the target object information of the target object generated by the object recognition unit, and the candidate object is selected according to the comparison result have.
Wherein the candidate object information includes color information of the candidate object and the image transformation unit uses the color information included in the candidate object information of the candidate object selected by the candidate selection unit to generate color information corresponding to the target object And a color conversion unit for converting the image signal so that the color distribution of the image signal of the target image region matches the color information of the candidate object, thereby generating the converted image.
Wherein the candidate object information includes position information in the image of the candidate object and the image transform unit transforms the candidate object using the position information in the image included in the candidate object information of the candidate object selected by the candidate selecting unit, And a position conversion unit for performing an image transformation for moving the position of the target object in the target image so that the target object in the image corresponds to the in-image position information of the candidate object, thereby generating the transformed image .
Wherein the candidate object information includes texture information of the candidate object and the image transformation unit uses the texture information included in the candidate object information of the candidate object selected by the candidate selection unit, And a texture conversion unit for converting the image signal of the target image area corresponding to the target object so that the object corresponds to the texture information of the candidate object, and generating the converted image.
Wherein the image conversion unit searches the word database for an associated word corresponding to the semantic information of the target object, searches an image database for an image patch corresponding to the retrieved related word, And an object adding unit for generating the transformed image.
Wherein the object recognition unit extracts a predetermined feature including a color feature or a texture feature in an area in the target image corresponding to the target object, and the emotion information classifier used by the emotion information extractor includes: Wherein the classifier is configured to learn a plurality of learning data in which the semantic information, the characteristic, and the emotion information are set in advance, parameters of a classifying function for inputting the semantic information and the characteristic and outputting the emotion information, The information extracting unit may extract the emotion information of the target object by inputting the semantic information of the target object and the feature including the color feature or texture feature to the classification function of the emotion information classifier.
The emotion information classifier may be a classifier that learns the learning data using a linear regression model or a support vector-based regression model, and the parameter of the classification function is set.
According to another aspect of the present invention, there is provided an image emotion conversion method including receiving an emotion conversion target image, extracting a target object to be emotion-transformed from the target image, An object recognizing step of generating object object information including information on a position or an area of the target image and recognizing the target object and generating semantic information which is information on the meaning of the target object; A candidate selecting step of comparing the at least one candidate object and the target object based on the semantic information and the emotion information and selecting the candidate object according to a result of the comparison; The candidate object information of the selected candidate object, Converts the video signal of the area within the target image corresponding to the object, it may include an image converting step of generating the image the target image is converted to indicate the emotion corresponding to the target emotion information.
The object recognition step may further include extracting a predetermined feature by analyzing a video signal of pixels included in the target image area corresponding to the extracted target object.
Here, the image emotion conversion method may include inputting the semantic information and the characteristic of the extracted target object to a previously learned emotion information classifier, extracting emotion information of the target object, and based on the extracted emotion information of the target object And an emotion information extracting step of determining emotion information of the target image.
Here, the candidate object is stored in a candidate database, and the semantic information, which is information on the meaning of the candidate object for each candidate object, and the emotion information previously set for the candidate object are stored in the candidate database, Step compares the semantic information and the emotion information of the candidate object stored in the candidate database with the semantic information of the target object generated in the object recognition step and the target emotion information, And selects the candidate object corresponding to the object.
Wherein the candidate object information includes color information of the candidate object and the image transformation step uses the color information included in the candidate object information of the candidate object selected in the candidate selection step to correspond to the target object And a color conversion step of converting the image signal so that the color distribution of the image signal of the area within the target image matches the color information of the candidate object to generate the converted image.
The image conversion step may include searching an associated word corresponding to the semantic information of the target object in a word database, searching an image database for an image patch corresponding to the retrieved related word, And adding an object to the image to generate the transformed image.
According to another aspect of the present invention, an image emotion conversion program may be a computer program stored in a medium for executing the image emotion conversion method in combination with a computer.
According to the present invention, the object of the present invention is to provide an image-emotion conversion apparatus and method, in which the emotion felt by a person is most affected by an object included in the photograph, There is an effect that the image can be converted so as to better express the target emotion desired by the user by using the information about the object together.
FIG. 1 is a block diagram of a video image conversion apparatus according to an exemplary embodiment of the present invention. Referring to FIG.
2 is a block diagram of an image-emotion conversion apparatus further including an emotion information extracting unit.
3 is a detailed block diagram of the object recognition unit.
4 is a block diagram for explaining the operation of the candidate selector.
5 is a block diagram for explaining the operation of the candidate selector in the case of selecting candidate objects by further using object information in addition to emotion information and semantic information.
FIG. 6 is a reference diagram for explaining the operation of the video image conversion apparatus according to the present invention.
7 is a detailed block diagram of the image conversion unit.
8 is a block diagram for explaining the operation of the emotion information extracting unit.
9 is a flowchart of a video emotion conversion method according to the present invention.
10 is a flowchart of a video emotion conversion method further including an emotion information extracting step.
Hereinafter, preferred embodiments of the present invention will be described in detail with reference to the accompanying drawings. In the drawings, the same reference numerals are used to designate the same or similar components throughout the drawings. In the following description of the present invention, a detailed description of known functions and configurations incorporated herein will be omitted when it may make the subject matter of the present invention rather unclear. In addition, the preferred embodiments of the present invention will be described below, but it is needless to say that the technical idea of the present invention is not limited thereto and can be variously modified by those skilled in the art.
There have been various image conversion technologies that process image signals from various smart devices such as a smart phone and a tablet or a computer and give specific effects to the images. Especially, as the use of social network services has been expanded, applications have been developed in which each user can take a picture by using a smart device, download a picture from the web, give a desired image effect, and store or retransmit the picture. have. Traditionally, image processing techniques that give specific effects to images have generally been used to adjust the saturation and brightness of the image, or to use blurring or sharpening effects on the image.
On the other hand, people tend to communicate their feelings to others by shooting images and posting or delivering them. Thus, many existing studies have been conducted to recognize emotion information felt by people from images. And existing studies have focused primarily on the combination of features to use to extract emotions from images. However, since a person perceives an object included in the image and interprets the information about the object, the emotion that the image brings to the person is greatly influenced by the meaning of the object existing in the image. However, existing studies have focused only on finding a more efficient combination of features, and there is a limit in that attempts to extract emotion from the images using the semantic information of the objects included in the images are neglected.
On the other hand, image processing methods have been proposed for converting the emotion of an image so that the image expresses the emotion desired by the user. In other words, there are methods of converting images using a preset image filter or using a color palette so as to more intuitively reflect a certain emotion desired by a user in an image. For example, if the user wishes to change the mood of a specific image more pleasantly or more unpleasantly, there are image emotion conversion methods of converting the image so that the color, lightness, etc. are changed by reflecting such emotion information. However, in the existing methods of converting emotions, a method of converting only the overall color information of an image is used in order to express effects according to specific emotion information. In this regard, But there is a limit in that it can not.
In the present invention, attention is paid to the fact that the emotion felt by a person viewing a photograph is most affected by an object included in the photograph, so that not only the color information included in the photograph but also the information about the object are used together to determine the target A method and a device associated therewith for converting an image to better represent emotions.
The image emotion conversion method according to the present invention does not perform image signal processing for emotion conversion for the entire image as in the conventional image emotion conversion methods, And performs video signal processing. More specifically, in the image emotion conversion method according to the present invention, the semantic information, which is the information about the meaning of the object existing in the image, and the feature according to the image signal are extracted and based on the extracted emotion, Lt; / RTI > To this end, the present invention uses an emotion information classifier configured by learning learning data in which semantic information, characteristics, and emotion information of an object included in an image are preset.
In addition, the image emotional conversion method according to the present invention extracts an object from a target image, searches for a candidate object having semantic information of the extracted object and conversion target emotion information for the target image in the candidate database, and uses the selected candidate object And converts the extracted object. More specifically, the candidate information is previously stored in the candidate database by using the emotion information and the semantic information, and information according to the position of the object in the image or the characteristics of the image signal, if necessary, A candidate object having information similar to the above information of the object is selected from the candidate data base and the image signal of the extracted object is transformed using the selected candidate object.
According to the above configuration, the image emotion conversion method according to the present invention can extract the emotion information containing the image more accurately based on the semantic information of the object, and can also extract the image based on the object, And convert the emotion of the target image so that the target emotion desired by the user is better represented.
Hereinafter, a video image conversion apparatus and a method thereof according to the present invention will be described in detail.
First, the image emotion conversion apparatus according to the present invention will be described.
FIG. 1 is a block diagram of a video image conversion apparatus according to an exemplary embodiment of the present invention. Referring to FIG.
The image emotion conversion apparatus according to the present invention may include an
The
The
The
Here, the video image conversion apparatus according to the present invention may further include an emotion information extraction unit 400 if necessary.
2 is a block diagram of an image-emotion conversion apparatus further including an emotion-information extracting unit 400. As shown in FIG.
In this case, the
The emotion information extracting unit 400 extracts emotion information of the target object by inputting the semantic information and the characteristic of the extracted target object into a previously learned emotion information classifier and extracts emotion information of the extracted target object The emotion information of the target image can be determined based on the emotion information.
Here, the video image conversion apparatus according to the present invention may be configured such that all of the components are implemented as one independent hardware, or a part or all of the components are selectively combined to form a part or a combination But may also be embodied as a computer program having a program module that performs all of its functions. Further, the video image conversion apparatus according to the present invention may be implemented as a software program and operated on a processor or a signal processing module, or may be implemented in hardware form to be included in various processors, chips, semiconductors, to be. Further, the video image conversion apparatus according to the present invention may be included in various embedded systems or devices such as a computer, a mobile phone, a tablet, a handheld device, a wearable device, or the like in the form of hardware or software modules.
Hereinafter, the operation of the
The
Here, the target object information may include information about an area in which the target object exists in the target image. Hereinafter, the target object may be extracted to a certain region within the target image through image segmentation, and the information about the extracted region may be included in the target object information. Also, the target object information may include position information in the target image of the target object determined according to the position in the target image of the extracted region as described above.
Here, the semantic information means information about the meaning of the target object. The semantic information may include a tag indicating the meaning of the target object, and may include information indicating a category to which the target object belongs. For example, the semantic information may include a tag such as 'pine', and may include information indicating a category such as 'tree'. The semantic information may include various information indicating the meaning of the other object.
First, the
Image segmentation is a technique of dividing an image into a plurality of segments - a set of pixels - and is a technique widely used for detecting the position of a boundary of an object or an object. Here, image segmentation is a technique of projecting or classifying each pixel having a different image signal value included in an image into a predetermined category according to a predetermined criterion. For example, it is possible to generate one segment by classifying pixels having similar signal values adjacent to each other, pixels corresponding to the same object into one category, and generating a plurality of segments in the image in this manner have.
Preferably, the
In addition, the
Here, the
Here, the
Alternatively, the
In the field of image processing, the saliency is a numerical value of a portion of a human being felt important when viewing the image. In other words, there are parts of the human eye that feel more important depending on the color, brightness, and contour of the object in the image, and the saliency is a numerical value of the degree. For example, the saliency can be set according to the degree of difference in color or brightness, and the degree of strong feature of the contour. For example, "Yan, Q., Xu, L., Shi, J., & Jia, J. (2013, June). Hierarchical saliency detection.In Computer Vision and Pattern Recognition (CVPR) (IEEE Std 1155-1162, IEEE), "images are classified by layers (from a coarse level where images are simplified to a fine level where fine details remain), and a regional contrast is calculated for each layer, And calculates a final importance score by combining the scores calculated by the respective layers. In the present invention, the importance value is calculated using various existing methods as described above, and a target object among the divided regions can be selected based on the importance value.
Here, the
In addition, the
Alternatively, the
Here, the
For this, the
Alternatively, the
Here, the class of the object is a predetermined class for the object, and is a concept used in the field of object recognition. Here, the class may be a value or a label set according to the purpose of recognizing the object.
Here, the
Here, the
FIG. 3 is a detailed block diagram of the
The
The
The
The
Here, the video signal may mean a video signal value that each pixel included in the image has according to a predefined color space. For example, when an RGB color space is used, it may be a signal value of each of the R, G, and B channels. Or may be a luminance signal or a color difference signal in a pixel. Here, the image signal includes all of image signal values in various color spaces such as YCbCr, CMYK, YIQ, and the like.
The semantic
Next, the operation of the
The
Here, the emotion information may include a label or a numerical value indicating a specific emotion as information indicating a specific emotion. For example, emotion information can include labels that express qualitative emotions such as 'joy', 'sadness', 'loveliness', and may include labels indicating the nature of emotions such as 'affirmative' And may include numerical values representing the degree of certain emotions.
Here, the candidate object refers to an object used for transforming a target image in the
Here, the
Here, the
At this time, the
4 is a block diagram for explaining the operation of the
Here, the
Here, the
Here, when the emotion information is set to the emotional label, if the emotion information of the candidate object and the target emotion information have the same emotional label or have the emotional label classified into the same category, the candidate object can be selected. That is, the plurality of emotion labels can be classified into the same category in advance according to the nature of the emotion, and can be judged to be similar if the emotion level belongs to the same category.
For example, when the target emotion information corresponds to the emotion label of 'joy', the
In the same manner, the
Or the
Here, the
Here, the
FIG. 5 is a block diagram of the
In this case, the
The candidate object information including the intra-image location information of the candidate object may be stored in the
In this case, the
In addition, the
The candidate object information including the texture information of the candidate object may be stored in the
In this case, the
Here, the method by which the
6 is a reference diagram for explaining the operations of the
As shown in FIG. 6A, the
Next, the
Next, the operation of the
The
The conversion of the video signal of the target image corresponding to the target object may be performed by converting the video signal in the area corresponding to the target object, And < / RTI >
Here, the candidate object information may include information on the color, position, texture, or the like of the candidate object, and the
7 is a detailed block diagram of the
The
Here, the candidate object information may include color information of the candidate object.
Here, the
Preferably, the color information may be color distribution information of the candidate object, and the
Here, the color distribution can be the histogram of each L, a, b in the three-dimensional color space of the Lab. In addition to Lab color space, it can include all color spaces such as RGB, HSV, HSI, CMYK, YCbCr.
Here, the
Preferably, the
Here, the
Referring to FIGS. 6B and 6C, the
The candidate object information may include position information in the image of the candidate object.
Here, the position of the candidate object in the image may be the position information of the candidate object in the candidate image including the candidate object.
Here, the
For example, in (a) of FIG. 6, it is possible to perform image transformation to move the position of the target object T in accordance with the position information of the candidate object.
The candidate object information may include texture information of the candidate object. Here, the texture information means texture information of a video signal of a candidate object.
Here, the
Here, the
Here, the existing related word search method can be used to find a related word corresponding to the word according to the semantic information. For example, the same method proposed by ConceptNet may be used. Here, the image database may be a database storing image patches and a plurality of data in which words corresponding to the patches are defined.
For example, when the semantic information of the target object is 'Tree', the
Referring to FIG. 6 (d), the
Next, the operation of the emotion information extracting unit 400 will be described in more detail.
As described above, the emotion information extracting unit 400 according to the present invention draws attention to the fact that emotions that people see and feel are affected most by the objects included in the photograph, Extracts emotions contained in the image based on the extracted semantic information and feature based on the image signal.
FIG. 8 is a block diagram for explaining the operation of the emotion information extracting unit 400. FIG.
In this case, the
The emotion information extracting unit 400 according to the present invention extracts emotion from the target image by using the extracted features together with the semantic information of the target object recognized by the
To this end, the present invention uses an emotion information classifier set by learning learning data in which semantic information, features, and emotion information of an object included in an image are preset.
The emotion information extracting unit 400 extracts emotion information of the target object by inputting the semantic information and the characteristic of the extracted target object into a previously learned emotion information classifier and extracts emotion information of the extracted target object The emotion information of the target image can be determined based on the emotion information.
If the target object is one, the extracted emotion information of the target object may be emotion information of the target image. If there are a plurality of target objects, the set of emotion information of each target object may be emotion information of the target image, Some of the emotion information of the object may be the emotion information of the target image. For example, the emotion information of the target object selected as the main target object may be the emotion information of the target image, or the emotion information of the target image may be the emotion information of the target image.
Here, the emotion information classifier can extract the emotion information of the target object from the semantic information and the feature by using the classification function in which parameters are set in advance. Here, the emotion information extracting unit 400 may extract the emotion information of the target object by inputting the semantic information of the target object and the feature including the color feature or texture feature to the classification function of the emotion information classifier have.
For example, the emotion information extracting unit 400 may calculate and extract emotion information of the target object according to the result of the classification function as shown in Equation (1).
Where S is the semantic information, F is the feature, C () is the classification function, and E is the result of the classification function. The emotion information extracted here may be the resultant value of the classification function itself or may be an emotional label or numeric value classified according to whether it belongs to a predetermined numerical range based on the resultant value.
Here, the emotion information classifier used by the emotion information extractor 400 learns a plurality of learning data in which the semantic information, the characteristic, and the emotion information are set in advance for a plurality of learning objects, and the semantic information and the characteristic And the parameter of the classification function for outputting the emotion information is set to be a set classifier.
Preferably, the emotion information classifier sets correspondence between the semantic information and the feature information and the emotion information using a linear regression model or a support vector-based regression model, and learns a model using the learning data have. That is, the emotion information classifier may learn the learning data using a linear regression model or a support vector-based regression model, and the classifier may be set to a parameter of the classification function.
In addition, the parameters of the classification function of the emotion information classifier can be set through machine learning. In this case, various conventional methods can be used for the classification function and the corresponding machine learning method. For example, an ensemble learning method, a boost-based learning method, a nearest neighborhood search-based learning method, a support vector machine-based learning method, and an AdaBoost-based learning method can be used. In addition, the emotion information extracting unit 400 can set parameters of the classification function of the emotion information classifier by using various types of classification functions and the corresponding machine learning methods, without being limited to the above-described methods.
The image emotion conversion method according to another embodiment of the present invention may include an object recognition step S100, a candidate selection step S200, and an image conversion step S300. The image emotion conversion method according to the present invention can operate in the same manner as the image emotion conversion apparatus described in detail with reference to FIGS. 1 to 8 above. Therefore, overlapping portions will be omitted and briefly described below. In addition, the image emotion conversion method according to the present invention can be included in a computer or an embedded system in the form of hardware or software modules.
9 is a flowchart of a video emotion conversion method according to the present invention.
The object recognizing step S100 receives the image to be emotionally transformed, extracts a target object to be emotion transformed from the target image, and outputs target object information including information about a position or an area in the target image of the target object And recognizes the target object and generates semantic information, which is information on the meaning of the target object.
The candidate selection step S200 includes inputting target emotion information indicating a target emotion to be transformed with respect to the target image, comparing at least one candidate object and the target object based on the semantic information and emotion information, And selects the candidate object according to the candidate object.
The image transformation step S300 transforms the image signal of the target image in the target image corresponding to the target object using the candidate object information of the candidate object selected in the candidate selection step, Lt; RTI ID = 0.0 > emotion < / RTI >
Here, the image emotion conversion method according to the present invention may further include the emotion information extraction step S50 as needed.
10 is a flowchart of a video emotion conversion method further including a feeling information extracting step (S50).
In this case, the object recognition step S100 may further include analyzing the image signal of the pixels included in the target image area corresponding to the extracted target object to extract a predetermined feature.
The emotion information extraction step S400 extracts emotion information of the target object by inputting the semantic information and the feature of the extracted target object to a previously learned emotion information classifier and extracting emotion information of the extracted target object The emotion information of the target image is determined based on the emotion information.
Here, the candidate object is stored in the
Here, the candidate selection step S200 compares the semantic information and the emotion information of the candidate object stored in the
Wherein the candidate object information includes color information of the candidate object,
Here, in the image transforming step S300, the color distribution of the image signal of the target image within the target image corresponding to the target object is calculated using the color information included in the candidate object information of the candidate object selected in the candidate selection step And a color conversion step of converting the image signal according to the color information of the candidate object to generate the converted image.
In the image conversion step S300, an associated word corresponding to the semantic information of the target object is searched in a word database, an image patch corresponding to the searched related word is searched in an image database, And adding an object on the target image to generate the transformed image.
The image emotion conversion computer program according to another embodiment of the present invention may be a computer program stored in the medium for executing the image emotion conversion method described above in combination with the computer.
It is to be understood that the present invention is not limited to these embodiments, and all elements constituting the embodiment of the present invention described above are described as being combined or operated in one operation. That is, within the scope of the present invention, all of the components may be selectively coupled to one or more of them.
In addition, although all of the components may be implemented as one independent hardware, some or all of the components may be selectively combined to perform a part or all of the functions in one or a plurality of hardware. As shown in FIG. In addition, such a computer program may be stored in a computer readable medium such as a USB memory, a CD disk, a flash memory, etc., and read and executed by a computer to implement an embodiment of the present invention. As the recording medium of the computer program, a magnetic recording medium, an optical recording medium, a carrier wave medium, and the like can be included.
Furthermore, all terms including technical or scientific terms have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs, unless otherwise defined in the Detailed Description. Commonly used terms, such as predefined terms, should be interpreted to be consistent with the contextual meanings of the related art, and are not to be construed as ideal or overly formal, unless expressly defined to the contrary.
It will be apparent to those skilled in the art that various modifications, substitutions and substitutions are possible, without departing from the scope and spirit of the invention as disclosed in the accompanying claims. will be. Therefore, the embodiments disclosed in the present invention and the accompanying drawings are intended to illustrate and not to limit the technical spirit of the present invention, and the scope of the technical idea of the present invention is not limited by these embodiments and the accompanying drawings . The scope of protection of the present invention should be construed according to the following claims, and all technical ideas within the scope of equivalents should be construed as falling within the scope of the present invention.
100: Object recognition unit
200: candidate selection unit
300:
400: Emotion information extracting unit
50: Candidate database
110: Image Segmentation Unit
120: object extracting unit
130: Feature extraction unit
140: Semantic information generating unit
310: Color conversion unit
320:
330: texture conversion unit
340:
S100: Object recognition
S200: Candidate selection
S300: Image conversion
Claims (21)
An object to be emotionally transformed is extracted and a target object to be emotion transformed is extracted from the target image to generate target object information including information about a position or an area in the target image of the target object, An object recognition unit for recognizing and generating semantic information which is information on the meaning of the target object;
The method includes receiving target emotion information representing a target emotion to be transformed by a user with respect to the target image, receiving semantic information and emotion information for each of at least one candidate object preset and stored in the candidate database, A candidate selecting unit for calculating the similarity between the semantic information and the target emotion information and selecting candidate objects corresponding to the target object, the candidate objects having the similarity calculated among the at least one candidate object being greater than a preset reference; And
Wherein the candidate selecting unit converts the video signal of the target intra-region corresponding to the target object using the candidate object information of the candidate object selected by the candidate selecting unit to convert the target image into an emotion corresponding to the target emotion information An image converter for generating an image; Lt; / RTI >
Wherein the object recognizer extracts a predetermined feature including a color feature or a texture feature by analyzing an image signal of pixels included in the target image region corresponding to the extracted target object,
Wherein the object recognition unit preliminarily learns a plurality of learning data in which the semantic information, the predetermined characteristic, and the feeling information are preliminarily learned, thereby obtaining a parameter of a classification function for inputting the semantic information and the predetermined characteristic and outputting the feeling information The method of claim 1, further comprising the steps of: inputting the semantic information and the predetermined characteristic of the extracted target object to a predetermined emotion information classifier to extract emotion information of the target object; And an emotion information extracting unit for determining emotion information.
Wherein the object recognition unit performs image segmentation on the target image to divide the target image into a plurality of regions, extracts the target image region corresponding to the target object by selecting a portion of the divided regions, And generates the target object information including information on a position or an area in the target image of the target object.
The object recognition unit determines which class of the predetermined object the class of the object corresponds to, using the image signal of the target image or the extracted characteristic, and generates the semantic information according to the determined class of the object And the image processing unit converts the image into a video signal.
An image segmentation unit dividing the target image into a plurality of regions by performing image segmentation;
An object extraction unit for selecting a part of the divided areas and extracting the target image area corresponding to the target object;
A feature extraction unit for analyzing the image signal of the pixels included in the target image region corresponding to the target object and extracting the predetermined feature; And
A semantic information generation unit for determining which class of the predetermined class of object the target object corresponds to using the predetermined feature extracted from the target image and generating the semantic information according to the determined class of the object; Further comprising: a video emotional conversion unit for generating a video emotional state;
Wherein the object recognition unit extracts an area in the target image corresponding to the target object and uses the position information of the extracted area as the position information of the target object to obtain the target object information including the position information of the target object Generate,
The candidate object information including position information of the candidate object in the image for each candidate object is stored in the candidate database,
Wherein the candidate selector is configured to classify the semantic information, the emotion information, and the candidate object information of the candidate object stored in the candidate database, the semantic information of the target object, the target emotion information, Compares the information, calculates the degree of similarity, and selects candidates corresponding to the target object according to the degree of similarity.
Wherein the object recognition unit extracts an area in the target image corresponding to the target object and uses the texture information of the extracted area as texture information of the target object to generate the target object information including the texture information of the target object and,
The candidate object information including texture information of the candidate object for each candidate object is stored in the candidate database,
Wherein the candidate selector is configured to classify the semantic information, the emotion information, and the candidate object information of the candidate object stored in the candidate database, the semantic information of the target object, the target emotion information, Compares the information, calculates the degree of similarity, and selects candidates corresponding to the target object according to the degree of similarity.
Wherein the candidate object information includes color information of the candidate object,
Wherein the image transformation unit uses the color information included in the candidate object information of the candidate object selected by the candidate selection unit so that a color distribution of the image signal of the target image in the target image corresponding to the target object is obtained from the color information of the candidate object And a color conversion unit for converting the image signal according to the color information to generate the converted image.
Wherein the candidate object information includes position information in the image of the candidate object,
Wherein the image transforming unit transforms the target object in the target image to the in-image position information of the candidate object using the intra-image position information included in the candidate object information of the candidate object selected by the candidate selecting unit And a position conversion unit for performing an image conversion for moving the position of the target object in the target image and generating the converted image.
Wherein the candidate object information includes texture information of the candidate object,
Wherein the image transformation unit uses the texture information included in the candidate object information of the candidate object selected by the candidate selection unit so that the target object in the target image matches the texture information of the candidate object, And a texture conversion unit for converting the image signal of the corresponding area in the target image to generate the converted image.
Wherein the image conversion unit searches the word database for an associated word corresponding to the semantic information of the target object, searches an image database for an image patch corresponding to the retrieved related word, And an object adding unit for generating the converted image in addition to the image adding unit.
Wherein the emotion information classifier learns the learning data using a linear regression model or a support vector-based regression model, and the parameter of the classification function is set by the classifier.
An object to be emotionally transformed is extracted and a target object to be emotion transformed is extracted from the target image to generate target object information including information about a position or an area in the target image of the target object, An object recognizing step of recognizing and generating semantic information which is information on the meaning of the target object;
The method includes receiving target emotion information representing a target emotion to be transformed by a user with respect to the target image, receiving semantic information and emotion information for each of at least one candidate object preset and stored in the candidate database, A candidate selection step of calculating a similarity between the semantic information and the target emotion information and selecting a candidate object corresponding to the target object whose calculated similarity among the at least one candidate object is larger than a preset reference; And
Wherein the target image is transformed into an image signal of an area in the target image corresponding to the target object using the candidate object information of the candidate object selected in the candidate selection step so that the target image is transformed to indicate the emotion according to the target emotion information And an image conversion step of generating an image,
The object recognition step
Extracting a predetermined feature including a color feature or a texture feature by analyzing a video signal of pixels included in the target intra-region corresponding to the extracted target object; And
Learning a plurality of learning data in which the semantic information, the predetermined feature, and the emotion information are set in advance, so that the parameter of the classification function for inputting the semantic information and the predetermined characteristic, Extracting emotion information of the target object by inputting the semantic information of the extracted target object and the predetermined feature including the color feature or the texture feature into the information classifier; And
Determining emotion information of the target image based on the extracted emotion information of the target object; Wherein the video emotional conversion method comprises:
Wherein the candidate object information includes color information of the candidate object,
Wherein the image transformation step uses the color information included in the candidate object information of the candidate object selected in the candidate selection step so that the color distribution of the image signal of the target image in the target image corresponding to the target object is the candidate object, And a color conversion step of converting the image signal to comply with the color information of the input image, and generating the converted image.
Wherein the image conversion step searches the word database for an associated word corresponding to the semantic information of the target object, searches the image database for an image patch corresponding to the retrieved related word, And an object adding step of creating the transformed image in addition to the transformed image.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020150104116A KR101606760B1 (en) | 2015-07-23 | 2015-07-23 | Apparatus and Method of Transforming Emotion of Image based on Object in Image |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020150104116A KR101606760B1 (en) | 2015-07-23 | 2015-07-23 | Apparatus and Method of Transforming Emotion of Image based on Object in Image |
Publications (1)
Publication Number | Publication Date |
---|---|
KR101606760B1 true KR101606760B1 (en) | 2016-03-28 |
Family
ID=57007635
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020150104116A KR101606760B1 (en) | 2015-07-23 | 2015-07-23 | Apparatus and Method of Transforming Emotion of Image based on Object in Image |
Country Status (1)
Country | Link |
---|---|
KR (1) | KR101606760B1 (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101830512B1 (en) * | 2016-07-08 | 2018-02-20 | 전자부품연구원 | Optimized Image Segmentation Methods and System with DL and PDE |
WO2019066373A1 (en) * | 2017-09-27 | 2019-04-04 | 삼성전자주식회사 | Method of correcting image on basis of category and recognition rate of object included in image and electronic device implementing same |
KR20190044762A (en) * | 2017-10-23 | 2019-05-02 | 연세대학교 산학협력단 | Apparatus Transforming Object based on Target Emotion and Method thereof |
WO2019156508A1 (en) * | 2018-02-08 | 2019-08-15 | Samsung Electronics Co., Ltd. | Method and electronic device for rendering background in image |
KR20200027794A (en) * | 2018-09-05 | 2020-03-13 | 삼성전자주식회사 | Image display device and operating method for the same |
CN117412450A (en) * | 2023-12-13 | 2024-01-16 | 深圳市千岩科技有限公司 | Atmosphere lamp equipment, lamp effect color matching method thereof, corresponding device and medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2003248819A (en) * | 2002-02-22 | 2003-09-05 | Toshiba Corp | Device and method for image processing, program, and storage medium |
KR20110000910A (en) * | 2009-06-29 | 2011-01-06 | 가톨릭대학교 산학협력단 | Method for converting color of image based on relationship between color and emotion |
JP2013011944A (en) * | 2011-06-28 | 2013-01-17 | Nikon Corp | Image processing apparatus, imaging apparatus, image processing method, and image processing program |
KR20140037373A (en) | 2012-09-17 | 2014-03-27 | 중앙대학교 산학협력단 | Apparatus and method for transferring a color of image |
KR20140091554A (en) * | 2011-11-17 | 2014-07-21 | 마이크로소프트 코포레이션 | Automatic tag generation based on image content |
-
2015
- 2015-07-23 KR KR1020150104116A patent/KR101606760B1/en active IP Right Grant
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2003248819A (en) * | 2002-02-22 | 2003-09-05 | Toshiba Corp | Device and method for image processing, program, and storage medium |
KR20110000910A (en) * | 2009-06-29 | 2011-01-06 | 가톨릭대학교 산학협력단 | Method for converting color of image based on relationship between color and emotion |
JP2013011944A (en) * | 2011-06-28 | 2013-01-17 | Nikon Corp | Image processing apparatus, imaging apparatus, image processing method, and image processing program |
KR20140091554A (en) * | 2011-11-17 | 2014-07-21 | 마이크로소프트 코포레이션 | Automatic tag generation based on image content |
KR20140037373A (en) | 2012-09-17 | 2014-03-27 | 중앙대학교 산학협력단 | Apparatus and method for transferring a color of image |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101830512B1 (en) * | 2016-07-08 | 2018-02-20 | 전자부품연구원 | Optimized Image Segmentation Methods and System with DL and PDE |
WO2019066373A1 (en) * | 2017-09-27 | 2019-04-04 | 삼성전자주식회사 | Method of correcting image on basis of category and recognition rate of object included in image and electronic device implementing same |
KR20190036168A (en) * | 2017-09-27 | 2019-04-04 | 삼성전자주식회사 | Method for correcting image based on category and recognition rate of objects included image and electronic device for the same |
US11270420B2 (en) | 2017-09-27 | 2022-03-08 | Samsung Electronics Co., Ltd. | Method of correcting image on basis of category and recognition rate of object included in image and electronic device implementing same |
KR102383129B1 (en) * | 2017-09-27 | 2022-04-06 | 삼성전자주식회사 | Method for correcting image based on category and recognition rate of objects included image and electronic device for the same |
KR20190044762A (en) * | 2017-10-23 | 2019-05-02 | 연세대학교 산학협력단 | Apparatus Transforming Object based on Target Emotion and Method thereof |
KR102022479B1 (en) * | 2017-10-23 | 2019-09-18 | 연세대학교 산학협력단 | Apparatus Transforming Object based on Target Emotion and Method thereof |
WO2019156508A1 (en) * | 2018-02-08 | 2019-08-15 | Samsung Electronics Co., Ltd. | Method and electronic device for rendering background in image |
KR20200027794A (en) * | 2018-09-05 | 2020-03-13 | 삼성전자주식회사 | Image display device and operating method for the same |
KR102579452B1 (en) | 2018-09-05 | 2023-09-15 | 삼성전자주식회사 | Image display device and operating method for the same |
CN117412450A (en) * | 2023-12-13 | 2024-01-16 | 深圳市千岩科技有限公司 | Atmosphere lamp equipment, lamp effect color matching method thereof, corresponding device and medium |
CN117412450B (en) * | 2023-12-13 | 2024-04-02 | 深圳市千岩科技有限公司 | Atmosphere lamp equipment, lamp effect color matching method thereof, corresponding device and medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR101606760B1 (en) | Apparatus and Method of Transforming Emotion of Image based on Object in Image | |
US10762608B2 (en) | Sky editing based on image composition | |
KR101289085B1 (en) | Images searching system based on object and method thereof | |
US10839573B2 (en) | Apparatus, systems, and methods for integrating digital media content into other digital media content | |
US10332266B2 (en) | Method and device for traffic sign recognition | |
KR101611895B1 (en) | Apparatus and Method of Automatic Text Design based on Emotion | |
CN105184763B (en) | Image processing method and device | |
Mehrani et al. | Saliency Segmentation based on Learning and Graph Cut Refinement. | |
US11651477B2 (en) | Generating an image mask for a digital image by utilizing a multi-branch masking pipeline with neural networks | |
US11393100B2 (en) | Automatically generating a trimap segmentation for a digital image by utilizing a trimap generation neural network | |
Ge et al. | Co-saliency detection via inter and intra saliency propagation | |
JP2010026603A (en) | Image processor, image processing method and computer program | |
KR101833943B1 (en) | Method and system for extracting and searching highlight image | |
CN113626444A (en) | Table query method, device, equipment and medium based on bitmap algorithm | |
CN110196917B (en) | Personalized LOGO format customization method, system and storage medium | |
KR20090065099A (en) | System for managing digital image features and its method | |
US11461880B2 (en) | Generating image masks from digital images utilizing color density estimation and deep learning models | |
Sravani et al. | Robust detection of video text using an efficient hybrid method via key frame extraction and text localization | |
JP2020087165A (en) | Learning data generation program, learning data generation device, and learning data generation method | |
Deshmukh et al. | Real-time traffic sign recognition system based on colour image segmentation | |
KR101374726B1 (en) | Feature descriptor generating device and method, image object recognition device and method using the same | |
Khan et al. | Systematic skin segmentation: merging spatial and non-spatial data | |
Ghandour et al. | Building shadow detection based on multi-thresholding segmentation | |
JP5158974B2 (en) | Attention area extraction method, program, and image evaluation apparatus | |
US20230169708A1 (en) | Image and video matting |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
E902 | Notification of reason for refusal | ||
E701 | Decision to grant or registration of patent right | ||
GRNT | Written decision to grant |