WO2013027893A1 - Apparatus and method for emotional content services on telecommunication devices, apparatus and method for emotion recognition therefor, and apparatus and method for generating and matching the emotional content using same - Google Patents

Apparatus and method for emotional content services on telecommunication devices, apparatus and method for emotion recognition therefor, and apparatus and method for generating and matching the emotional content using same Download PDF

Info

Publication number
WO2013027893A1
WO2013027893A1 PCT/KR2011/008399 KR2011008399W WO2013027893A1 WO 2013027893 A1 WO2013027893 A1 WO 2013027893A1 KR 2011008399 W KR2011008399 W KR 2011008399W WO 2013027893 A1 WO2013027893 A1 WO 2013027893A1
Authority
WO
WIPO (PCT)
Prior art keywords
emotion
image
face
user
video call
Prior art date
Application number
PCT/KR2011/008399
Other languages
French (fr)
Korean (ko)
Inventor
강준규
Original Assignee
Kang Jun-Kyu
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Kang Jun-Kyu filed Critical Kang Jun-Kyu
Publication of WO2013027893A1 publication Critical patent/WO2013027893A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management
    • G06Q10/101Collaborative creation, e.g. joint development of products or services
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/10Services
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/20Scenes; Scene-specific elements in augmented reality scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/174Facial expression recognition
    • G06V40/176Dynamic expression
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/04Protocols specially adapted for terminals or networks with limited capabilities; specially adapted for terminal portability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/131Protocols for games, networked simulations or virtual reality
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects

Definitions

  • the present invention analyzes the emotions and facial expressions recognized by the image pickup apparatus of the transmitting communication terminal device, and the emotion contents for realizing mixing the virtual object on the screen of the receiving communication terminal device in real time so as to effectively deliver the analyzed emotion or communication.
  • Service apparatus and method an emotion recognition apparatus and method for serving the emotion content, an apparatus and method for generating and matching the emotion content through the emotion recognition, and an apparatus and method for generating the emotion content .
  • Augmented reality which is the second key keyword in the IT field, is a technology derived from one field of virtual reality and refers to a technology that combines the real world and the virtual experience. Augmented reality is regarded as one of the top 10 innovations to lead the future, and it is a technology that gives the user a better sense of reality by interacting with virtual objects based on the real world.
  • Augmented reality is a field of virtual reality, a computer graphics technique that synthesizes virtual objects in the real environment and looks like objects existing in the original environment.
  • Augmented reality is an existing method that targets only virtual space and virtual objects. Unlike virtual reality, it is a technology that synthesizes virtual objects on the basis of real world and reinforces and provides additional information that is difficult to obtain in real world alone.
  • augmented reality technology is being actively used in various forms in the fields of broadcasting, advertising, exhibition, games, theme parks, military, education and promotion.
  • augmented reality differs from the virtual reality technology, which excludes interaction with the real world and processes interactions only in the pre-established virtual space, based on real-time processing.
  • the information is overlaid on the image of the real world input through the terminal, and thus it is distinguished from the virtual reality that provides only the image generated by the computer in that it enables interaction with the real world.
  • Marker-based mobile augmented reality technology is a technology for recognizing a building by recognizing a specific sign after taking a specific sign corresponding to a specific building when shooting a specific building. It is a technology that overlays POI (Point of Interests) information corresponding to the image in the inferred direction by inferring the current position of the terminal and the viewing direction by using the installed GPS and digital compass. .
  • POI Point of Interests
  • 3D video content the third key keyword in the IT field, is exploding in related industries due to James Cameron's 'Avatar', and it is expected that the time for enjoying video calls with 3D content will come.
  • Android phones such as Samsung's Galaxy SI II are supporting video calls, and since Android version 2.3 (Gingerbread) officially supports video calls, Android phones from mid-2011 have been able to support video calls. It is expected to be equipped with call function.
  • Android-based Tablet computers and iPad 2 generations are also expected to provide video call service through front and back cameras, so video calls are becoming more common.
  • 1: 1 or multi-party video calls may grow into core services in addition to voice and data services.
  • interest in video calls through mobile terminals is gradually increasing.
  • an object of the present invention is to provide an emotional content service apparatus and method that can provide fun to a video call by providing a virtual object representing the emotional state of both callers with the video during a video call, and the emotional content.
  • another object of the present invention is to provide an emotional content service apparatus and method for superimposing the emotional state of the caller to the video of the caller through a virtual object, to experience augmented reality that gives the callers a more realistic feeling and And an emotion recognition apparatus and method for servicing the emotion content, an apparatus and method for generating the emotion content through the emotion recognition, and the emotion content generated by the apparatus and method for generating the emotion content. have.
  • the present invention analyzes the emotions and facial expressions recognized from the image pickup device of the transmitting communication terminal device, and the emotion that is implemented by real-time mixing a virtual object on the screen of the receiving communication terminal device to effectively deliver the analyzed emotion or communication
  • An apparatus and method for providing content services are provided.
  • the present invention analyzes the emotions and facial expressions recognized from the image pickup device of the transmitting communication terminal device, and the emotion that is implemented by real-time mixing a virtual object on the screen of the receiving communication terminal device to effectively deliver the analyzed emotion or communication
  • the present invention provides an apparatus and method for emotion recognition for providing a content service.
  • the present invention analyzes the emotions and facial expressions recognized from the image pickup device of the transmitting communication terminal device, and the emotion that is implemented by real-time mixing a virtual object on the screen of the receiving communication terminal device to effectively deliver the analyzed emotion or communication
  • An apparatus and method for generating the emotion content through an emotion recognition apparatus and method for providing a content service are provided.
  • the present invention analyzes the emotions and facial expressions recognized from the image pickup device of the transmitting communication terminal device, and the emotion that is implemented by real-time mixing a virtual object on the screen of the receiving communication terminal device to effectively deliver the analyzed emotion or communication
  • the present invention provides an emotional content generated by an apparatus and a method for generating the emotional content through an emotion recognition apparatus and a method for providing a content service.
  • the present invention for achieving the above object is to analyze the emotion to match the avatar to the avatar instead of a specific emotion to analyze, and the emoticon matching to add the effect on the specific expression by analyzing the expression and the expression for a particular emotion It has at least one feature which exaggerates the specific part of the face or body which it represents.
  • the present invention proposes a face detection, face recognition, and emotion recognition core technology for recognizing a user's emotions and expressions, and accordingly proposes a technology for generating and matching emoticons maximizing avatars and expressions for the recognized emotions.
  • a video call service through a smartphone.
  • the present invention can increase the effect of communication by recognizing a change in facial expression and matching the corresponding content (avatar) on a person's real face, thereby enabling expression of emotions that are impossible in the real world through facial recognition.
  • Apparatus and method for emotion content service of a communication terminal device for achieving the above objects, apparatus and method for emotion recognition for the same, apparatus and method for generating and matching emotion content using the same are voice recognition, object recognition Face area detection technology of face recognition technology, face area normalization technology, feature extraction technology within face area, facial component (expression analysis) relationship technology of emotion recognition technology of object recognition, object or hand gesture, object recognition behavior and behavior
  • voice recognition object recognition Face area detection technology of face recognition technology
  • face area normalization technology feature extraction technology within face area
  • object recognition behavior and behavior On the basis of cognitive technology, the real-time matching technology of the real picture and the virtual image is used to match the mixed virtual objects (including characters) through the gesture and facial expression analysis on both the face and the body to make a video call. Characterized in that the mixed reality is implemented through a video call.
  • the emotional content service of the communication terminal device is registered in advance the expression analysis relation function of the specific expression and gesture of the voice, face and body, if a similar voice, facial expression and gesture is transmitted through the image On the output video screen, the virtual object responding to voice, facial expression, and gesture is matched in real time on the face and body to enjoy a video call.
  • the emotional content service of the communication terminal apparatus for achieving the above objects is at least one of a gesture and facial expression of the user photographed through the image pickup means of the communication terminal device having at least an image pickup means and a display means
  • the virtual object may further include a character.
  • the virtual object may be changed by the user.
  • the virtual object is changed in real time corresponding to the emotional state.
  • the virtual object is characterized in that the position superimposed on the body and face of the user is changed.
  • the communication terminal device further comprises a voice input means and a voice output means, the emotional state of the user from the user's voice input through the voice input means Characterized in further extracting.
  • the emotional content service of the communication terminal apparatus for achieving the above object is a video call service providing terminal having at least an imaging means and a display means, the gesture and facial expression of the user to be photographed through the imaging means
  • the emotional content service method of the communication terminal device for achieving the above object, the process comprising: inputting a face image of the user to the communication terminal device; Extracting face components from the input face image; Preprocessing the extracted facial components; Extracting facial features from the preprocessed facial components; Registering the extracted feature of the face in a face database; And a step of recognizing an emotion by comparing a feature registered in the face database with a face component extracted in the feature extraction process.
  • the face image of the user from the camera module of the communication terminal device Receiving a process; Preprocessing the input face image; Detecting only valid data in the preprocessed face image; Estimating the position of the face and the camera information from the detected valid data; And generating a 3D image from the camera information and the position information of the face, and matching the generated 3D image with the face image of the user.
  • the method for generating and matching the emotion content through the emotion recognition method for the emotion content service of the communication terminal device comprises the steps of: outputting the 3D image matched with the face image of the user on the screen; And transmitting the 3D image matched with the face image of the user to a counterpart communication terminal through a network.
  • the emotional content service method of the communication terminal device for achieving the above object, the process of capturing a frame image to be analyzed in the video data source from the camera module of the communication terminal device; Preprocessing the captured image in an easy to analyze state; Detecting a face in the preprocessed image; Recognizing a posture estimation and a facial expression based on the recognition information extracted through the face detection process, and selecting a posture and facial expression of the avatar with respect to the posture and facial expression taken by the face; Determining position coordinates of the selected avatar on the 3D space through the analyzed information, selecting an avatar animation for a corresponding expression and emotion, and transmitting a control signal to the 3D engine; Composing a 3D space in which the avatar image and the video are to be represented, and performing a function of placing the analyzed avatar at a corresponding position; Matching the avatar and the video source represented through the 3D space into a single image source; Performing video encoding on the matched image together with a voice source; And
  • the process of detecting a face in the pre-processed image, applying a learning algorithm or the like to analyze the facial feature point, and the position and relationship data on the image of the component It is characterized in that the extraction.
  • the step of transmitting the matched image to the other terminal through the network establishes a session via SIP, and transmits to the Internet via RTP / RTCP Characterized in that.
  • the emotional content service apparatus of the communication terminal apparatus for achieving the above object, the server communication unit interworking with the video call service providing terminal; And recognize the emotion state of the user from at least one of a gesture and an expression of the user from the image information received from the video call service providing terminal, and compare the recognized emotion state with previously stored object related information.
  • the emotion content service apparatus of the communication terminal device characterized in that it further comprises a server storage unit for storing the object-related data corresponding to the emotional state.
  • an emotion recognition apparatus for an emotion content service of a communication terminal apparatus for achieving the above objects includes a display unit for displaying an image of the other party and an object overlapping the image according to the video call; A communication unit interworking with a video call service providing server; An imaging unit which acquires image information of a user according to a video call; And recognize the emotion state of the user from the image information obtained by the image pickup unit, extract emotion information related to the recognized emotion state, and transmit the extracted emotion information to the video call service providing server, and the image from the video call service providing server.
  • a controller configured to receive an object corresponding to the emotion information of the other party according to a call and to superimpose the received object on a position associated with the received object in the image of the other party according to the video call and to output the object to the display unit. It is characterized by.
  • the apparatus for generating and matching the emotion content through the emotion recognition method for the emotion content service of the communication terminal device for achieving the above object, the video of the other party and the video according to the video call
  • a display unit displaying overlapping objects;
  • a communication unit interworking with a video call service providing server;
  • An imaging unit which acquires image information of a user according to a video call; And recognize the emotion state of the user from the image information obtained by the image pickup unit, extract emotion information related to the recognized emotion state, and transmit the extracted emotion information to the video call service providing server, and the image from the video call service providing server.
  • a controller configured to receive an object corresponding to the emotion information of the other party according to a call and to superimpose the received object on a position associated with the received object in the image of the other party according to the video call and to output the object to the display unit. It is characterized by.
  • the display unit of the device for generating and matching the emotion content through the emotion recognition method for the emotion content service of the communication terminal device according to the present invention characterized in that for further displaying the image of the user according to the video call do.
  • the apparatus for generating and matching the emotion content through the emotion recognition method for the emotion content service of the communication terminal device further includes a key input unit for determining whether to apply the object. do.
  • the apparatus for generating and matching the emotion content through the emotion recognition method for the emotion content service of the communication terminal device according to the present invention may further include a storage unit for storing the object.
  • the emotional content service of the communication terminal apparatus for achieving the above objects is at least one of a gesture and facial expression of the user photographed through the image pickup means of the communication terminal device having at least an image pickup means and a display means
  • a communication terminal device of the other party that extracts the emotional state of the user from the user, exaggerates at least one of the body and the face of the user representing the gesture and expression corresponding to the extracted emotional state, and makes a video call with the user It is characterized by displaying on the display means.
  • the present invention can provide abundant sights to the video call by implementing various mixed reality not seen in the real world through the video call.
  • the present invention has an effect that can provide fun to the video call by providing a virtual object representing the emotional state of both callers with the video call during the video call.
  • the present invention by superimposing the emotional state of the caller to the video of the caller through a virtual object, it is possible to experience the augmented reality that gives the caller a more realistic feeling to deliver the emotional state of the caller freshly It has an effect.
  • the present invention has the effect of enabling the user to experience both virtual and reality by shaping the expressions of voices, faces and bodies of specific expressions and gestures into virtual objects through the video call screen.
  • FIG. 1 is a view showing a concept of avatar matching according to the present invention
  • FIG. 2 is a view showing a concept of emoticon matching according to the present invention
  • FIG. 3 is a view showing composite data matched to a change in facial expression of a user and standard data thereof according to an embodiment of the present invention
  • FIG. 4 is a control flow diagram illustrating an emotion and facial expression recognition and matching procedure of a 3D avatar
  • FIG. 6 is a diagram illustrating a service movement scenario according to user movement between access networks to which the emotion content service method according to the present invention is applied;
  • FIG. 7 is a control flowchart for face detection in an emotion recognition method for an emotion content service of a communication terminal device according to the present invention.
  • FIG. 9 is a diagram illustrating a basic message and status code scheme of a SIP
  • FIG. 10 is a diagram illustrating a SIP protocol stack
  • FIG. 11 is a diagram showing a basic procedure of call setup of a SIP protocol
  • FIG. 12 is a conceptual diagram of a content matching system according to the present invention.
  • FIG. 14 is a schematic diagram of an avatar video communication operation procedure through emotion recognition and image registration according to the present invention.
  • FIG. 15 is a view illustrating an avatar video communication operation procedure through emotion recognition and image registration of FIG. 14 using actual content
  • 16 is a view showing various fields to which the present invention is applicable.
  • the present invention has at least one feature of avatar matching to analyze an emotion to allow an avatar to express it for a specific emotion, and emoticon matching to analyze an expression and add an effect to the specific expression.
  • FIG. 1 is a view showing the concept of avatar matching according to the present invention
  • Figure 2 is a view showing the concept of emoticon matching according to the present invention.
  • the avatar matching according to the present invention recognizes an emotion expressed in an actual image and replaces the actual image with an avatar expressed in augmented reality.
  • the entire screen of the actual image may be replaced or only a specific part of the face may be expressed in augmented reality.
  • matching the emoticon according to the present invention is to use a variety of emoticons for recognizing the emotions expressed in the actual image to increase the transmission effect on the recognized expression.
  • the emotional content according to the present invention may be composed of, for example, ten male emotional recognition reactions, ten female emotional recognition reactions, and ten animal emotional recognition reactions.
  • emotion contents can express characters and motions linked with motion expression scripts through 3D modeling that can be matched to standard data.
  • FIG. 3 is a diagram showing synthetic data and standard data thereof matched to a change in facial expression of a user according to an exemplary embodiment of the present invention.
  • standard data can be used to make a more interesting and pleasant video call between users by substituting an avatar or an emotion that is difficult to express.
  • avatar creation and video call application on behalf of.
  • the present invention can be implemented to be applicable to the video call by producing a man, a woman, an anthropomorphic character according to the emotion and facial expression changes and each type of performance.
  • FIG. 4 is a control flowchart illustrating a process of recognizing and matching emotions and facial expressions of a 3D avatar. As illustrated in FIG. 4, a source of an avatar produced according to each emotion and facial expression is applied to a user application through the following recognition and matching steps. Is reflected.
  • the procedure indicated in blue is a procedure implemented by a commercial library
  • the procedure indicated in green is a content implementation procedure
  • gray indicates a system area
  • red indicates a user's emotion and facial expression during a video call.
  • This is a procedure implemented by a technology for developing smartphone application content in which an avatar acting on a corresponding emotion and an emoticon adding an effect to a corresponding expression are matched.
  • the present invention develops core technologies for face recognition, emotion recognition, video registration, and video communication necessary for emotion and facial expression matching support video call service, and develops application content and interworking server.
  • avatars and emoticons for the user are produced and applied to the smartphone application.
  • the emotion content generated by the device and method for generating the emotion content through the emotion recognition device and method for providing a content service utilizes the creative, future-oriented value and smart image of the character, a new theme, a new trend. It is used as content in mobile video call.
  • the emotional content according to the present invention enhances the educational utility as a differentiated theme using advanced smart devices that will change the future life, and maximizes artistic value through differentiated content storytelling and the shaping of universal values.
  • it maximizes the differentiation factor that 3D animation is possible not only through video player but also through video call with others.
  • the emotional content according to the present invention can be helpful for the emotional development of our children by inducing a didactic value and various facial expression changes that our lives are happy and precious through smart video calls that accompany us and always communicate in everyday life.
  • the visual image of the character according to the present invention constructs a character image that anyone can like and accept through the composition of figures 3-4.
  • Figure 5 is a table summarized the embodiments of the main characters of storytelling according to the present invention, understanding the importance of living and breathing with me, the importance of universal values and communication for imagination and adventure, friendship and family It shows the main characters, their images, and their personalities and roles to convey them.
  • the background story structure of the embodiment of FIG. 5 deals with a left-right doldol episode that occurs while the main character 'Ava' (user) meets a race 'Bata' in a virtual space living in his mobile phone.
  • the protagonist is a user who uses a video call
  • the character 'Bata' personified in the virtual space that is the friend of the protagonist, is a friend of mine who can meet at any time when the protagonist makes a video call with the other party.
  • the emotion content according to the present invention in a 3D object in a video communication terminal such as a smartphone, it is expected that the processing time for image processing from face detection to recognition and emotion recognition during a video call is large, and also recognized information. On the basis of this, delays are expected to occur in expressing matching between video and 3D objects. Therefore, in terms of 3D matching, it is possible to consider the method of minimizing the matching delay time by using the 3D engine. For example, the development speed can be improved by using the Unity3D engine in the 3D object representation method using OpenGL ES.
  • FIG. 6 is a diagram illustrating a service movement scenario according to user movement between access networks to which the emotion content service method according to the present invention is applied, and a method for solving a communication network access problem in a mobile environment through a smart phone may be sought. .
  • face detection is to find a location of a face in an image.
  • a face of a person is determined based on an angle of the front or side according to the gazing direction, the degree of tilting the head from side to side, various expressions, and distance from the camera.
  • the image may vary depending on external changes such as morphological changes such as the size of the face image, differences in brightness levels within the face due to lighting, complex backgrounds, or other objects of indistinguishable color from the face. Face detection studies from MM include many difficulties.
  • Face detection is a pre-processing step before face recognition, such as knowledge-based methods, feature-based methods, template-matching methods, and appearance-based methods. Divided by. This is summarized in Table 1 below.
  • Knowledge-based Face Detection Methods is a method in which the face of the human face, such as eyebrows, eyes, nose, mouth, each face component using a constant distance and positional relationship to each other. In this method, partial contrast is concentrated in the center area of the face image, and the contrast distribution of the face image and the image is detected through comparison, and a top-down approach is mainly used.
  • the knowledge-based detection method has a disadvantage in that it is difficult to detect a face in an image having various changes in the face, such as a tilt of a face, an angle of looking at a camera, and an expression, so that it can be applied only in a special case.
  • Feature-based Face Detection Methods are a combination of the size and shape of the facial feature components (eye, nose, mouth, outline, and contrast), their correlations, the color and texture information of the face and the components It is a method of detecting a face using shape information.
  • the bottom-up approach is used to find partial features of the face and to integrate the candidate regions (face specific components) to find the face.
  • Feature-based face detection methods have the advantage of not being sensitive to poses or face orientations, because the processing time can be found quickly and easily. However, it can be mistaken for a background or an object similar to the skin color, and the color and texture information of the face may be lost as the brightness of the light changes. In addition, there is a disadvantage that can not detect the feature components of the face according to the degree of inclination of the face.
  • Template matching-based Face Detection Methods create a standard template for all of the target faces and then detect the face by comparing the similarity with the input image. There are algorithms and variant template algorithms.
  • the template matching-based face detection method generates information using partial regions or outlines from the prepared image data, and then transforms the generated information through algorithms to increase the amount of similar information and use it for face detection.
  • the template matching-based face detection method is sensitive to the change in the size of the face according to the distance, the rotation angle and the tilt of the face according to the gaze direction, and it is difficult to define templates for different poses like the knowledge-based method.
  • An appearance-based method is a method of detecting a face using a model trained by a set of training images using pattern recognition.
  • Appearance-based methods are one of the most used methods in the face detection field, and are based on eigenface, linear discriminant analysis (LDA), and neural network generated by Principal Component Analysis (PCA). NN), Adaboost, and Support Vector Machines (SVMs).
  • Appearance-based methods use the existing face and non-face learning data groups to detect face regions in complex images and generate learned eigenvectors to find faces. This method has the advantage of high recognition rate because the constraints mentioned in other detection methods are overcome by learning.
  • appearance-based methods such as PCA, NN, and SVM require a lot of time to learn the database, and also have a disadvantage of having to learn again when the database changes.
  • face recognition technology is a method used to identify a face after detecting a face through a multimedia image. Face recognition technology can be classified as shown in Table 1 below, in the present invention is used to identify the components of the face.
  • the input of the face recognition system uses the entire face area.
  • the holistic face recognition method has an advantage that can be easily implemented, but it does not take enough detail of the face, so it is difficult to obtain sufficient results.
  • Holistic face recognition methods include principal component analysis (PCA), linear discriminant analysis (LDA), independent component analysis (ICA), tensor face and probabilistic decision-based neural networks (PDBNN).
  • Feature-based methods first extract spatial features (eyes, nose and mouth), and then the location and spatial characteristics (geometry and appearance) of the spatial features are input to the recognition system.
  • the feature-based method is quite complicated because there is a variety of feature information on the face, so it is necessary to determine how to select the best features to improve face recognition performance.
  • the typical feature-based methods such as Pure Geometry, Dynamic Link Architecture, and Hidden Markov model, have much better performance than the above holistic matching method. Are utilized.
  • Hybrid methods are very complicated because they use the entire face area to recognize a face along with the location characteristics, but the recognition rate is much superior to the holistic matching and feature-based matching methods.
  • Hybrid methods include linear feature analysis (LFA), shape-normalization, and component-based methods.
  • FIG. 8 is a diagram illustrating a configuration and operation (Req / Resp) sequence for a SIP service
  • FIG. 9 is a diagram illustrating a basic message and status code scheme of SIP
  • FIGS. 10 and 11 are a basic SIP protocol stack and call setup.
  • SIP is a protocol for managing sessions or calls in multimedia communication, and is a technique that focuses on multimedia communication management through signaling rather than multimedia data transmission itself.
  • Table 3 summarizes the components of SIP service and its main functions.
  • a caller sends an INVITE request message for creating a session to a callee. These messages go through several SIP servers to be delivered to the receiver.
  • the received proxy server parses the message to recognize the recipient and delivers the received message to the appropriate proxy server or recipient's user agent (UA).
  • UA user agent
  • the receiver receiving the INVITE message sends a response message to the INVITE message.
  • the response message has a status code indicating the result of processing. If the receiver receives and processes the message correctly, it sends a “200 OK” response message to the sender.
  • the sender who receives the response sends a ACK request message back to the receiver to inform the receiver that the response message is correctly received.
  • the wired / wireless convergence service environment is an environment in which the mobility of terminals is generalized, and various types of access networks are selected based on criteria such as service quality and user preferences, rather than access network access in a simple sense for existing communication access. It is evolving into a mobile terminal environment between heterogeneous networks that connect and communicate with each other. Therefore, in order to access a terminal between heterogeneous networks, a mobility support technology between heterogeneous networks is required, and a function for mobility support technology must be mounted in the terminal.
  • a multi-mode terminal having a plurality of communication interfaces to access networks to which the terminals can be connected is required.
  • multi-mode terminals The need for more is growing.
  • the current approach takes the form of changing the communication mode for connection to a heterogeneous network, which requires a reset of the terminal's power and services.
  • an automatic access control technique between heterogeneous networks is required in a multimode terminal in which handover between heterogeneous networks is automatically controlled without user terminal setting and service disconnection.
  • FIG. 12 is a conceptual diagram of a content matching system according to an embodiment of the present invention.
  • an avatar is recognized to a counterpart by recognizing a user's face and emotion during a video call through an Avatar video call program included in each user's terminal.
  • the facial expressions and emoticons are matched on the screen and transmitted.
  • the user performs a continuous face recognition and emotion during the call. At this time, more effective and enjoyable video call with the other party is possible through avatar matching and emoticon matching with perceived emotion or facial expression.
  • Basic video call is made through SIP-based video conference and performs data transmission and reception using RTP / RTCP.
  • the switch to HTTP streaming may be considered.
  • FIG. 13 is a diagram illustrating a basic operation procedure for image registration according to the present invention.
  • an avatar and a video are matched.
  • image input is performed through a camera module of a smart phone (terminal) to perform image preprocessing for face recognition, facial expression recognition, and emotion recognition.
  • it extracts possible face candidates and analyzes the components of the face to extract information for posture estimation and emotion recognition.
  • the facial expression and motion of the avatar are selected, the position in the 3D space is calculated, matched with the video, and displayed on the screen. It also encodes this image and sends it over the network to a remote video call smartphone.
  • Step 1 is a step of capturing a frame image to be analyzed in a video data source from a camera module
  • step 2 is a analysis of the captured image.
  • the preprocessing is performed in an easy-to-use state, so that the boundary between objects in the image can be grasped by an edge detection algorithm or the like.
  • step (3) detects a face from the preprocessed image. Analyzes facial feature points by applying a learning algorithm, extracts the position and relationship data on the image of the component, and step (4) uses the face detection step. At the stage of posture estimation and facial expression recognition based on the extracted recognition information, the attitude and facial expression of the avatar regarding the posture and facial expression taken by the face are selected. Next, step 5 determines the avatar (face) position coordinates in the 3D space through the analyzed information, selects an avatar animation for the expression and emotion, and transmits a control signal (message) to the 3D engine.
  • a control signal messagessage
  • step 6 the 3D space in which the avatar image and video are represented is composed, and the analyzed avatar is placed in the corresponding position (controlling the avatar and 3D space through the 3D engine API).
  • the avatar and the video source to be represented are matched to a single image source, and in step 8, the video is encoded along with the voice source.
  • the audio source is extracted from the video source and processed.
  • the network is sent to the other terminal configured for the video call.
  • the session is configured through the SIP and transmitted through the Internet through the RTP / RTCP.
  • FIG. 15 illustrates an avatar video communication operation process through emotion recognition and image registration of FIG. 14 using actual content.
  • face tracking and eye, nose, and mouth are recognized from the captured image, and through this, standard analysis and relationship technology are applied, emotional inference and real-time matching to match a virtual model to the face of the real world.
  • standard analysis and relationship technology are applied, emotional inference and real-time matching to match a virtual model to the face of the real world.
  • a real-time avatar is implemented on the user's face.
  • an apparatus and method for emotion content service of a communication terminal device As described above, an apparatus and method for emotion recognition therefor, an apparatus and method for generating and matching emotion content using the same, and whether the emotion content is voice recognition or an object Facial region detection technology of face recognition technology, facial region normalization technology, feature extraction technology in face region, facial component (expression analysis) relationship technology of emotion recognition technology of object recognition, object hand gesture, object recognition action and
  • object Facial region detection technology of face recognition technology facial region normalization technology
  • feature extraction technology in face region feature extraction technology in face region
  • facial component (expression analysis) relationship technology of emotion recognition technology of object recognition object hand gesture
  • object recognition action On the basis of behavioral cognitive technology, real-time matching technology of real-life and virtual images is used to match mixed virtual objects (including characters) through gesture and facial expression analysis on both faces and bodies making video calls, which cannot be seen in the real world.
  • Mixed reality is realized through video call.
  • an apparatus and method for emotion contents service of a communication terminal device includes voice, a specific expression of a face and a body If the similar voice, facial expressions and gestures are transmitted through the video, the virtual objects responding to the voices, facial expressions and gestures are matched on the face and body in real time. It is a wonderful way to enjoy your video calls.
  • an apparatus and method for emotion content service of a communication terminal device may be expressed through a mobile device.
  • a system that recognizes gestures and expresses them through avatars, it will become a foundation to take a leap forward in the cutting-edge video industry, such as domestic film, animation, and cyber characters, and bring human emotions and facial expressions to third parties (Avatar). It will greatly contribute to enhancing the competitiveness of domestic mobile contents and video contents industry by making the process of expressing 3D virtual objects on the face of the real world more natural.
  • FIG. 16 is a diagram illustrating various fields to which the present invention is applicable, and it is expected that the protagonist and the viewer of the movie may communicate in the future due to the spread of the smart device, and the main character of the camera installed on the upper part of the device. It can be applied to the high-tech cultural industry that can provide diversity to see the same video but have different experiences by showing the appropriate response and gestures through the recognition.
  • the present invention is a script-based expression and gesture expression technology used in movies, animation, cyber characters, interface design using user emotion reaction, facial recognition for security and surveillance, consumer's emotional response measurement for products and designs
  • the field of application is endless. Therefore, the scope of the present invention should not be limited to the described embodiments, but should be determined not only by the scope of the following claims, but also by the equivalents of the claims.

Landscapes

  • Engineering & Computer Science (AREA)
  • Business, Economics & Management (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Resources & Organizations (AREA)
  • Tourism & Hospitality (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Multimedia (AREA)
  • Strategic Management (AREA)
  • Signal Processing (AREA)
  • Marketing (AREA)
  • Health & Medical Sciences (AREA)
  • Economics (AREA)
  • General Business, Economics & Management (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Primary Health Care (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Processing Or Creating Images (AREA)
  • Telephonic Communication Services (AREA)
  • Telephone Function (AREA)

Abstract

The present invention relates to augmented reality providing a user with mixed data from a real environment and a virtual environment. More particularly, the present invention relates to an apparatus and method for emotional content services on telecommunication devices, to an apparatus and method for emotion recognition therefor, and to an apparatus and method for generating and matching the emotional content using same, which may utilize a real-time matching technique of matching real images and virtual images on the basis of voice recognition, a technique of extracting facial characteristics, a technique of facial normalization, a technique of facial detection for a facial recognition technique for object recognition, a facial feature (expression analysis) relationship technique for an emotion recognition technique for object recognition, a hand motion recognition technique for object recognition, and a motion and behavior recognition technique for object recognition, and which may match mixed virtual objects (including characters) to the faces and bodies of both sides making a video call using gesture and expression analyses to implement various mixed realities, which cannot be seen in the real world, for the video call.

Description

[규칙 제26조에 의한 보정 16.11.2011] 통신단말장치의 감정 컨텐츠 서비스 장치 및 방법, 이를 위한 감정 인지 장치 및 방법, 이를 이용한 감정 컨텐츠를 생성하고 정합하는 장치 및 방법[Correction 16.11.2011] according to Rule 26. Apparatus and method for emotion content service of communication terminal device, apparatus and method for emotion recognition therefor, apparatus and method for generating and matching emotion content using same
본 발명은 송신 통신단말장치의 촬상장치로부터 인식된 감정과 표정을 분석하고, 분석된 감정이나 의사전달을 효과적으로 전달할 수 있도록 수신 통신단말장치의 화면에 가상의 객체를 실시간 혼합시켜 구현하기 위한 감정 컨텐츠 서비스 장치 및 방법과, 상기 감정 컨텐츠를 서비스하기 위한 감정 인지 장치 및 방법과, 상기 감정 인지를 통해 상기 감정 컨텐츠를 생성하고 정합하는 장치 및 방법과, 상기 감정 컨텐츠를 생성하는 장치 및 방법에 관한 것이다.The present invention analyzes the emotions and facial expressions recognized by the image pickup apparatus of the transmitting communication terminal device, and the emotion contents for realizing mixing the virtual object on the screen of the receiving communication terminal device in real time so as to effectively deliver the analyzed emotion or communication. Service apparatus and method, an emotion recognition apparatus and method for serving the emotion content, an apparatus and method for generating and matching the emotion content through the emotion recognition, and an apparatus and method for generating the emotion content .
최근의 국내외 IT분야의 핵심키워드는 스마트 디바이스(Smart Device)와 증강현실(Augmented Reality), 그리고 콘텐츠(Contents)로 요약된다. 이 중, 스마트 디바이스는 애플사의 아이폰을 통해 본격적으로 확산된 스마트폰이 대표적인 장치이다. 스마트폰은 가입자 수가 2011년 3월을 기점으로 천만명을 돌파할 정도로 단기간에 빠른 성장세를 보이고 있다. 한국 IDC 조사자료에 의하면 올해 국내 태블릿PC시장은 90~100만대 정도로 예측하고 있으며, 아이패드의 경우 35~45%의 비중을 차지할 것으로 예상된다. Recent key keywords in domestic and international IT are summarized as Smart Device, Augmented Reality, and Contents. Among them, smart devices are smartphones that have spread in earnest through Apple's iPhone. Smartphones are growing rapidly in the short term, with subscribers surpassing 10 million since March 2011. According to the Korea IDC survey data, the domestic tablet PC market is expected to be 90 ~ 1 million units this year, and iPad is expected to account for 35 ~ 45%.
IT분야의 핵심키워드 중 두 번째인 증강현실은 가상현실의 한 분야에서 파생된 기술이며, 현실세계와 가상의 체험을 결합하는 기술을 의미한다. 증강현실은 미래를 이끌 10대 혁신기술로 꼽히고 있으며, 현실 세계를 바탕으로 사용자가 가상의 물체와 상호작용 함으로써 향상된 현실감을 부여하는 기술이다.Augmented reality, which is the second key keyword in the IT field, is a technology derived from one field of virtual reality and refers to a technology that combines the real world and the virtual experience. Augmented reality is regarded as one of the top 10 innovations to lead the future, and it is a technology that gives the user a better sense of reality by interacting with virtual objects based on the real world.
증강현실은 가상현실의 한 분야로서 실제환경에 가상의 사물을 합성하여 원래의 환경에 존재하는 사물처럼 보이도록 하는 컴퓨터 그래픽 기법으로서, 증강현실은 가상의 공간과 가상의 사물만을 대상으로 하는 기존의 가상현실과 달리 현실세계의 기반에 가상의 사물을 합성하여 현실세계만으로는 얻기 어려운 부가적인 정보들을 보강해 제공할 수 있는 기술이다. 현재 증강현실 기술은 방송, 광고, 전시, 게임, 테마 파크, 군용, 교육 및 프로모션 등의 분야에서 다양한 형태로 사용되고 활발하게 개발되고 있는 상태이다.Augmented reality is a field of virtual reality, a computer graphics technique that synthesizes virtual objects in the real environment and looks like objects existing in the original environment.Augmented reality is an existing method that targets only virtual space and virtual objects. Unlike virtual reality, it is a technology that synthesizes virtual objects on the basis of real world and reinforces and provides additional information that is difficult to obtain in real world alone. Currently, augmented reality technology is being actively used in various forms in the fields of broadcasting, advertising, exhibition, games, theme parks, military, education and promotion.
즉, 증강현실은 실제세계와의 상호작용을 배제하고 오직 미리 구축해 놓은 가상공간 내에서의 상호작용을 처리하는 가상현실(Virtual Reality) 기술과 달리, 실시간 처리를 바탕으로 하여 미리 획득된 실제세계에 관한 정보가 단말기를 통해 입력되는 실제세계에 대한 영상에 겹쳐서 표시되어 실제세계와의 상호작용을 가능케 한다는 점에서 컴퓨터에 의해 생성된 영상만을 제공하는 가상현실과 구분된다.In other words, augmented reality differs from the virtual reality technology, which excludes interaction with the real world and processes interactions only in the pre-established virtual space, based on real-time processing. The information is overlaid on the image of the real world input through the terminal, and thus it is distinguished from the virtual reality that provides only the image generated by the computer in that it enables interaction with the real world.
이러한 증강현실 기술은 특히 통신 단말기에서 사용되는 모바일 증강현실 기술분야에서 각광받고 있는 추세로서, 현재 마커 기반의 모바일 증강현실 기술 또는 센서 기반의 모바일 증강현실 기술에 많은 연구와 투자가 이루어지고 있다. 마커 기반의 모바일 증강현실 기술은 특정 건물을 촬영할 때, 특정 건물과 대응되는 특정 기호를 같이 촬영한 후 특정 기호를 인식하여 해당하는 건물을 인식하는 기술이며, 센서 기반의 모바일 증강현실 기술은 단말기에 탑재된 GPS와 전자 나침반(Digital Compass) 등을 이용하여 단말기의 현재 위치와 바라보고 있는 방향을 유추하여 유추된 방향으로 영상에 해당하는 POI(Point of Interests) 정보를 오버레이(Overlay) 시켜주는 기술이다. Such augmented reality technology has been in the spotlight in the field of mobile augmented reality technology used in communication terminals in particular, a lot of research and investment in the mobile augmented reality technology based on a marker or a mobile augmented reality technology is currently being made. Marker-based mobile augmented reality technology is a technology for recognizing a building by recognizing a specific sign after taking a specific sign corresponding to a specific building when shooting a specific building. It is a technology that overlays POI (Point of Interests) information corresponding to the image in the inferred direction by inferring the current position of the terminal and the viewing direction by using the installed GPS and digital compass. .
이러한 종래의 기술들은 서비스 제공자가 미리 지정해 놓은 건물이나 장소에 대한 정보만을 제공하는 것이 일반적이어서 사용자가 서비스 제공자에 의해 지정되지 않은 객체에 대한 적절한 정보를 제공해 주는 것이 불가능하고, 현재 위치와 단말기가 바라보고 있는 방향을 유추할 뿐 단말기를 통해 입력된 영상을 정확하게 인식하는 기술을 제공하지 못하고 있기 때문에, 현재 대부분의 연구가 획득한 영상 내에 존재하는 실제 물체를 정확하게 인식하고 해당 물체의 지역정보를 매핑하여 직관적이고 편리한 영상인식 기반의 증상현실을 제공하고자 하는 연구나, 단말기를 통해 입력되는 입력영상에 포함된 객체의 위치에 상기 객체의 상세정보에 접근 가능하도록 하기 위한 아이콘을 증강현실의 형태로 디스플레이하여 사용자가 편리하게 관심객체의 위치를 인지하고 해당 관심객체의 상세정보에 접근할 수 있도록 하는 등의 연구 등 제공하는 정보의 정확도와 양적인 확장에 연구가 국한되고 있다.These conventional techniques generally provide only information about a building or a place previously designated by a service provider, and thus it is impossible for a user to provide appropriate information about an object not designated by the service provider. Since it does not provide the technology to accurately recognize the image input through the terminal by inferring the viewing direction, most current researches accurately recognize the actual object existing in the acquired image and map the local information of the object. In order to provide an intuitive and convenient image recognition-based symptom reality, or to display the icon in the form of augmented reality to access the detailed information of the object in the position of the object included in the input image input through the terminal The user can conveniently Research is limited to the accuracy and quantitative expansion of information provided, such as research on location and access to detailed information of the object of interest.
따라서 증강현실 기술의 개발에 편중되지 않고 증강현실 기술의 발전과 더불어 일상생활에서 통신 단말기를 사용하는데 있어서 증강현실 기술을 통해 사용자들에게 즐거움을 줄 수 있는 다양한 어플리케이션의 개발이 요망된다.Therefore, the development of augmented reality technology, rather than the development of augmented reality technology in the use of a communication terminal in everyday life, it is desired to develop a variety of applications that can give users a pleasure through augmented reality technology.
IT분야의 핵심키워드 중 세 번째인 3D 영상 콘텐츠는 제임스카메론 감독의 '아바타'로 인해 관련 산업이 폭발적으로 증가하고 있으며, 영상통화도 3D컨텐츠와 함께 즐기는 시기가 도래할 것으로 예상된다. 3D video content, the third key keyword in the IT field, is exploding in related industries due to James Cameron's 'Avatar', and it is expected that the time for enjoying video calls with 3D content will come.
정보통신정책연구원의 '모바일 애플리케이션의 동향과 전망'에 따르면 Telecom Asia는 2011년에 유행할 5가지 앱 트렌드로 자동차, 모바일 영상통화, 소셜 미디어, 증강현실, 성인용 콘텐츠를 선정하였고, 마케팅 리서치 회사 In-stat 발행보고서에 의하면 모바일 영상통화의 매출규모는 2015년까지 10억 달러 돌파를 예상하고 있으며, IDC는 올해 전세계 모바일 앱 다운로드 건수가 109억 건에 달할 것이고, 2014년에는 769억 건으로 가파르게 늘어날 뿐만 아니라, 모바일 앱(Mobile App) 매출 규모도 2014년에 350억 달러를 넘어설 것으로 전망하는 등 앱 생태계가 앞으로 10년은 거뜬히 이어질 것이라는 전망이 쏟아지고 있다.According to the 'Information and Trends of Mobile Applications' of the Korea Institute of Information and Telecommunications Policy, Telecom Asia selected automotive, mobile video calls, social media, augmented reality and adult content as the top five app trends in 2011. According to the -stat issuance report, mobile video call sales are expected to exceed $ 1 billion by 2015. IDC expects worldwide mobile app downloads to reach 10.9 billion this year and steeply climbed to 76.9 billion by 2014. In addition, mobile app sales are expected to exceed $ 35 billion in 2014, and the app ecosystem is expected to continue for the next 10 years.
우리나라와 달리 해외 스마트폰은 영상통화 서비스 보급이 더딘 편이었으나, 애플사의 아이폰(iPhone 4)이 FaceTime이라는 서비스를 강조하면서 영상통화에 대한 관심이 부쩍 커져가고 있는 실정이다. Unlike in Korea, overseas smartphones have been slow in distributing video call services, but as Apple's iPhone (iPhone 4) emphasizes the service called FaceTime, interest in video calls is growing.
또한, 삼성전자의 Galaxy S I?II를 비롯한 최근 안드로이드(Android)폰들이 영상통화를 지원하고 있으며, 안드로이드 2.3버전(Gingerbread)가 공식적으로 영상통화를 지원하기 때문에 2011년 중반부터 안드로이드 계열의 폰들은 영상통화 기능을 기본탑재 할 것으로 전망된다. 여기에 안드로이드 기반의 테블릿(Tablet) 컴퓨터나 아이패드(iPad 2)세대 역시 전면과 후면 카메라를 통한 영상통화 서비스 제공이 확실해 보이기 때문에 영상통화가 일반화될 가능성이 높아지고 있다.In addition, Android phones such as Samsung's Galaxy SI II are supporting video calls, and since Android version 2.3 (Gingerbread) officially supports video calls, Android phones from mid-2011 have been able to support video calls. It is expected to be equipped with call function. In addition, Android-based Tablet computers and iPad 2 generations are also expected to provide video call service through front and back cameras, so video calls are becoming more common.
또한, 4G 서비스가 본격화되면 음성과 데이터 서비스 외에 1:1 혹은 다자간 영상통화가 핵심 서비스로 성장할 가능성도 있으며, 국내와 달리 해외에서의 휴대 단말기를 통한 영상통화에 대한 관심은 점차 상승하는 추세이다. In addition, if 4G services become full-fledged, 1: 1 or multi-party video calls may grow into core services in addition to voice and data services. Unlike domestic markets, interest in video calls through mobile terminals is gradually increasing.
한편, SNS(Social networking service )분야에서는 음성 및 영상통화서비스가 유행처럼 번지고 있으며, 트위터와 자자의 결합서비스인 음성통화서비스 역시 이러한 연장선상에 있다. 또한 앞으로 온라인상에서 음성과 영상을 활용한 커뮤니케이션 서비스는 더욱 증가할 것으로 보인다. 따라서 소셜네트워크 서비스 업체 및 통신 업체들의 서비스 트랜드는 증강현실과 영상통화 그리고 특화된 서비스가 결합된 방향으로 점차 이동해 갈 것으로 예상된다. On the other hand, in the field of social networking services (SNS), voice and video call services are spreading like fashion, and voice call service, which is a combination service of Twitter and its own, is also in line with this extension. In addition, communication services using voice and video are expected to increase further online. Therefore, the service trends of social network service providers and telecommunications companies are expected to gradually move toward a combination of augmented reality, video call, and specialized services.
이에 영상통화 서비스의 보급을 한 계 진화시킬 수 있는 감정인식과 증강현실기법을 활용한 신개념 서비스를 위한 관련 기술의 개발 및 이를 이용한 서비스의 개발이 절실히 요망된다. Therefore, there is an urgent need for the development of related technologies for new concept services using emotional recognition and augmented reality techniques that can further evolve the spread of video call services, and the development of services using them.
따라서 본 발명의 목적은 영상통화 시에 통화자 쌍방의 감정상태를 표현하는 가상의 객체를 통화영상과 함께 제공함으로써 영상통화에 재미를 선사할 수 있는 감정 컨텐츠 서비스 장치 및 방법과, 상기 감정 컨텐츠를 서비스하기 위한 감정 인지 장치 및 방법과, 상기 감정 인지를 통해 상기 감정 컨텐츠를 생성하는 장치 및 방법과, 상기 감정 컨텐츠를 생성하는 장치 및 방법에 의하여 생성된 상기 감정 컨텐츠를 제공하는데 있다. Accordingly, an object of the present invention is to provide an emotional content service apparatus and method that can provide fun to a video call by providing a virtual object representing the emotional state of both callers with the video during a video call, and the emotional content. An emotion recognition apparatus and method for service, an apparatus and method for generating the emotion content through the emotion recognition, and the emotion content generated by the apparatus and method for generating the emotion content.
또한, 본 발명의 다른 목적은 통화자의 감정상태를 가상의 객체를 통해 통화자의 영상에 중첩시킴으로써, 통화자들의 영상통화에 더욱 현실감을 부여하는 증강현실을 경험할 수 있도록 하는 감정 컨텐츠 서비스 장치 및 방법과, 상기 감정 컨텐츠를 서비스하기 위한 감정 인지 장치 및 방법과, 상기 감정 인지를 통해 상기 감정 컨텐츠를 생성하는 장치 및 방법과, 상기 감정 컨텐츠를 생성하는 장치 및 방법에 의하여 생성된 상기 감정 컨텐츠를 제공하는데 있다. In addition, another object of the present invention is to provide an emotional content service apparatus and method for superimposing the emotional state of the caller to the video of the caller through a virtual object, to experience augmented reality that gives the callers a more realistic feeling and And an emotion recognition apparatus and method for servicing the emotion content, an apparatus and method for generating the emotion content through the emotion recognition, and the emotion content generated by the apparatus and method for generating the emotion content. have.
또한, 본 발명은 송신 통신단말장치의 촬상장치로부터 인식된 감정과 표정을 분석하고, 분석된 감정이나 의사전달을 효과적으로 전달할 수 있도록 수신 통신단말장치의 화면에 가상의 객체를 실시간 혼합시켜 구현하는 감정 컨텐츠 서비스 장치 및 방법을 제공하는데 있다.In addition, the present invention analyzes the emotions and facial expressions recognized from the image pickup device of the transmitting communication terminal device, and the emotion that is implemented by real-time mixing a virtual object on the screen of the receiving communication terminal device to effectively deliver the analyzed emotion or communication An apparatus and method for providing content services are provided.
또한, 본 발명은 송신 통신단말장치의 촬상장치로부터 인식된 감정과 표정을 분석하고, 분석된 감정이나 의사전달을 효과적으로 전달할 수 있도록 수신 통신단말장치의 화면에 가상의 객체를 실시간 혼합시켜 구현하는 감정 컨텐츠 서비스를 제공하기 위한 감정 인지 장치 및 방법을 제공하는데 있다.In addition, the present invention analyzes the emotions and facial expressions recognized from the image pickup device of the transmitting communication terminal device, and the emotion that is implemented by real-time mixing a virtual object on the screen of the receiving communication terminal device to effectively deliver the analyzed emotion or communication The present invention provides an apparatus and method for emotion recognition for providing a content service.
또한, 본 발명은 송신 통신단말장치의 촬상장치로부터 인식된 감정과 표정을 분석하고, 분석된 감정이나 의사전달을 효과적으로 전달할 수 있도록 수신 통신단말장치의 화면에 가상의 객체를 실시간 혼합시켜 구현하는 감정 컨텐츠 서비스를 제공하기 위한 감정 인지 장치 및 방법을 통해 상기 감정 컨텐츠를 생성하는 장치 및 방법을 제공하는데 있다. In addition, the present invention analyzes the emotions and facial expressions recognized from the image pickup device of the transmitting communication terminal device, and the emotion that is implemented by real-time mixing a virtual object on the screen of the receiving communication terminal device to effectively deliver the analyzed emotion or communication An apparatus and method for generating the emotion content through an emotion recognition apparatus and method for providing a content service are provided.
또한, 본 발명은 송신 통신단말장치의 촬상장치로부터 인식된 감정과 표정을 분석하고, 분석된 감정이나 의사전달을 효과적으로 전달할 수 있도록 수신 통신단말장치의 화면에 가상의 객체를 실시간 혼합시켜 구현하는 감정 컨텐츠 서비스를 제공하기 위한 감정 인지 장치 및 방법을 통해 상기 감정 컨텐츠를 생성하는 장치 및 방법에 의하여 생성하는 상기 감정 컨텐츠를 제공하는데 있다.In addition, the present invention analyzes the emotions and facial expressions recognized from the image pickup device of the transmitting communication terminal device, and the emotion that is implemented by real-time mixing a virtual object on the screen of the receiving communication terminal device to effectively deliver the analyzed emotion or communication The present invention provides an emotional content generated by an apparatus and a method for generating the emotional content through an emotion recognition apparatus and a method for providing a content service.
상기와 같은 목적들을 달성하기 위한 본 발명은 감정을 분석하여 특정 감정에 대해 아바타가 대신 표현하도록 하는 아바타 정합과, 표정을 분석하여 특정 표정에 대한 효과를 첨가하는 이모티콘 정합 및 특정 감정에 대한 표정을 나타내는 얼굴이나 신체의 특정부위를 과장하는 특징을 적어도 하나 이상 갖는다.The present invention for achieving the above object is to analyze the emotion to match the avatar to the avatar instead of a specific emotion to analyze, and the emoticon matching to add the effect on the specific expression by analyzing the expression and the expression for a particular emotion It has at least one feature which exaggerates the specific part of the face or body which it represents.
본 발명은 사용자의 감정과 표정을 인식하기 위한 얼굴검출, 얼굴인식, 감정인식 핵심 기술을 제안하고, 이에 따라 인식된 감정을 대행하는 아바타 및 표정을 극대화하는 이모티콘을 생성하고 정합하는 기술을 제안하며, 이를 이용한 스마트폰을 통한 화상통화 서비스 기술을 제안한다. The present invention proposes a face detection, face recognition, and emotion recognition core technology for recognizing a user's emotions and expressions, and accordingly proposes a technology for generating and matching emoticons maximizing avatars and expressions for the recognized emotions. In this paper, we propose a video call service through a smartphone.
본 발명은 표정변화를 인지하고 해당콘텐츠(아바타)를 사람의 실제 얼굴위에 정합함으로써, 표정인식을 통해 현실세계에 불가능한 감정표현을 가능하여 의사전달의 효과를 증대시킬 수 있다. The present invention can increase the effect of communication by recognizing a change in facial expression and matching the corresponding content (avatar) on a person's real face, thereby enabling expression of emotions that are impossible in the real world through facial recognition.
상기와 같은 목적들을 달성하기 위한 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스 장치 및 방법, 이를 위한 감정 인지 장치 및 방법, 이를 이용한 감정 컨텐츠를 생성하고 정합하는 장치 및 방법은 음성인식, 객체인지의 얼굴인지기술의 얼굴영역검출기술, 얼굴영역 정규화 기술, 얼굴영역 내 특징추출기술, 객체인지의 감정인지기술의 얼굴구성요소(표정분석)관계기술, 객체인지의 손동작인지, 객체인지의 동작과 행동인지기술 기반 위에 실사와 가상영상의 실시간 정합기술을 활용하여 영상통화를 하는 쌍방의 얼굴과 몸체 위에 제스처 및 표정분석을 통한 혼합된 가상의 오브젝트(문자포함)를 정합시켜 현실세계에서 볼 수 없는 다양한 혼합현실을 영상통화를 통해 구현하는 것을 특징으로 한다.Apparatus and method for emotion content service of a communication terminal device according to the present invention for achieving the above objects, apparatus and method for emotion recognition for the same, apparatus and method for generating and matching emotion content using the same are voice recognition, object recognition Face area detection technology of face recognition technology, face area normalization technology, feature extraction technology within face area, facial component (expression analysis) relationship technology of emotion recognition technology of object recognition, object or hand gesture, object recognition behavior and behavior On the basis of cognitive technology, the real-time matching technology of the real picture and the virtual image is used to match the mixed virtual objects (including characters) through the gesture and facial expression analysis on both the face and the body to make a video call. Characterized in that the mixed reality is implemented through a video call.
또한, 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스는 본 발명은 음성, 얼굴과 몸체의 특정한 표정과 제스처의 표정분석 관계연산함수를 미리 등록하여, 유사한 음성, 표정과 제스처가 영상을 통해 전송되면, 출력되는 영상화면에서 음성, 표정과 제스처에 반응한 가상의 객체를 얼굴과 몸체 위에 실시간으로 정합시켜, 영상통화를 즐기도록 하는데 특징이 있다.In addition, the emotional content service of the communication terminal device according to the present invention is registered in advance the expression analysis relation function of the specific expression and gesture of the voice, face and body, if a similar voice, facial expression and gesture is transmitted through the image On the output video screen, the virtual object responding to voice, facial expression, and gesture is matched in real time on the face and body to enjoy a video call.
또한, 상기와 같은 목적들을 달성하기 위한 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스는 촬상수단 및 디스플레이수단을 적어도 구비하는 통신단말장치의 상기 촬상수단 통해 촬영되는 사용자의 제스처 및 표정 중 적어도 어느 하나로부터 상기 사용자의 감정상태를 추출하고, 상기 추출된 감정상태에 대응하는 가상의 객체를 생성하여, 상기 사용자의 바디 및 얼굴 중 적어도 어느 하나 위에 중첩시켜, 상기 사용자와 영상통화 하는 상대방의 통신단말장치의 디스플레이수단에 표시하는 것을 특징으로 한다.In addition, the emotional content service of the communication terminal apparatus according to the present invention for achieving the above objects is at least one of a gesture and facial expression of the user photographed through the image pickup means of the communication terminal device having at least an image pickup means and a display means A communication terminal device of a counterpart that extracts the emotional state of the user from the user, generates a virtual object corresponding to the extracted emotional state, superimposes on at least one of the body and the face of the user, and makes a video call with the user It is characterized by displaying on the display means.
이때, 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스 방법에 있어서 상기 가상의 객체는, 문자를 더 포함하는 것을 특징으로 한다.In this case, in the emotional content service method of the communication terminal device according to the present invention, the virtual object may further include a character.
또한, 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스 방법에 있어서 상기 가상의 객체는, 상기 사용자에 의하여 변경 가능한 것을 특징으로 한다.In the emotional content service method of the communication terminal device according to the present invention, the virtual object may be changed by the user.
또한, 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스 방법에 있어서 상기 가상의 객체는, 상기 감정상태에 대응하여 실시간으로 변화하는 것을 특징으로 한다.In the emotional content service method of the communication terminal device according to the present invention, the virtual object is changed in real time corresponding to the emotional state.
또한, 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스 방법에 있어서 상기 가상의 객체는, 상기 사용자의 바디 및 얼굴에 중첩된 위치가 변경되는 것을 특징으로 한다.In addition, in the emotional content service method of the communication terminal device according to the present invention, the virtual object is characterized in that the position superimposed on the body and face of the user is changed.
또한, 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스 방법에 있어서 상기 통신단말장치가 음성입력수단 및 음성출력수단을 더 구비하면, 상기 음성입력수단을 통해 입력되는 사용자의 음성으로부터 상기 사용자의 감정상태를 더 추출하는 것을 특징으로 한다.In addition, in the emotional content service method of the communication terminal device according to the present invention, if the communication terminal device further comprises a voice input means and a voice output means, the emotional state of the user from the user's voice input through the voice input means Characterized in further extracting.
또한, 상기와 같은 목적들을 달성하기 위한 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스는 촬상수단 및 디스플레이수단을 적어도 구비하는 영상통화서비스 제공단말기에 있어서, 상기 촬상수단 통해 촬영되는 사용자의 제스처 및 표정 중 적어도 어느 하나로부터 상기 사용자의 감정상태를 추출하고, 상기 추출된 감정상태에 대응하는 가상의 객체를 생성하여, 상기 사용자의 바디 및 얼굴 중 적어도 어느 하나 위에 중첩시켜, 상기 사용자와 영상통화 하는 상대방의 영상통화장치의 디스플레이수단에 표시하는 것을 특징으로 한다. In addition, the emotional content service of the communication terminal apparatus according to the present invention for achieving the above object is a video call service providing terminal having at least an imaging means and a display means, the gesture and facial expression of the user to be photographed through the imaging means A counterpart that extracts the emotional state of the user from at least one of the following, generates a virtual object corresponding to the extracted emotional state, and superimposes on at least one of a body and a face of the user to make a video call with the user It is characterized in that displayed on the display means of the video call device.
또한, 상기와 같은 목적들을 달성하기 위한 본 발명에 따른 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스 방법은, 통신단말장치에 사용자의 얼굴 영상을 입력하는 과정; 상기 입력된 얼굴 영상에서 얼굴 구성 요소를 추출하는 과정; 상기 추출된 얼굴 구성 요소를 전처리하는 과정; 상기 전처리된 얼굴 구성 요소에서 얼굴의 특징을 추출하는 과정; 상기 추출된 얼굴의 특징을 얼굴 데이터베이스에 등록하는 과정; 및 상기 얼굴 데이터베이스에 등록된 특징과 상기 특징 추출 과정에서 추출된 얼굴 구성 요소를 비교하여 감정을 인지하는 과정;을 포함하는 것을 특징으로 한다. In addition, the emotional content service method of the communication terminal device according to the present invention for achieving the above object, the process comprising: inputting a face image of the user to the communication terminal device; Extracting face components from the input face image; Preprocessing the extracted facial components; Extracting facial features from the preprocessed facial components; Registering the extracted feature of the face in a face database; And a step of recognizing an emotion by comparing a feature registered in the face database with a face component extracted in the feature extraction process.
또한, 상기와 같은 목적들을 달성하기 위한 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 방법을 통해 상기 감정 컨텐츠를 생성하고 정합하는 방법은, 통신단말장치의 카메라 모듈로부터 사용자의 얼굴 영상을 입력받는 과정; 상기 입력된 얼굴 영상을 전처리하는 과정; 상기 전처리된 얼굴 영상에서 유효 데이터만을 검출하는 과정; 상기 검출된 유효 데이터로부터 얼굴의 위치 및 상기 카메라 정보를 추정하는 과정; 및 상기 카메라 정보와 얼굴의 위치정보로부터 3D 영상을 생성하고, 생성된 상기 3D영상을 상기 사용자의 얼굴 영상에 정합하는 과정;을 포함하는 것을 특징으로 한다. In addition, the method for generating and matching the emotion content through the emotion recognition method for the emotion content service of the communication terminal device according to the present invention for achieving the above object, the face image of the user from the camera module of the communication terminal device Receiving a process; Preprocessing the input face image; Detecting only valid data in the preprocessed face image; Estimating the position of the face and the camera information from the detected valid data; And generating a 3D image from the camera information and the position information of the face, and matching the generated 3D image with the face image of the user.
이때, 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 방법을 통해 상기 감정 컨텐츠를 생성하고 정합하는 방법은, 상기 사용자의 얼굴 영상과 정합된 상기 3D영상을 화면에 출력하는 과정; 및 상기 사용자의 얼굴 영상과 정합된 상기 3D영상을 네트워크를 통해 상대방 통신단말기로 전송하는 과정;을 더 포함하는 것을 특징으로 한다.In this case, the method for generating and matching the emotion content through the emotion recognition method for the emotion content service of the communication terminal device according to the present invention comprises the steps of: outputting the 3D image matched with the face image of the user on the screen; And transmitting the 3D image matched with the face image of the user to a counterpart communication terminal through a network.
또한, 상기와 같은 목적들을 달성하기 위한 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스 방법은, 통신단말장치의 카메라 모듈로부터 동영상 데이터 소스에서 분석할 프레임 이미지를 캡쳐하는 과정; 상기 캡쳐한 이미지를 분석하기 쉬운 상태로 전처리를 하는 과정; 상기 전처리된 이미지에서 얼굴을 검출하는 과정; 상기 얼굴 검출 과정을 통해 추출된 인식 정보를 바탕으로 자세추정과 표정을 인식하여, 얼굴이 취하는 자세와 표정에 대한 아바타의 자세와 표정을 선정하는 과정; 분석된 정보를 통해서 3D 공간상에 상기 선정된 아바타의 위치 좌표를 결정하고, 해당 표정과 감정에 대한 아바타 애니메이션을 선정하며, 3D엔진에 제어 신호를 전달하는 과정; 상기 아바타 영상과 동영상이 표현될 3D공간을 구성하고, 분석된 아바타를 해당 위치에 놓는 기능을 수행하는 과정; 상기 3D공간을 통해 표현되는 아바타와 동영상소스를 단일 이미지 소스로 정합하는 과정; 상기 정합된 이미지를 음성소스와 함께 동영상 인코딩을 수행하는 과정; 및 화상통화를 위해 구성된 네트워크를 통해 상대단말로 상기 정합된 이미지를 전송하는 과정;을 포함하는 것을 특징으로 한다. In addition, the emotional content service method of the communication terminal device according to the present invention for achieving the above object, the process of capturing a frame image to be analyzed in the video data source from the camera module of the communication terminal device; Preprocessing the captured image in an easy to analyze state; Detecting a face in the preprocessed image; Recognizing a posture estimation and a facial expression based on the recognition information extracted through the face detection process, and selecting a posture and facial expression of the avatar with respect to the posture and facial expression taken by the face; Determining position coordinates of the selected avatar on the 3D space through the analyzed information, selecting an avatar animation for a corresponding expression and emotion, and transmitting a control signal to the 3D engine; Composing a 3D space in which the avatar image and the video are to be represented, and performing a function of placing the analyzed avatar at a corresponding position; Matching the avatar and the video source represented through the 3D space into a single image source; Performing video encoding on the matched image together with a voice source; And transmitting the matched image to a counterpart terminal through a network configured for video call.
이때, 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스 방법에 있어서, 상기 전처리된 이미지에서 얼굴을 검출하는 과정은, 학습 알고리즘 등을 적용하여 얼굴 특징점을 분석하고, 구성 요소의 이미지상 위치와 관계 데이터를 추출하는 것을 특징으로 한다. At this time, in the emotional content service method of the communication terminal device according to the present invention, the process of detecting a face in the pre-processed image, applying a learning algorithm or the like to analyze the facial feature point, and the position and relationship data on the image of the component It is characterized in that the extraction.
또한, 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스 방법에 있어서, 상기 네트워크를 통해 상대단말로 상기 정합된 이미지를 전송하는 과정은, SIP를 통해 세션을 구성하고, RTP/RTCP를 통해 인터넷으로 전송하는 것을 특징으로 한다. In addition, in the emotional content service method of the communication terminal device according to the present invention, the step of transmitting the matched image to the other terminal through the network, establishes a session via SIP, and transmits to the Internet via RTP / RTCP Characterized in that.
또한, 상기와 같은 목적들을 달성하기 위한 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스 장치는, 영상통화서비스 제공단말기이기와 연동하는 서버 통신부; 및 상기 영상통화서비스 제공단말기로부터 수신된 영상정보로부터 사용자의 제스처 및 표정 중 적어도 어느 하나에서 상기 사용자의 감정상태를 인지하고, 상기 인지된 감정상태를 기 저장된 객체 관련 정보와 비교하여, 상기 인지된 감정상태와 매칭되는 객체를 추출하며, 상기 추출된 객체를 상기 사용자의 바디 및 얼굴 중 적어도 어느 하나 위에 중첩시켜, 상기 영상통화서비스 제공단말기이기와 통신하는 상대방 영상통화서비스 제공단말기이기로 전송하는 서버 제어부;를 포함하는 것을 특징으로 한다. In addition, the emotional content service apparatus of the communication terminal apparatus according to the present invention for achieving the above object, the server communication unit interworking with the video call service providing terminal; And recognize the emotion state of the user from at least one of a gesture and an expression of the user from the image information received from the video call service providing terminal, and compare the recognized emotion state with previously stored object related information. A server for extracting an object matching an emotional state and superimposing the extracted object on at least one of a body and a face of the user and transmitting the object to a counterpart video call service providing terminal communicating with the video call service providing terminal. And a control unit.
이때, 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스 장치는, 상기 감정상태와 대응하는 객체 관련 데이터를 저장하는 서버 저장부;를 더 포함하는 것을 특징으로 한다. In this case, the emotion content service apparatus of the communication terminal device according to the present invention, characterized in that it further comprises a server storage unit for storing the object-related data corresponding to the emotional state.
또한, 상기와 같은 목적들을 달성하기 위한 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 장치는 영상통화에 따른 상대방의 영상 및 상기 영상에 중첩되는 객체를 표시하는 표시부; 영상통화서비스 제공서버와 연동하는 통신부; 영상통화에 따른 사용자의 영상정보를 획득하는 촬상부; 및 상기 촬상부에서 획득된 상기 영상정보로부터 상기 사용자의 감정상태를 인지하고, 인지된 감정상태와 관련된 감정정보를 추출하여 상기 영상통화서비스 제공서버로 전송하며, 상기 영상통화서비스 제공서버로부터 상기 영상통화에 따른 상대방의 감정정보에 대응하는 객체를 수신하여, 상기 영상통화에 따른 상대방의 영상에서 상기 수신된 객체와 연관되는 위치에 상기 수신된 객체를 중첩하여 상기 표시부에 출력하는 제어부;를 포함하는 것을 특징으로 한다.In addition, an emotion recognition apparatus for an emotion content service of a communication terminal apparatus according to the present invention for achieving the above objects includes a display unit for displaying an image of the other party and an object overlapping the image according to the video call; A communication unit interworking with a video call service providing server; An imaging unit which acquires image information of a user according to a video call; And recognize the emotion state of the user from the image information obtained by the image pickup unit, extract emotion information related to the recognized emotion state, and transmit the extracted emotion information to the video call service providing server, and the image from the video call service providing server. And a controller configured to receive an object corresponding to the emotion information of the other party according to a call and to superimpose the received object on a position associated with the received object in the image of the other party according to the video call and to output the object to the display unit. It is characterized by.
또한, 상기와 같은 목적들을 달성하기 위한 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 방법을 통해 상기 감정 컨텐츠를 생성하고 정합하는 장치는, 영상통화에 따른 상대방의 영상 및 상기 영상에 중첩되는 객체를 표시하는 표시부; 영상통화서비스 제공서버와 연동하는 통신부; 영상통화에 따른 사용자의 영상정보를 획득하는 촬상부; 및 상기 촬상부에서 획득된 상기 영상정보로부터 상기 사용자의 감정상태를 인지하고, 인지된 감정상태와 관련된 감정정보를 추출하여 상기 영상통화서비스 제공서버로 전송하며, 상기 영상통화서비스 제공서버로부터 상기 영상통화에 따른 상대방의 감정정보에 대응하는 객체를 수신하여, 상기 영상통화에 따른 상대방의 영상에서 상기 수신된 객체와 연관되는 위치에 상기 수신된 객체를 중첩하여 상기 표시부에 출력하는 제어부;를 포함하는 것을 특징으로 한다. In addition, the apparatus for generating and matching the emotion content through the emotion recognition method for the emotion content service of the communication terminal device according to the present invention for achieving the above object, the video of the other party and the video according to the video call A display unit displaying overlapping objects; A communication unit interworking with a video call service providing server; An imaging unit which acquires image information of a user according to a video call; And recognize the emotion state of the user from the image information obtained by the image pickup unit, extract emotion information related to the recognized emotion state, and transmit the extracted emotion information to the video call service providing server, and the image from the video call service providing server. And a controller configured to receive an object corresponding to the emotion information of the other party according to a call and to superimpose the received object on a position associated with the received object in the image of the other party according to the video call and to output the object to the display unit. It is characterized by.
이때, 본 발명에 따른 상기 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 방법을 통해 상기 감정 컨텐츠를 생성하고 정합하는 장치의 상기 표시부는, 상기 영상통화에 따른 사용자의 영상을 더 표시하는 것을 특징으로 한다.In this case, the display unit of the device for generating and matching the emotion content through the emotion recognition method for the emotion content service of the communication terminal device according to the present invention, characterized in that for further displaying the image of the user according to the video call do.
또한, 본 발명에 따른 상기 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 방법을 통해 상기 감정 컨텐츠를 생성하고 정합하는 장치는 상기 객체를 적용할지 여부를 결정하는 키입력부;를 더 포함하는 것을 특징으로 한다.The apparatus for generating and matching the emotion content through the emotion recognition method for the emotion content service of the communication terminal device according to the present invention further includes a key input unit for determining whether to apply the object. do.
또한, 본 발명에 따른 상기 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 방법을 통해 상기 감정 컨텐츠를 생성하고 정합하는 장치는 상기 객체를 저장하는 저장부;를 더 포함하는 것을 특징으로 한다. The apparatus for generating and matching the emotion content through the emotion recognition method for the emotion content service of the communication terminal device according to the present invention may further include a storage unit for storing the object.
또한, 상기와 같은 목적들을 달성하기 위한 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스는 촬상수단 및 디스플레이수단을 적어도 구비하는 통신단말장치의 상기 촬상수단 통해 촬영되는 사용자의 제스처 및 표정 중 적어도 어느 하나로부터 상기 사용자의 감정상태를 추출하고, 상기 추출된 감정상태에 대응하는 제스처 및 표정을 나타내는 상기 사용자의 바디 및 얼굴 중 적어도 어느 하나의 부위를 과장시켜, 상기 사용자와 영상통화 하는 상대방의 통신단말장치의 디스플레이수단에 표시하는 것을 특징으로 한다.In addition, the emotional content service of the communication terminal apparatus according to the present invention for achieving the above objects is at least one of a gesture and facial expression of the user photographed through the image pickup means of the communication terminal device having at least an image pickup means and a display means A communication terminal device of the other party that extracts the emotional state of the user from the user, exaggerates at least one of the body and the face of the user representing the gesture and expression corresponding to the extracted emotional state, and makes a video call with the user It is characterized by displaying on the display means.
상술한 바와 같이 본 발명은 현실세계에서 볼 수 없는 다양한 혼합현실을 영상통화를 통해 구현함으로써, 영상통화에 풍성한 볼거리를 제공할 수 있다.As described above, the present invention can provide abundant sights to the video call by implementing various mixed reality not seen in the real world through the video call.
특히, 본 발명은 영상통화 시에 통화자 쌍방의 감정상태를 표현하는 가상의 객체를 통화영상과 함께 제공함으로써 영상통화에 재미를 선사할 수 있는 효과가 있다. In particular, the present invention has an effect that can provide fun to the video call by providing a virtual object representing the emotional state of both callers with the video call during the video call.
또한, 본 발명은 통화자의 감정상태를 가상의 객체를 통해 통화자의 영상에 중첩시킴으로써, 통화자들의 영상통화에 더욱 현실감을 부여하는 증강현실을 경험할 수 있도록 하여 통화자들의 감정상태를 신선하게 전달할 수 있는 효과가 있다. In addition, the present invention by superimposing the emotional state of the caller to the video of the caller through a virtual object, it is possible to experience the augmented reality that gives the caller a more realistic feeling to deliver the emotional state of the caller freshly It has an effect.
뿐만 아니라, 본 발명은 음성, 얼굴과 몸체의 특정한 표정과 제스처의 표정을 영상통화 화면을 통해 가상의 객체로 형상화함으로써, 사용자들에게 가상과 현실을 모두 경험할 수 있도록 하는 효과가 있다. In addition, the present invention has the effect of enabling the user to experience both virtual and reality by shaping the expressions of voices, faces and bodies of specific expressions and gestures into virtual objects through the video call screen.
도 1은 본 발명에 따른 아바타 정합의 개념을 나타낸 도면, 1 is a view showing a concept of avatar matching according to the present invention;
도 2는 본 발명에 따른 이모티콘 정합의 개념을 나타낸 도면,2 is a view showing a concept of emoticon matching according to the present invention,
도 3은 본 발명의 실시예에 따라 사용자의 표정변화에 정합되는 합성 데이터 및 이에 대한 표준 데이터를 나타낸 도면,3 is a view showing composite data matched to a change in facial expression of a user and standard data thereof according to an embodiment of the present invention;
도 4는 3D 아바타의 감정 및 표정 인식과 정합 절차를 나타낸 제어흐름도,4 is a control flow diagram illustrating an emotion and facial expression recognition and matching procedure of a 3D avatar;
도 5는 본 발명에 따른 스토리텔링의 주요 캐릭터의 실시예를 표로 정리한 도면,5 is a table summarized the embodiments of the main characters of storytelling according to the present invention;
도 6은 본 발명에 따른 감정 컨텐츠 서비스 방법이 적용되는 접속망간의 사용자 이동에 따른 서비스 이동 시나리오를 나타낸 도면,6 is a diagram illustrating a service movement scenario according to user movement between access networks to which the emotion content service method according to the present invention is applied;
도 7은 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 방법 중 얼굴검출을 위한 제어흐름도,7 is a control flowchart for face detection in an emotion recognition method for an emotion content service of a communication terminal device according to the present invention;
도 8은 SIP 서비스를 위한 구성 및 동작(Req/Resp) 순서를 나타낸 도면,8 is a view showing a configuration and operation (Req / Resp) sequence for the SIP service,
도 9는 SIP의 기본 메시지 및 상태 코드 체계를 나타낸 도면, 9 is a diagram illustrating a basic message and status code scheme of a SIP;
도 10은 SIP 프로토콜 스텍을 나타낸 도면,10 is a diagram illustrating a SIP protocol stack;
도 11은 SIP 프로토콜의 호 설정 기본 과정을 나타낸 도면,11 is a diagram showing a basic procedure of call setup of a SIP protocol;
도 12는 본 발명에 따른 컨텐츠 정합시스템의 개념도,12 is a conceptual diagram of a content matching system according to the present invention;
도 13은 본 발명에 따른 영상정합을 위한 기본 동작 절차도,13 is a basic operation procedure for image registration according to the present invention,
도 14는 본 발명에 따른 감정인식 및 영상정합을 통한 아바타 화상통신 동작 절차 개요도,14 is a schematic diagram of an avatar video communication operation procedure through emotion recognition and image registration according to the present invention;
도 15는 도 14의 감정인식 및 영상정합을 통한 아바타 화상통신 동작 절차를 실제 콘텐츠를 이용하여 나타낸 도면,FIG. 15 is a view illustrating an avatar video communication operation procedure through emotion recognition and image registration of FIG. 14 using actual content; FIG.
도 16은 본 발명이 적용 가능한 다양한 분야를 나타낸 도면.16 is a view showing various fields to which the present invention is applicable.
이하 본 발명의 바람직한 실시 예들의 상세한 설명이 첨부된 도면들을 참조하여 설명될 것이다. 도면들 중 동일한 구성들은 가능한 한 어느 곳에서든지 동일한 부호들을 나타내고 있음을 유의하여야 한다. 하기 설명에서 구체적인 특정 사항들이 나타나고 있는데, 이는 본 발명의 보다 전반적인 이해를 돕기 위해 제공된 것이다. 그리고 본 발명을 설명함에 있어, 관련된 공지 기능 혹은 구성에 대한 구체적인 설명이 본 발명의 요지를 불필요하게 흐릴 수 있다고 판단되는 경우 그 상세한 설명을 생략한다. DETAILED DESCRIPTION Hereinafter, detailed descriptions of preferred embodiments of the present invention will be described with reference to the accompanying drawings. It should be noted that the same components in the figures represent the same numerals wherever possible. Specific details are set forth in the following description, which is provided to aid a more general understanding of the present invention. In the following description of the present invention, if it is determined that a detailed description of a related known function or configuration may unnecessarily obscure the subject matter of the present invention, the detailed description thereof will be omitted.
본 발명은 감정을 분석하여 특정 감정에 대해 아바타가 대신 표현하도록 하는 아바타 정합과, 표정을 분석하여 특정 표정에 대한 효과를 첨가하는 이모티콘 정합의 특징을 적어도 하나 이상 갖는다.The present invention has at least one feature of avatar matching to analyze an emotion to allow an avatar to express it for a specific emotion, and emoticon matching to analyze an expression and add an effect to the specific expression.
도 1은 본 발명에 따른 아바타 정합의 개념을 나타낸 도면이고, 도 2는 본 발명에 따른 이모티콘 정합의 개념을 나타낸 도면이다.1 is a view showing the concept of avatar matching according to the present invention, Figure 2 is a view showing the concept of emoticon matching according to the present invention.
도 1에 도시된 바와 같이, 본 발명에 따른 아바타 정합은 실제 영상에서 표현된 감정을 인지하여 감정이 증강현실로 표현된 아바타로 실제 영상을 교체한다. 이때 실제 영상의 화면 전체를 교체할 수도 있고 얼굴의 특정 부위만 증강현실로 표현할 수도 있다.As shown in FIG. 1, the avatar matching according to the present invention recognizes an emotion expressed in an actual image and replaces the actual image with an avatar expressed in augmented reality. In this case, the entire screen of the actual image may be replaced or only a specific part of the face may be expressed in augmented reality.
한편, 도 2에 도시된 바와 같이, 본 발명에 따른 이모티콘 정합은 실제 영상에서 표현된 감정을 인지하여 인지된 표정에 대한 전달 효과를 증대하기 위한 다양한 이모티콘들을 사용하는 것이다.On the other hand, as shown in Figure 2, matching the emoticon according to the present invention is to use a variety of emoticons for recognizing the emotions expressed in the actual image to increase the transmission effect on the recognized expression.
이에 따라, 본 발명에 따른 감정 컨텐츠는 일례로 남성적 감정인식 반응 10종, 여성적 감정인식 반응 10종 및 동물적 감정인식 반응 10종으로 구성될 수 있다. 이러한 감정 컨텐츠는 표준 데이터에 매칭이 가능한 3D 모델링을 통해서 동작 표현 스크립트와 연동되는 캐릭터와 동작을 표현할 수 있다.Accordingly, the emotional content according to the present invention may be composed of, for example, ten male emotional recognition reactions, ten female emotional recognition reactions, and ten animal emotional recognition reactions. Such emotion contents can express characters and motions linked with motion expression scripts through 3D modeling that can be matched to standard data.
도 3은 본 발명의 실시예에 따라 사용자의 표정변화에 정합되는 합성 데이터 및 이에 대한 표준 데이터를 나타낸 도면이다. 도 3에 도시된 바와 같이, 영상통화 시 사용자간 표정변화에 정합하기 위한 3D아바타를 제작하여, 표현하기 어려운 감정이나 표정을 아바타에게 대행시켜 사용자간 보다 재미있고 즐거운 영상통화가 될 수 있도록 표준 데이터를 대행하는 아바타 생성과 화상통화 어플리케이션을 제공한다. 이때, 본 발명은 감정 및 표정변화에 따른 남자, 여자, 의인화 캐릭터 제작 및 각 종별 퍼포먼스를 제작하여 화상통화 시 적용가능 하도록 구현할 수 있다.FIG. 3 is a diagram showing synthetic data and standard data thereof matched to a change in facial expression of a user according to an exemplary embodiment of the present invention. As shown in FIG. 3, by making a 3D avatar to match the change of facial expressions between users during a video call, standard data can be used to make a more interesting and pleasant video call between users by substituting an avatar or an emotion that is difficult to express. Provides avatar creation and video call application on behalf of. At this time, the present invention can be implemented to be applicable to the video call by producing a man, a woman, an anthropomorphic character according to the emotion and facial expression changes and each type of performance.
도 3에 도시된 바와 같이, 평온, 기쁨, 화남, 슬픔, 놀람, 공포, 뽀뽀(2종), 윙크(2종)에 대한 표준화 데이터를 구축하고, 하기에서 설명할 도 5의 '아바', '바타', '스폰', '바쓰'의 기본형 캐릭터 및 응용형 캐릭터를 3D 아바타로 제작한다. 이때, 캐릭터 별로 기본형 1종과 응용형으로 2종 이상 제작한다. 그리고 '아바', '바타', '스폰', '바쓰'의 감정 및 표정 반응에 따른 동작을 구현한 감정 및 표정 애니메이션을 제작한다. 이에 따라, 평온, 기쁨, 화남, 슬픔, 놀람, 공포, 뽀뽀(2종), 윙크(2종)에 대한 3D객체 및 애니메이션 데이터를 제작(예를 들어, 아바:10종, 바타:10종, 스폰:10종)한다.As shown in Figure 3, to build the standardized data for calm, joy, anger, sadness, surprise, fear, kiss (two species), wink (two species), and will be described below 'Ava', Basic characters and application characters of 'Bata', 'Sponsor' and 'Bath' are created as 3D avatars. In this case, two or more types of the basic type and the application type are produced for each character. In addition, the film produces emotion and facial expression animations that embody actions based on the emotion and facial response of Ava, Bata, Spawn, and Bath. Accordingly, 3D object and animation data for calm, joy, anger, sadness, surprise, horror, kiss (2 kinds), and wink (2 kinds) are produced (for example, Ava: 10 species, Bata: 10 species, Spawn: 10 species).
도 4는 3D 아바타의 감정 및 표정 인식과 정합 절차를 나타낸 제어흐름도로서, 도 4에 도시한 바와 같이 각 감정과 표정에 따라 제작된 아바타의 소스는 다음의 인식과 정합 단계를 통해서, 사용자 어플리케이션에 반영된다.FIG. 4 is a control flowchart illustrating a process of recognizing and matching emotions and facial expressions of a 3D avatar. As illustrated in FIG. 4, a source of an avatar produced according to each emotion and facial expression is applied to a user application through the following recognition and matching steps. Is reflected.
도 4에서, 청색으로 표시한 절차는 상용라이브러리로 구현되는 절차이고, 초록색으로 표시한 절차는 콘텐츠 구현 절차이며, 회색은 시스템영역을 나타낸 것이고, 적색은 화상통화시 사용자의 감정과 표정을 인식하여 해당 감정을 대행하는 아바타와 해당 표정에 효과를 더하는 이모티콘이 정합되는 스마트폰 어플리케이션 콘텐츠를 개발하기 위한 기술에 의하여 구현되는 절차이다. In FIG. 4, the procedure indicated in blue is a procedure implemented by a commercial library, the procedure indicated in green is a content implementation procedure, gray indicates a system area, and red indicates a user's emotion and facial expression during a video call. This is a procedure implemented by a technology for developing smartphone application content in which an avatar acting on a corresponding emotion and an emoticon adding an effect to a corresponding expression are matched.
즉, 본 발명은 감정 및 표정 정합지원 화상통화 서비스를 위해 필요한 얼굴인식, 감정인식, 영상정합, 화상통신 핵심 기술을 개발하고, 어플리케이션 콘텐츠와 연동 서버를 개발한다. 또한, 사용자를 대행할 아바타 및 이모티콘 제작을 수행하여, 스마트폰 어플리케이션에 적용한다.That is, the present invention develops core technologies for face recognition, emotion recognition, video registration, and video communication necessary for emotion and facial expression matching support video call service, and develops application content and interworking server. In addition, avatars and emoticons for the user are produced and applied to the smartphone application.
본 발명에 있어서, 송신 통신단말장치의 촬상장치로부터 인식된 감정과 표정을 분석하고, 분석된 감정이나 의사전달을 효과적으로 전달할 수 있도록 수신 통신단말장치의 화면에 가상의 객체를 실시간 혼합시켜 구현하는 감정 컨텐츠 서비스를 제공하기 위한 감정 인지 장치 및 방법을 통해 상기 감정 컨텐츠를 생성하는 장치 및 방법에 의하여 생성하는 상기 감정 컨텐츠는 캐릭터의 창의적, 미래지향적 가치 및 스마트 이미지를 활용하는 것으로서, 새로운 테마, 새로운 트랜드의 모바일 영상통화 속 콘텐츠로 활용된다. In the present invention, the emotion and facial expressions recognized from the image pickup device of the transmitting communication terminal device, and the emotion to implement the real-time mixing of the virtual object on the screen of the receiving communication terminal device to effectively deliver the analyzed emotion or communication The emotion content generated by the device and method for generating the emotion content through the emotion recognition device and method for providing a content service utilizes the creative, future-oriented value and smart image of the character, a new theme, a new trend. It is used as content in mobile video call.
특히, 본 발명에 따른 감정 컨텐츠는 미래의 삶을 바꿀 첨단 스마트기기를 활용한 차별화된 테마로 교육적 효용성 제고하고, 차별화된 콘텐츠 스토리텔링 및 보편적 가치의 형상화를 통해 예술적 가치를 극대화한다. 즉, 3D애니메이션은 영상플레이어를 통해서만 보는 것이 아닌 다른 이와의 영상통화를 통해서도 가능하다는 차별화 요소를 극대화시킨 것이다.In particular, the emotional content according to the present invention enhances the educational utility as a differentiated theme using advanced smart devices that will change the future life, and maximizes artistic value through differentiated content storytelling and the shaping of universal values. In other words, it maximizes the differentiation factor that 3D animation is possible not only through video player but also through video call with others.
이에 따라 본 발명에 따른 감정 컨텐츠는 우리와 함께 하고 일상에서 언제나 교류하는 스마트 영상통화를 통해 우리의 삶이 행복하고 소중하다는 교훈적 가치와 다양한 표정변화를 유도하여 우리의 아이들의 정서함양에 도움이 될 수 있는 구성안을 제시한다.Accordingly, the emotional content according to the present invention can be helpful for the emotional development of our children by inducing a didactic value and various facial expression changes that our lives are happy and precious through smart video calls that accompany us and always communicate in everyday life. Present a proposal.
이를 위하여, 차별화된 콘텐츠 스토리텔링의 일례로서, 외로움을 이겨내고 소중한 친구들과 함께 성장하는 '아바'를 통해 부모와 아이의 소통, 친구와 친구의 소통, 선생님과 학생의 소통, 직장상사와 직장동료의 소통을 유도하여 오늘날 문제로 떠오르는 계층 간 '소통'의 문제를 주인공 '아바'와 그의 친구 '바타'종족을 통해 풀어보고자 한다. 이때, 본 발명에 따른 캐릭터의 비주얼 이미지는 3~4등신의 인물구성을 통해 누구나 좋아하고 받아들일 수 있는 캐릭터 이미지를 구축하는 것이 바람직하다.To this end, as an example of differentiated content storytelling, through 'Ava', which overcomes loneliness and grows with precious friends, parent and child communication, friend and friend communication, teacher and student communication, workplace boss and coworker By inducing communication, I will try to solve the problem of 'communication' that emerges as a problem through the main character 'Ava' and his friend 'Bata' race. At this time, it is preferable that the visual image of the character according to the present invention constructs a character image that anyone can like and accept through the composition of figures 3-4.
도 5는 본 발명에 따른 스토리텔링의 주요 캐릭터의 실시예를 표로 정리한 도면으로서, 나와 함께 살아 숨 쉬는 생명에 대한 소중함을 이해하고, 상상과 모험, 우정과 가족에 대한 보편적 가치와 소통의 중요성을 전달하기 위한 주요 캐릭터와 그 이미지, 그리고 성격 및 역할을 도시한 것이다.Figure 5 is a table summarized the embodiments of the main characters of storytelling according to the present invention, understanding the importance of living and breathing with me, the importance of universal values and communication for imagination and adventure, friendship and family It shows the main characters, their images, and their personalities and roles to convey them.
도 5의 실시예의 배경 스토리 구조는 주인공 '아바'(사용자)가 자신의 휴대폰 속에 살고 있는 가상공간 속 종족 '바타'를 만나면서 벌어지는 좌충우돌 에피소드를 다룬다. 주인공은 영상통화를 사용하는 사용자이며, 주인공의 친구인 가상공간 속에서 의인화된 캐릭터 '바타'는 주인공이 상대방과 영상통화를 할 때 언제든지 만날 수 있는 휴대폰 속 나만의 친구이다. 가상공간 속에서 살고 있는 종족 '바타'가 주인공과 만날 수 있는 유일한 통로는 휴대폰 영상통화의 통신 단말기를 통해서이며, 따라 하기를 좋아하고, 장난을 즐기는 성격 탓에 주인공의 감정반응에 따라 주인공처럼 행동하고, 주인공을 대신해 다양한 표정 이모티콘을 현실처럼 보여주는 신비한 능력을 지닌 종족으로 표현된다.The background story structure of the embodiment of FIG. 5 deals with a left-right doldol episode that occurs while the main character 'Ava' (user) meets a race 'Bata' in a virtual space living in his mobile phone. The protagonist is a user who uses a video call, and the character 'Bata', personified in the virtual space that is the friend of the protagonist, is a friend of mine who can meet at any time when the protagonist makes a video call with the other party. The only passage through which the race 'Bata', who lives in the virtual space, meets the main character through the communication terminal of the mobile phone video call, and because he likes to follow and enjoys playing, he acts like the main character according to the emotional reaction of the main character. It is represented as a race with mysterious ability to show various facial expression emoticons on behalf of the main character.
도 5의 배경 스토리 구조에 따라, 평온, 기쁨, 화남, 슬픔, 놀람, 공포, 뽀뽀, 윙크에 대한 표준화 데이터 구축하고, 아바, 바타, 스폰, 바쓰의 기본형 캐릭터 및 다수의 응용형 캐릭터 제작하며, 아바, 바타, 스폰, 바쓰의 감정 및 표정 반응에 따른 동작 구현함과 아울러, 각각의 캐릭터의 평온, 기쁨, 화남, 슬픔, 놀람, 공포, 뽀뽀, 윙크에 대한 3D 객체 및 애니메이션 데이터를 제작한다.According to the background story structure of FIG. 5, standardized data on calmness, joy, anger, sadness, surprise, horror, kiss, wink, and aba, bata, spawn, bass basic characters and a number of applied characters are produced. In addition to embodying the emotions and facial expressions of Ava, Bata, Spawn, and Bath, 3D objects and animation data for calm, joy, anger, sadness, surprise, horror, kiss and wink of each character are produced.
본 발명에 따른 감정 컨텐츠를 스마트폰 등 영상통신단말에서 3D객체로 표현하는데 있어서, 화상통화 중에 얼굴검출에서 인지 및 감정인지까지의 이미지 처리를 위한 프로세스 소요 시간이 클 것으로 예상되며, 또한 인지한 정보를 바탕으로 동영상과 3D객체간의 정합 등을 표현함에 있어, 딜레이가 발생할 것으로 예상된다. 따라서 3D정합 측면에서 3D엔진사용을 통한 정합 지연시간을 최소화 하는 방안을 고려할 수 있으며, 일례로 OpenGL ES를 이용한 3D오브젝트 표현 방식에서 Unity3D 엔진을 사용하여 개발속도를 개선시킬 수 있다.In expressing the emotion content according to the present invention as a 3D object in a video communication terminal such as a smartphone, it is expected that the processing time for image processing from face detection to recognition and emotion recognition during a video call is large, and also recognized information. On the basis of this, delays are expected to occur in expressing matching between video and 3D objects. Therefore, in terms of 3D matching, it is possible to consider the method of minimizing the matching delay time by using the 3D engine. For example, the development speed can be improved by using the Unity3D engine in the 3D object representation method using OpenGL ES.
도 6은 본 발명에 따른 감정 컨텐츠 서비스 방법이 적용되는 접속망간의 사용자 이동에 따른 서비스 이동 시나리오를 나타낸 도면으로서, 스마트폰을 통한 이동환경에서의 통신망 접속 문제을 해결하기 위한 방안을 모색해 볼 수 있다. FIG. 6 is a diagram illustrating a service movement scenario according to user movement between access networks to which the emotion content service method according to the present invention is applied, and a method for solving a communication network access problem in a mobile environment through a smart phone may be sought. .
도 6에 도시된 바와 같이, 사용자의 위치 이동에 따라 가용한 접속망을 선택함에 있어 네트워크 환경 측면의 문제가 존재한다. 즉, 다른 접속망에 접속함으로써 할당되는 IP 주소의 변경은 사용중인 서비스에 대한 단절 및 오류를 유발 할 수 있다. 특히, 화상통화와 같은 실시간성 서비스에 대해서는 품질 저하 및 서비스 불가 상황이 발생할 수 있으므로, 서비스 이동성 기술에 대한 적용 방안을 모색하고, 이러한 문제점에 대한 대응안을 도출할 필요성이 있다.As shown in FIG. 6, there is a problem in terms of network environment in selecting an available access network according to a user's location movement. In other words, changing the IP address assigned by connecting to another access network may cause disconnection and error of the service in use. In particular, since quality degradation and service unavailability may occur for a real-time service such as a video call, there is a need to find an application method for the service mobility technology and derive a countermeasure for such a problem.
도 7은 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 방법 중 얼굴검출을 위한 제어흐름도이다. 도 7을 참조하면, 얼굴검출은 영상에서 얼굴이 있는 위치를 알아내는 것으로, 사람의 얼굴은 응시하는 방향에 따른 정면 혹은 측면의 각도, 고개를 좌우로 기울이는 정도, 다양한 표정, 카메라와의 거리에 따른 얼굴 영상의 크기 등과 같은 형태적 변화와 조명에 따른 얼굴 내에서의 밝기 정도의 차이, 복잡한 배경 혹은 얼굴과 구분이 어려운 색상의 다른 객체 등과 같은 외부적 변화에 따라 매우 다양하게 나타날 수 있기 때문에 영상으로부터의 얼굴검출 연구는 많은 어려움을 포함하고 있다.7 is a control flowchart for face detection in an emotion recognition method for an emotion content service of a communication terminal device according to the present invention. Referring to FIG. 7, face detection is to find a location of a face in an image. A face of a person is determined based on an angle of the front or side according to the gazing direction, the degree of tilting the head from side to side, various expressions, and distance from the camera. The image may vary depending on external changes such as morphological changes such as the size of the face image, differences in brightness levels within the face due to lighting, complex backgrounds, or other objects of indistinguishable color from the face. Face detection studies from MM include many difficulties.
얼굴검출은 얼굴인식 이전의 전처리 단계로 지식 기반 방법(Knowledge-based Methods), 특징 기반 방법(Feature-based Methods), 템플릿 매칭 방법(Template-matching Methods), 외형 기반 방법(Appearance-based Methods) 등으로 나눠진다. 이를 하기의 표 1에 정리하였다. Face detection is a pre-processing step before face recognition, such as knowledge-based methods, feature-based methods, template-matching methods, and appearance-based methods. Divided by. This is summarized in Table 1 below.
표 1
Figure PCTKR2011008399-appb-T000001
Table 1
Figure PCTKR2011008399-appb-T000001
지식 기반 얼굴 검출 방법(Knowledge-based Face Detection Methods)은 사람의 얼굴이 눈썹, 눈, 코, 입 등 각각의 얼굴 성분들은 서로 일정한 거리와 위치 관계를 이용하여 얼굴을 검출하는 방법이다. 이 방법에서는 얼굴 영상의 중앙영역에 부분적인 명암도 집중이 나타나 얼굴영상과 영상의 명암도 분포를 비교를 통해 검출하며, 주로 하향식 접근법을 사용한다. 지식기반 검출 방법은 얼굴의 기울기, 카메라를 바라보는 각도, 표정 등과 같은 얼굴의 다양한 변화가 있는 영상에서는 얼굴의 검출이 어렵게 되기 때문에 변화가 없거나 특별한 경우에만 적용이 가능하다는 단점이 있다.Knowledge-based Face Detection Methods (Knowledge-based Face Detection Methods) is a method in which the face of the human face, such as eyebrows, eyes, nose, mouth, each face component using a constant distance and positional relationship to each other. In this method, partial contrast is concentrated in the center area of the face image, and the contrast distribution of the face image and the image is detected through comparison, and a top-down approach is mainly used. The knowledge-based detection method has a disadvantage in that it is difficult to detect a face in an image having various changes in the face, such as a tilt of a face, an angle of looking at a camera, and an expression, so that it can be applied only in a special case.
특징 기반 얼굴 검출 방법(Feature-based Face Detection Methods)이란 얼굴 특징 성분(눈, 코, 입, 외곽선, 명암도)들의 크기와 모양, 상호 연관성, 얼굴의 색상과 질감(texture) 정보와 성분들의 혼합된 형태의 정보를 이용해서 얼굴을 검출하는 방법이다. 특정 기반 방법에서는 상향식 접근법을 사용하여 얼굴의 부분적인 특징을 찾고 후보영역(얼굴 특정 성분)을 통합하여 얼굴을 찾는 방법이다. Feature-based Face Detection Methods are a combination of the size and shape of the facial feature components (eye, nose, mouth, outline, and contrast), their correlations, the color and texture information of the face and the components It is a method of detecting a face using shape information. In the specific based method, the bottom-up approach is used to find partial features of the face and to integrate the candidate regions (face specific components) to find the face.
특징 기반 얼굴 검출 방법들은 처리 시간이 빠르고 쉽게 얼굴을 찾을 수 있다고 포즈나 얼굴의 방향에 민감하지 않는 장점을 가지고 있다. 하지만, 피부색과 유사한 배경이나 개체에 대해 착오인식 할 수 있으며, 조명의 밝기 변화에 따라 얼굴의 색상 및 질감 정보를 잃을 수도 있다. 또한 얼굴의 기울어진 정도에 따라 얼굴의 특징 성분들을 검출하지 못할 수 있는 단점이 있다. Feature-based face detection methods have the advantage of not being sensitive to poses or face orientations, because the processing time can be found quickly and easily. However, it can be mistaken for a background or an object similar to the skin color, and the color and texture information of the face may be lost as the brightness of the light changes. In addition, there is a disadvantage that can not detect the feature components of the face according to the degree of inclination of the face.
템플릿 매칭 기반 얼굴 검출 방법(Template matching-based Face Detection Methods)은 대상이 되는 모든 얼굴에 대한 표준 템플릿을 만든 후, 입력 영상과 유사상관도를 비교하여 얼굴을 검출하는 방법으로 종류로는 사전 정의 템플릿알고리즘과 변형 템플릿 알고리즘이 있다.Template matching-based Face Detection Methods create a standard template for all of the target faces and then detect the face by comparing the similarity with the input image. There are algorithms and variant template algorithms.
템플릿 매칭 기반 얼굴 검출 방법은 준비된 이미지 자료에서 부분영역이나 외곽선을 이용하여 정보를 생성한 후, 만들어진 정보들은 각각의 알고리즘을 통해 변형시켜 유사 정보의 양을 늘려 얼굴 검출에 활용하게 된다. 그러나 템플릿 매칭 기반 얼굴 검출 방법은 거리에 따른 얼굴의 크기 변화와 응시 방향에 따른 얼굴의 회전각도, 기울어짐 등에 민감하며, 지식기반 방법과 같이 각각 다른 포즈에 대하여 템플릿을 정의하기 어려운 단점이 있다. The template matching-based face detection method generates information using partial regions or outlines from the prepared image data, and then transforms the generated information through algorithms to increase the amount of similar information and use it for face detection. However, the template matching-based face detection method is sensitive to the change in the size of the face according to the distance, the rotation angle and the tilt of the face according to the gaze direction, and it is difficult to define templates for different poses like the knowledge-based method.
외형 기반 방법(Appearance-based Methods)이란 패턴 인식을 이용하여 학습 영상 집합에 의해 학습된 모델을 이용해서 얼굴을 검출하는 방법이다. 외형 기반 방법은 얼굴 검출 분야에서 가장 많이 사용되는 방법 중 하나로 주성분 분석(Principal Component Analysis: PCA)에 의해 생성되는 고유 얼굴(eigenface), 선형판별식 해석(Linear Discriminant Analysis: LDA), 신경망(Neural Network: NN), 아다부스트(Adaboost), 서포트 벡터 머신(Support Vector Machine: SVM)을 이용하는 방법 등이 있다.An appearance-based method is a method of detecting a face using a model trained by a set of training images using pattern recognition. Appearance-based methods are one of the most used methods in the face detection field, and are based on eigenface, linear discriminant analysis (LDA), and neural network generated by Principal Component Analysis (PCA). NN), Adaboost, and Support Vector Machines (SVMs).
외형 기반 방법들은 복잡한 영상에서 얼굴영역을 검출하기 위해 기존의 얼굴 영역과 얼굴이 아닌 영역의 학습 데이터 집단을 이용해서 학습된 고유 벡터들을 만들어내고 이를 이용해 얼굴을 찾는다. 이 방법은 다른 검출 방법들에서 언급된 제약 조건들이 학습을 통해 극복되어지기 때문에 높은 인식률의 장점을 가진다. 하지만, PCA나 NN, SVM 등 외형기반 방법들은 데이터베이스 학습에 많은 시간이 필요하게 되고, 또한 데이터베이스가 변하게 되면 다시 학습을 시켜야 하는 단점이 있다.Appearance-based methods use the existing face and non-face learning data groups to detect face regions in complex images and generate learned eigenvectors to find faces. This method has the advantage of high recognition rate because the constraints mentioned in other detection methods are overcome by learning. However, appearance-based methods such as PCA, NN, and SVM require a lot of time to learn the database, and also have a disadvantage of having to learn again when the database changes.
다음으로, 얼굴인식(Face Recognition) 단계로서 얼굴인식 기술은 멀티미디어 영상을 통하여 얼굴을 검출한 뒤, 검출된 얼굴의 신원을 밝히기 위해 사용하는 방법이다. 얼굴인식 기술은 하기의 표 1과 같이 분류될 수 있으며, 본 발명에서는 얼굴의 구성요소를 파악하기 위해 사용된다.Next, as a face recognition step, face recognition technology is a method used to identify a face after detecting a face through a multimedia image. Face recognition technology can be classified as shown in Table 1 below, in the present invention is used to identify the components of the face.
표 2
Figure PCTKR2011008399-appb-T000002
TABLE 2
Figure PCTKR2011008399-appb-T000002
전체론적인 방법(Holistic methods)에 따른 얼굴 인식 방법에서 얼굴 인식 시스템의 입력은 얼굴 전체 영역을 사용한다. 일반적으로 전체론적인 얼굴 인식 방법은 쉽게 구현할 수 있는 장점을 가지고 있지만, 얼굴의 자세한 부분까지 고려하지 않기 때문에 충분한 결과를 얻기는 힘든 단점을 가진다. 전체론적인 얼굴 인식 방법은 주성분 분석(PCA), 선형판별 분석(LDA), 독립성분 분석(ICA), 텐서 얼굴(Tensorface)과 확률론적 결정 기반의 신경망(PDBNN) 방법 등을 포함한다. In the face recognition method according to holistic methods, the input of the face recognition system uses the entire face area. In general, the holistic face recognition method has an advantage that can be easily implemented, but it does not take enough detail of the face, so it is difficult to obtain sufficient results. Holistic face recognition methods include principal component analysis (PCA), linear discriminant analysis (LDA), independent component analysis (ICA), tensor face and probabilistic decision-based neural networks (PDBNN).
특징 기반의 매칭 방법(Feature-based methods)은 공간 특징(눈, 코 그리고 입)을 먼저 추출한 후, 공간 특징들의 위치와 공간 특성(기하학과 외형)이 인식 시스템에 입력된다. 특징 기반 방법은 얼굴에는 다양한 특징 정보들이 존재하기 때문에 얼굴 인식 성능을 높이기 위해 최적의 특징들을 어떻게 선택해야 하는지 등을 결정해야 하기 때문에 상당히 복잡하다. 그러나 대표적인 특징 기반의 방법들인 순수 기하학적(Pure Geometry) 방법, 동적 연결 구조(Dynamic Link Architecture) 방법, 은닉 마르코프 모델(Hidden Markov model) 방법 등은 위 전체론적인 매칭 방법보다 더 좋은 성능을 가짐으로 많이 활용된다. Feature-based methods first extract spatial features (eyes, nose and mouth), and then the location and spatial characteristics (geometry and appearance) of the spatial features are input to the recognition system. The feature-based method is quite complicated because there is a variety of feature information on the face, so it is necessary to determine how to select the best features to improve face recognition performance. However, the typical feature-based methods, such as Pure Geometry, Dynamic Link Architecture, and Hidden Markov model, have much better performance than the above holistic matching method. Are utilized.
하이브리드 방법(Hybrid methods)에서는 위치 특성과 더불어 하나의 얼굴을 인식하기 위하여 얼굴 전체 영역을 사용하기 때문에 매우 복잡하지만 인식률은 전체론적인 매칭 방법과 특징 기반의 매칭 방법들에 비하여 월등히 우수하다. 하이브리드 방법에는 선형 특징 분석(LFA), 형태 정규화(Shape-Norm alized) 방법과 컴포넌트 기반(Component-based) 방법 등이 있다.Hybrid methods are very complicated because they use the entire face area to recognize a face along with the location characteristics, but the recognition rate is much superior to the holistic matching and feature-based matching methods. Hybrid methods include linear feature analysis (LFA), shape-normalization, and component-based methods.
한편, 본 발명과 직접적인 연관성이 없는 기술요소이나, 영상정합을 통해 MoIP 부가서비스 기술에 적용할 수 있으므로 기본적인 SIP(Session Initiation Protocol) 기반 MoIP 서비스를 위한 요소기술로 구현될 수 있다. On the other hand, it can be implemented as an element technology for basic SIP (Session Initiation Protocol) based MoIP service because it can be applied to MoIP supplementary service technology through technology matching or image matching that is not directly related to the present invention.
도 8은 SIP 서비스를 위한 구성 및 동작(Req/Resp) 순서를 나타낸 도면이고, 도 9는 SIP의 기본 메시지 및 상태 코드 체계를 나타낸 도면이며, 도 10 및 도 11은 SIP 프로토콜 스텍 및 호 설정 기본 과정을 각각 나타낸 도면으로서, SIP는 멀티미디어 통신에 있어 세션이나 호(Call)을 관리하는 프로토콜로 멀티미디어 데이터 전송 자체보다는 시그널링(Signaling)을 통한 멀티미디어 통신 관리에 중점을 두고 있는 기법이다. 하기의 표 3에 SIP 서비스의 구성요소 및 그 주요기능을 정리하였다.FIG. 8 is a diagram illustrating a configuration and operation (Req / Resp) sequence for a SIP service, FIG. 9 is a diagram illustrating a basic message and status code scheme of SIP, and FIGS. 10 and 11 are a basic SIP protocol stack and call setup. As a diagram illustrating each process, SIP is a protocol for managing sessions or calls in multimedia communication, and is a technique that focuses on multimedia communication management through signaling rather than multimedia data transmission itself. Table 3 below summarizes the components of SIP service and its main functions.
표 3
Figure PCTKR2011008399-appb-T000003
TABLE 3
Figure PCTKR2011008399-appb-T000003
도 8 내지 도 11을 참조하면, SIP 프로토콜의 기본적인 호 설정과정을 설명하면, 먼저 발신자(Caller)가 수신자(Callee)에게 하나의 세션을 생성하기 위한 INVITE 요청(Request) 메시지를 보내게 된다. 이러한 메시지는 수신자에게 전달되기 위해서 몇 개의 SIP 서버를 거치게 된다. 메시지를 전달받은 프록시 서버는 메시지를 파싱하여 수신자를 인식하고 받은 메시지를 적절한 프록시 서버나 수신자의 사용자 에이젼트(User Agent : UA)로 전달하게 된다.8 to 11, the basic call setup process of the SIP protocol will be described. First, a caller sends an INVITE request message for creating a session to a callee. These messages go through several SIP servers to be delivered to the receiver. The received proxy server parses the message to recognize the recipient and delivers the received message to the appropriate proxy server or recipient's user agent (UA).
INVITE 메시지를 받은 수신자는 INVITE 메시지에 대한 응답(Response) 메시지를 보내게 되는데 응답 메시지는 처리 결과를 나타내기 위한 상태 코드를 가지고 있다. 만약 수신자가 제대로 메시지를 받아서 처리했다면 “200 OK”라는 응답 메시지를 발신자에게 보내게 된다. 이러한 응답을 받은 발신자는 수신자가 보낸 응답 메시지를 제대로 받았음을 알리기 위해서 ACK 요청 메시지를 수신자에게 다시 보내는 과정을 거쳐 하나의 세션이 생성된다.The receiver receiving the INVITE message sends a response message to the INVITE message. The response message has a status code indicating the result of processing. If the receiver receives and processes the message correctly, it sends a “200 OK” response message to the sender. The sender who receives the response sends a ACK request message back to the receiver to inform the receiver that the response message is correctly received.
한편, 유·무선 융합 서비스 환경은 단말의 이동성이 일반화된 환경으로서, 기존의 통신접속을 위한 단순한 의미에서의 액세스망 접속이 아니라 서비스 품질 및 사용자 기호 등의 기준을 통한 다양한 종류의 액세스 네트워크를 선택적으로 접속하여 통신을 수행하는 이종망간의 단말이동 환경으로 진화하고 있다. 따라서 이러한 이종망간의 단말 접속을 위해서는 이종망간의 이동성 지원 기술이 요구되며, 이동성 지원 기술에 대한 기능이 단말에 탑재되어야 한다.On the other hand, the wired / wireless convergence service environment is an environment in which the mobility of terminals is generalized, and various types of access networks are selected based on criteria such as service quality and user preferences, rather than access network access in a simple sense for existing communication access. It is evolving into a mobile terminal environment between heterogeneous networks that connect and communicate with each other. Therefore, in order to access a terminal between heterogeneous networks, a mobility support technology between heterogeneous networks is required, and a function for mobility support technology must be mounted in the terminal.
또한, 서로 다른 접속망에서의 통신이 가능하기 위해서는 단말이 접속 가능한 접속망에 대한 다수의 통신 인터페이스를 갖는 멀티모드 단말이 요구된다. 최근 펨토셀과 관련하여, 싱글모드 단말의 활용성을 고취시키려는 움직임도 있으나, 향후 서비스 품질, 비용 및 사용자 요구에 의한 이종망간의 이동성뿐만 아니라, 서비스의 이동 및 변환 등의 서비스를 고려할 경우 멀티모드 단말의 필요성이 더욱 커지고 있다. 더욱이, 이러한 멀티모드 단말의 이용에 있어, 현재의 접근은 이종망으로의 접속을 위해 통신 모드를 바꾸는 형태를 취하며, 이는 단말의 전원 및 서비스의 리셋을 필요로 한다.In addition, in order to be able to communicate in different access networks, a multi-mode terminal having a plurality of communication interfaces to access networks to which the terminals can be connected is required. Recently, there is a movement to enhance the utilization of single-mode terminals in relation to femtocells, but in the future, considering the services such as service migration and conversion as well as mobility between heterogeneous networks due to service quality, cost and user demand, multi-mode terminals The need for more is growing. Moreover, in the use of such a multimode terminal, the current approach takes the form of changing the communication mode for connection to a heterogeneous network, which requires a reset of the terminal's power and services.
따라서 앞으로의 물리적 이동이 자유로운 멀티모드 단말을 이용한 이종망간 이동통신 환경에서는 사용자의 단말 설정 및 서비스의 단절 없이 이종망간의 핸드오버가 자동적으로 제어되는 멀티모드 단말에서의 이종망간 자동접속제어 기술이 필요하다. Therefore, in a heterogeneous mobile communication environment using a multimode terminal free of physical movement in the future, an automatic access control technique between heterogeneous networks is required in a multimode terminal in which handover between heterogeneous networks is automatically controlled without user terminal setting and service disconnection. Do.
도 12는 본 발명에 따른 컨텐츠 정합시스템의 개념도로서, 개괄적인 동작절차로는 각 사용자의 단말에 들어있는 '아바타 비디오콜' 프로그램을 통해서, 화상통화 중 사용자의 얼굴과 감정을 인지하여 상대방에게 아바타 표정 및 이모티콘이 화면상에 정합되어 전송되는 형태를 나타낸다.12 is a conceptual diagram of a content matching system according to an embodiment of the present invention. As a general operation procedure, an avatar is recognized to a counterpart by recognizing a user's face and emotion during a video call through an Avatar video call program included in each user's terminal. The facial expressions and emoticons are matched on the screen and transmitted.
사용자는 화상통화 어플리케이션을 통해서, 상대와 화상통화를 수행함에 있어, 통화중 지속적인 얼굴인지와 감정인지를 수행하게 된다. 이때 인지된 감정이나 표정에 대한 아바타 매칭과 이모티콘 정합을 통해 상대방과의 보다 효과적이고 즐거운 화상통화가 가능하다.In the video call application, the user performs a continuous face recognition and emotion during the call. At this time, more effective and enjoyable video call with the other party is possible through avatar matching and emoticon matching with perceived emotion or facial expression.
기본적인 화상통화는 SIP 기반의 비디오 컨퍼런스를 통해 이루어지며, RTP/RTCP를 이용한 데이터 송수신을 수행한다. 네트워크 환경 및 상황에 따라, HTTP 스트리밍 방식으로의 전환도 고려할 수 있다.Basic video call is made through SIP-based video conference and performs data transmission and reception using RTP / RTCP. Depending on the network environment and situation, the switch to HTTP streaming may be considered.
도 13은 본 발명에 따른 영상정합을 위한 기본 동작 절차도로서, 도 13에 도시된 기 본 동작 절차를 통해서, 아바타와 동영상을 정합하게 된다. 도 13에 도시된 바와 같이, 스마트폰(단말)의 카메라 모듈을 통해 영상입력을 받아 얼굴인식, 표정인식, 감정인식을 위한 이미지 전처리를 수행한다. 그리고 가능성 있는 얼굴 후보를 추출하고 해당 얼굴의 구성요소를 분석하여 자세추정 및 감정인식을 위한 정보를 추출한다. 이들 정보를 통해서 아바타의 표정과 동작을 선택하고, 3D 공간상에서의 위치를 계산하여 동영상과 정합하여 화면에 출력한다. 또한 이 이미지를 동영상 인코딩을 하여 원격의 화상통화 상대방 스마트폰으로 네트워크를 통해 전송한다.FIG. 13 is a diagram illustrating a basic operation procedure for image registration according to the present invention. Through the basic operation procedure illustrated in FIG. 13, an avatar and a video are matched. As shown in FIG. 13, image input is performed through a camera module of a smart phone (terminal) to perform image preprocessing for face recognition, facial expression recognition, and emotion recognition. In addition, it extracts possible face candidates and analyzes the components of the face to extract information for posture estimation and emotion recognition. Through this information, the facial expression and motion of the avatar are selected, the position in the 3D space is calculated, matched with the video, and displayed on the screen. It also encodes this image and sends it over the network to a remote video call smartphone.
도 14는 본 발명에 따른 감정인식 및 영상정합을 통한 아바타 화상통신 동작 절차 개요도로서, ①단계는 카메라 모듈로부터 동영상 데이터 소스에서 분석할 프레임 이미지를 캡쳐하는 단계이고, ②단계는 캡쳐한 이미지를 분석하기 쉬운 상태로 전처리를 하는 단계로, 엣지 검출 알고리즘 등을 통한 이미지내의 객체간 경계를 파악할 수 있도록 한다.14 is a schematic diagram of an avatar video communication operation procedure through emotion recognition and image registration according to the present invention. Step 1 is a step of capturing a frame image to be analyzed in a video data source from a camera module, and step 2 is a analysis of the captured image. The preprocessing is performed in an easy-to-use state, so that the boundary between objects in the image can be grasped by an edge detection algorithm or the like.
다음으로, ③단계는 전처리된 이미지에서 얼굴을 검출하는 단계로, 학습 알고리즘 등을 적용하여 얼굴 특징점을 분석하고, 구성 요소의 이미지상 위치와 관계 데이터를 추출하고, ④단계는 얼굴검출 단계를 통해 추출된 인식 정보를 바탕으로 자세추정과 표정인식을 하는 단계로, 얼굴이 취하는 자세와 표정에 대한 아바타의 자세와 표정을 선정한다. 이어, ⑤단계는 분석된 정보를 통해서 3D 공간상의 아바타(얼굴) 위치 좌표를 결정하고, 해당 표정과 감정에 대한 아바타 애니메이션을 선정하고, 3D엔진에 제어 신호(메시지)를 전달한다. Next, step (3) detects a face from the preprocessed image. Analyzes facial feature points by applying a learning algorithm, extracts the position and relationship data on the image of the component, and step (4) uses the face detection step. At the stage of posture estimation and facial expression recognition based on the extracted recognition information, the attitude and facial expression of the avatar regarding the posture and facial expression taken by the face are selected. Next, step ⑤ determines the avatar (face) position coordinates in the 3D space through the analyzed information, selects an avatar animation for the expression and emotion, and transmits a control signal (message) to the 3D engine.
⑥단계에서는 아바타 영상과 동영상이 표현될 3D공간을 구성하고, 분석된 아바타를 해당 위치에 놓는 기능을 수행하고(3D엔진 API를 통한 아바타 및 3D공간 제어), ⑦단계에서 현재, 3D공간을 통해 표현되는 아바타와 동영상소스를 단일 이미지 소스로 정합하며, ⑧단계에서 정합된 이미지를 음성소스와 함께 동영상 인코딩을 수행한다. 음성소스는 동영상 소스로부터 추출하여 처리한다.In step ⑥, the 3D space in which the avatar image and video are represented is composed, and the analyzed avatar is placed in the corresponding position (controlling the avatar and 3D space through the 3D engine API). The avatar and the video source to be represented are matched to a single image source, and in step ⑧, the video is encoded along with the voice source. The audio source is extracted from the video source and processed.
마지막으로, ⑨단계에서 화상통화를 위해 구성된 상대단말로의 네트워크 전송단계로 SIP를 통해 세션을 구성하며, RTP/RTCP를 통해 인터넷을 통해 전송한다.Finally, in the step ⑨, the network is sent to the other terminal configured for the video call. The session is configured through the SIP and transmitted through the Internet through the RTP / RTCP.
도 15에 도 14의 감정인식 및 영상정합을 통한 아바타 화상통신 동작 절차를 실제 콘텐츠를 이용하여 나타내었다. 도 15에 도시된 바와 같이, 촬영된 영상으로부터 얼굴추적 및 눈, 코, 입을 인지하고, 이를 통해 표준분석 및 관계기술을 적용하여, 감정추론 및 실시간 정합하여 현실세계의 얼굴에 가상의 모형을 매칭함으로써, 영상통화 시 해당 표정을 지으면 사용자 얼굴 위에 실시간 아바타를 구현하게 된다.FIG. 15 illustrates an avatar video communication operation process through emotion recognition and image registration of FIG. 14 using actual content. As shown in FIG. 15, face tracking and eye, nose, and mouth are recognized from the captured image, and through this, standard analysis and relationship technology are applied, emotional inference and real-time matching to match a virtual model to the face of the real world. Thus, if the expression is made during the video call, a real-time avatar is implemented on the user's face.
전술한 바와 같이, 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스 장치 및 방법, 이를 위한 감정 인지 장치 및 방법, 이를 이용한 감정 컨텐츠를 생성하고 정합하는 장치 및 방법, 이 감정 컨텐츠는 음성인식, 객체인지의 얼굴인지기술의 얼굴영역검출기술, 얼굴영역 정규화 기술, 얼굴영역 내 특징추출기술, 객체인지의 감정인지기술의 얼굴구성요소(표정분석)관계기술, 객체인지의 손동작인지, 객체인지의 동작과 행동인지기술 기반 위에 실사와 가상영상의 실시간 정합기술을 활용하여 영상통화를 하는 쌍방의 얼굴과 몸체 위에 제스처 및 표정분석을 통한 혼합된 가상의 오브젝트(문자포함)를 정합시켜 현실세계에서 볼 수 없는 다양한 혼합현실을 영상통화를 통해 구현한다.As described above, an apparatus and method for emotion content service of a communication terminal device according to the present invention, an apparatus and method for emotion recognition therefor, an apparatus and method for generating and matching emotion content using the same, and whether the emotion content is voice recognition or an object Facial region detection technology of face recognition technology, facial region normalization technology, feature extraction technology in face region, facial component (expression analysis) relationship technology of emotion recognition technology of object recognition, object hand gesture, object recognition action and On the basis of behavioral cognitive technology, real-time matching technology of real-life and virtual images is used to match mixed virtual objects (including characters) through gesture and facial expression analysis on both faces and bodies making video calls, which cannot be seen in the real world. Various mixed reality is realized through video call.
또한, 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스 장치 및 방법, 이를 위한 감정 인지 장치 및 방법, 이를 이용한 감정 컨텐츠를 생성하고 정합하는 장치 및 방법, 이 감정 컨텐츠는 음성, 얼굴과 몸체의 특정한 표정과 제스처의 표정분석 관계연산함수를 미리 등록하여, 유사한 음성, 표정과 제스처가 영상을 통해 전송되면, 출력되는 영상화면에서 음성, 표정과 제스처에 반응한 가상의 객체를 얼굴과 몸체 위에 실시간으로 정합시켜, 사용자의 영상통화에 놀라운 즐거움을 준다. In addition, an apparatus and method for emotion contents service of a communication terminal device according to the present invention, an apparatus and method for emotion recognition for the same, an apparatus and method for generating and matching emotion contents using the same, the emotion contents include voice, a specific expression of a face and a body If the similar voice, facial expressions and gestures are transmitted through the video, the virtual objects responding to the voices, facial expressions and gestures are matched on the face and body in real time. It is a wonderful way to enjoy your video calls.
또한, 본 발명에 따른 통신단말장치의 감정 컨텐츠 서비스 장치 및 방법, 이를 위한 감정 인지 장치 및 방법, 이를 이용한 감정 컨텐츠를 생성하고 정합하는 장치 및 방법, 이 감정 컨텐츠는 모바일기기를 통해 사람의 표정이나 몸짓을 인식하고 이를 아바타를 통해 표현하는 시스템의 개발로 국내 영화, 애니메이션, 사이버캐릭터 등 첨단 영상산업분야를 한 단계 도약시키는 기반이 될 것이며, 인간의 감정과 표정을 제3의 대상(아바타)에 정합하여 현실세계의 얼굴위에 3D가상의 물체를 자연스럽게 표현하는 과정을 보다 효율적으로 만들어 국내 모바일 콘텐츠 및 영상 콘텐츠 산업분야의 경쟁력을 제고하는데 크게 기여할 것이다.In addition, an apparatus and method for emotion content service of a communication terminal device according to the present invention, an apparatus and method for emotion recognition for the same, an apparatus and method for generating and matching emotion content using the same, and the emotion content may be expressed through a mobile device. By developing a system that recognizes gestures and expresses them through avatars, it will become a foundation to take a leap forward in the cutting-edge video industry, such as domestic film, animation, and cyber characters, and bring human emotions and facial expressions to third parties (Avatar). It will greatly contribute to enhancing the competitiveness of domestic mobile contents and video contents industry by making the process of expressing 3D virtual objects on the face of the real world more natural.
한편 본 발명의 상세한 설명에서는 구체적인 실시 예에 관해 설명하였으나, 본 발명의 범위에서 벗어나지 않는 한도 내에서 여러 가지 변형이 가능함은 물론이다. 도 16은 본 발명이 적용 가능한 다양한 분야를 나타낸 도면으로서, 스마트 디바이스의 보급으로 향후에는 영화의 주인공과 관람객이 소통할 수 있을 것이라 예상하며, 주인공이 특정 관람객의 감정과 표정을 디바이스 상부에 설치된 카메라를 통해 인식한 후 그에 합당한 반응과 제스처를 보임으로 써, 같은 영상을 보지만 모두 다른 경험을 할 수 있도록 다양성을 제공할 수 있는 첨단문화산업분야에 적용 가능하다. Meanwhile, in the detailed description of the present invention, specific embodiments have been described, but various modifications are possible without departing from the scope of the present invention. FIG. 16 is a diagram illustrating various fields to which the present invention is applicable, and it is expected that the protagonist and the viewer of the movie may communicate in the future due to the spread of the smart device, and the main character of the camera installed on the upper part of the device. It can be applied to the high-tech cultural industry that can provide diversity to see the same video but have different experiences by showing the appropriate response and gestures through the recognition.
뿐만 아니라, 본 발명은 영화, 애니메이션, 사이버캐릭터 등에서 사용되는 스크립트 기반의 표정ㆍ몸짓 표현 기술, 사용자 감정 반응을 이용한 인터페이스 설계, 보안 및 감시를 위한 표정인식, 제품 및 디자인에 대한 소비자의 감성반응 측정 등 활용분야가 무궁무진하다. 그러므로 본 발명의 범위는 설명된 실시 예에 국한되어 정해져서는 안되며 후술하는 특허청구의 범위뿐 아니라 이 특허청구의 범위와 균등한 것들에 의해서 정해져야 한다.In addition, the present invention is a script-based expression and gesture expression technology used in movies, animation, cyber characters, interface design using user emotion reaction, facial recognition for security and surveillance, consumer's emotional response measurement for products and designs The field of application is endless. Therefore, the scope of the present invention should not be limited to the described embodiments, but should be determined not only by the scope of the following claims, but also by the equivalents of the claims.

Claims (20)

  1. 촬상수단 및 디스플레이수단을 적어도 구비하는 통신단말장치의 상기 촬상수단 통해 촬영되는 사용자의 제스처 및 표정 중 적어도 어느 하나로부터 상기 사용자의 감정상태를 추출하고, 상기 추출된 감정상태에 대응하는 가상의 객체를 생성하여, 상기 사용자의 바디 및 얼굴 중 적어도 어느 하나 위에 중첩시켜, 상기 사용자와 영상통화 하는 상대방의 통신단말장치의 디스플레이수단에 표시하는 통신단말장치의 감정 컨텐츠 서비스 방법.Extracting the emotional state of the user from at least one of a gesture and a facial expression of the user photographed through the imaging means of the communication terminal apparatus having at least an image pickup means and a display means, and extracting a virtual object corresponding to the extracted emotional state; Generating and superimposing on at least one of a body and a face of the user to display on the display means of the communication terminal device of the other party having a video call with the user.
  2. 제 1항에 있어서, 상기 가상의 객체는,The method of claim 1, wherein the virtual object,
    문자를 더 포함하는 것을 특징으로 하는 통신단말장치의 감정 컨텐츠 서비스 방법.Emotional content service method of a communication terminal device further comprising a character.
  3. 제 1항에 있어서, 상기 가상의 객체는,The method of claim 1, wherein the virtual object,
    상기 사용자에 의하여 변경 가능한 것을 특징으로 하는 통신단말장치의 감정 컨텐츠 서비스 방법.Emotional content service method of a communication terminal device, characterized in that changeable by the user.
  4. 제 1항에 있어서, 상기 가상의 객체는,The method of claim 1, wherein the virtual object,
    상기 감정상태에 대응하여 실시간으로 변화하는 것을 특징으로 하는 통신단말장치의 감정 컨텐츠 서비스 방법.Emotion content service method of a communication terminal device, characterized in that for changing in real time corresponding to the emotional state.
  5. 제 1항에 있어서, 상기 가상의 객체는,The method of claim 1, wherein the virtual object,
    상기 사용자의 바디 및 얼굴에 중첩된 위치가 변경되는 것을 특징으로 하는 통신단말장치의 감정 컨텐츠 서비스 방법.Emotion contents service method of a communication terminal device, characterized in that for changing the position superimposed on the body and face of the user.
  6. 제 1항에 있어서, 상기 통신단말장치가 음성입력수단 및 음성출력수단을 더 구비하면, The communication terminal according to claim 1, further comprising a voice input means and a voice output means.
    상기 음성입력수단을 통해 입력되는 사용자의 음성으로부터 상기 사용자의 감정상태를 더 추출하는 것을 특징으로 하는 통신단말장치의 감정 컨텐츠 서비스 방법.Emotion content service method of a communication terminal device, characterized in that further extracting the emotional state of the user from the user's voice input through the voice input means.
  7. 통신단말장치에 사용자의 얼굴 영상을 입력하는 과정;Inputting a face image of the user to the communication terminal device;
    상기 입력된 얼굴 영상에서 얼굴 구성 요소를 추출하는 과정;Extracting face components from the input face image;
    상기 추출된 얼굴 구성 요소를 전처리하는 과정;Preprocessing the extracted facial components;
    상기 전처리된 얼굴 구성 요소에서 얼굴의 특징을 추출하는 과정;Extracting facial features from the preprocessed facial components;
    상기 추출된 얼굴의 특징을 얼굴 데이터베이스에 등록하는 과정; 및Registering the extracted feature of the face in a face database; And
    상기 얼굴 데이터베이스에 등록된 특징과 상기 특징 추출 과정에서 추출된 얼굴 구성 요소를 비교하여 감정을 인지하는 과정;을 포함하는 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 방법.And a step of recognizing an emotion by comparing a feature registered in the face database with a face component extracted in the feature extraction process.
  8. 통신단말장치의 카메라 모듈로부터 사용자의 얼굴 영상을 입력받는 과정;Receiving a face image of a user from a camera module of the communication terminal device;
    상기 입력된 얼굴 영상을 전처리하는 과정;Preprocessing the input face image;
    상기 전처리된 얼굴 영상에서 유효 데이터만을 검출하는 과정;Detecting only valid data in the preprocessed face image;
    상기 검출된 유효 데이터로부터 얼굴의 위치 및 상기 카메라 정보를 추정하는 과정; 및Estimating the position of the face and the camera information from the detected valid data; And
    상기 카메라 정보와 얼굴의 위치정보로부터 3D 영상을 생성하고, 생성된 상기 3D영상을 상기 사용자의 얼굴 영상에 정합하는 과정;을 포함하는 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 방법을 통해 상기 감정 컨텐츠를 생성하고 정합하는 방법. Generating a 3D image from the camera information and position information of the face, and matching the generated 3D image to the face image of the user; the emotion through a emotion recognition method for an emotion content service of a communication terminal device How to create and match content.
  9. 제 8항에 있어서,The method of claim 8,
    상기 사용자의 얼굴 영상과 정합된 상기 3D영상을 화면에 출력하는 과정; 및Outputting the 3D image matched with the face image of the user on a screen; And
    상기 사용자의 얼굴 영상과 정합된 상기 3D영상을 네트워크를 통해 상대방 통신단말기로 전송하는 과정;을 더 포함하는 것을 특징으로 하는 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 방법을 통해 상기 감정 컨텐츠를 생성하고 정합하는 방법.Transmitting the 3D image matched with the face image of the user to a counterpart communication terminal through a network; generating the emotion content through an emotion recognition method for an emotion content service of a communication terminal device; And how to match.
  10. 통신단말장치의 카메라 모듈로부터 동영상 데이터 소스에서 분석할 프레임 이미지를 캡쳐하는 과정;Capturing a frame image to be analyzed in a moving image data source from a camera module of the communication terminal device;
    상기 캡쳐한 이미지를 분석하기 쉬운 상태로 전처리를 하는 과정;Preprocessing the captured image in an easy to analyze state;
    상기 전처리된 이미지에서 얼굴을 검출하는 과정;Detecting a face in the preprocessed image;
    상기 얼굴 검출 과정을 통해 추출된 인식 정보를 바탕으로 자세추정과 표정을 인식하여, 얼굴이 취하는 자세와 표정에 대한 아바타의 자세와 표정을 선정하는 과정; Recognizing a posture estimation and a facial expression based on the recognition information extracted through the face detection process, and selecting a posture and facial expression of the avatar with respect to the posture and facial expression taken by the face;
    분석된 정보를 통해서 3D 공간상에 상기 선정된 아바타의 위치 좌표를 결정하고, 해당 표정과 감정에 대한 아바타 애니메이션을 선정하며, 3D엔진에 제어 신호를 전달하는 과정; Determining position coordinates of the selected avatar on the 3D space through the analyzed information, selecting an avatar animation for a corresponding expression and emotion, and transmitting a control signal to the 3D engine;
    상기 아바타 영상과 동영상이 표현될 3D공간을 구성하고, 분석된 아바타를 해당 위치에 놓는 기능을 수행하는 과정; Composing a 3D space in which the avatar image and the video are to be represented, and performing a function of placing the analyzed avatar at a corresponding position;
    상기 3D공간을 통해 표현되는 아바타와 동영상소스를 단일 이미지 소스로 정합하는 과정; Matching the avatar and the video source represented through the 3D space into a single image source;
    상기 정합된 이미지를 음성소스와 함께 동영상 인코딩을 수행하는 과정; 및Performing video encoding on the matched image together with a voice source; And
    화상통화를 위해 구성된 네트워크를 통해 상대단말로 상기 정합된 이미지를 전송하는 과정;을 포함하는 통신단말장치의 감정 컨텐츠 서비스 방법.And transmitting the matched image to a counterpart terminal through a network configured for a video call.
  11. 제 10항에 있어서, 상기 전처리된 이미지에서 얼굴을 검출하는 과정은,The method of claim 10, wherein the detecting of the face in the preprocessed image comprises:
    학습 알고리즘 등을 적용하여 얼굴 특징점을 분석하고, 구성 요소의 이미지상 위치와 관계 데이터를 추출하는 것을 특징으로 하는 통신단말장치의 감정 컨텐츠 서비스 방법.Emotion content service method of a communication terminal device characterized by analyzing the facial feature points by applying a learning algorithm, and extracts the position and the relationship data on the image of the component.
  12. 제 10항에 있어서, 상기 네트워크를 통해 상대단말로 상기 정합된 이미지를 전송하는 과정은,The method of claim 10, wherein the transmitting of the matched image to the opposite terminal through the network comprises:
    SIP를 통해 세션을 구성하고, RTP/RTCP를 통해 인터넷으로 전송하는 것을 특징으로 하는 통신단말장치의 감정 컨텐츠 서비스 방법.Emotional content service method of a communication terminal device, characterized in that for establishing a session through the SIP, and transmitting to the Internet via RTP / RTCP.
  13. 영상통화서비스 제공단말기와 연동하는 서버 통신부; 및A server communication unit interworking with a video call service providing terminal; And
    상기 영상통화서비스 제공단말기로부터 수신된 영상정보로부터 사용자의 제스처 및 표정 중 적어도 어느 하나에서 상기 사용자의 감정상태를 인지하고, 상기 인지된 감정상태를 기 저장된 객체 관련 정보와 비교하여, 상기 인지된 감정상태와 매칭되는 객체를 추출하며, 상기 추출된 객체를 상기 사용자의 바디 및 얼굴 중 적어도 어느 하나 위에 중첩시켜, 상기 영상통화서비스 제공단말기이기와 통신하는 상대방 영상통화서비스 제공단말기로 전송하는 서버 제어부;를 포함하는 통신단말장치의 감정 컨텐츠 서비스 장치.Recognizing the emotional state of the user in at least one of the gesture and facial expression of the user from the image information received from the video call service providing terminal, and compared the recognized emotional state with the pre-stored object-related information, the recognized emotion A server controller which extracts an object matching a state and superimposes the extracted object on at least one of a body and a face of the user and transmits the extracted object to a counterpart video call service providing terminal communicating with the video call service providing terminal; Emotion content service device of a communication terminal device comprising a.
  14. 제 13항에 있어서, The method of claim 13,
    상기 감정상태와 대응하는 객체 관련 데이터를 저장하는 서버 저장부;를 더 포함하는 통신단말장치의 감정 컨텐츠 서비스 장치.And a server storage unit for storing object related data corresponding to the emotional state.
  15. 영상통화에 따른 상대방의 영상 및 상기 영상에 중첩되는 객체를 표시하는 표시부;A display unit displaying an image of the other party and an object overlapping the image according to a video call;
    영상통화서비스 제공서버와 연동하는 통신부;A communication unit interworking with a video call service providing server;
    영상통화에 따른 사용자의 영상정보를 획득하는 촬상부; 및An imaging unit which acquires image information of a user according to a video call; And
    상기 촬상부에서 획득된 상기 영상정보로부터 상기 사용자의 감정상태를 인지하고, 인지된 감정상태와 관련된 감정정보를 추출하여 상기 영상통화서비스 제공서버로 전송하며, 상기 영상통화서비스 제공서버로부터 상기 영상통화에 따른 상대방의 감정정보에 대응하는 객체를 수신하여, 상기 영상통화에 따른 상대방의 영상에서 상기 수신된 객체와 연관되는 위치에 상기 수신된 객체를 중첩하여 상기 표시부에 출력하는 제어부;를 포함하는 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 장치.Recognizing the emotional state of the user from the image information obtained by the image pickup unit, extracts the emotion information related to the recognized emotional state and transmits to the video call service providing server, the video call from the video call service providing server And a controller configured to receive an object corresponding to the emotion information of the other party according to the video call, and to superimpose the received object on a position associated with the received object in the video of the other party according to the video call and output the object to the display unit. Emotion recognition device for emotion content service of the terminal device.
  16. 영상통화에 따른 상대방의 영상 및 상기 영상에 중첩되는 객체를 표시하는 표시부;A display unit displaying an image of the other party and an object overlapping the image according to a video call;
    영상통화서비스 제공서버와 연동하는 통신부;A communication unit interworking with a video call service providing server;
    영상통화에 따른 사용자의 영상정보를 획득하는 촬상부; 및An imaging unit which acquires image information of a user according to a video call; And
    상기 촬상부에서 획득된 상기 영상정보로부터 상기 사용자의 감정상태를 인지하고, 인지된 감정상태와 관련된 감정정보를 추출하여 상기 영상통화서비스 제공서버로 전송하며, 상기 영상통화서비스 제공서버로부터 상기 영상통화에 따른 상대방의 감정정보에 대응하는 객체를 수신하여, 상기 영상통화에 따른 상대방의 영상에서 상기 수신된 객체와 연관되는 위치에 상기 수신된 객체를 중첩하여 상기 표시부에 출력하는 제어부;를 포함하는 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 방법을 통해 상기 감정 컨텐츠를 생성하고 정합하는 장치.Recognizing the emotional state of the user from the image information obtained by the image pickup unit, extracts the emotion information related to the recognized emotional state and transmits to the video call service providing server, the video call from the video call service providing server And a controller configured to receive an object corresponding to the emotion information of the other party according to the video call, and to superimpose the received object on a position associated with the received object in the video of the other party according to the video call and output the object to the display unit. And generating and matching the emotion content through an emotion recognition method for an emotion content service of a terminal device.
  17. 제 16항에 있어서, 상기 표시부는, The method of claim 16, wherein the display unit,
    상기 영상통화에 따른 사용자의 영상을 더 표시하는 것을 특징으로 하는 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 방법을 통해 상기 감정 컨텐츠를 생성하고 정합하는 장치.And generating and matching the emotion content through an emotion recognition method for an emotion content service of a communication terminal device, further comprising displaying an image of a user according to the video call.
  18. 제 16항에 있어서, The method of claim 16,
    상기 객체를 적용할지 여부를 결정하는 키입력부;를 더 포함하는 것을 특징으로 하는 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 방법을 통해 상기 감정 컨텐츠를 생성하고 정합하는 장치. And a key input unit configured to determine whether to apply the object. The apparatus for generating and matching the emotion content through an emotion recognition method for an emotion content service of a communication terminal device.
  19. 제 16항에 있어서, The method of claim 16,
    상기 객체를 저장하는 저장부;를 더 포함하는 것을 특징으로 하는 통신단말장치의 감정 컨텐츠 서비스를 위한 감정 인지 방법을 통해 상기 감정 컨텐츠를 생성하고 정합하는 장치.And a storage unit for storing the object. The apparatus for generating and matching the emotion content through an emotion recognition method for an emotion content service of a communication terminal device.
  20. 촬상수단 및 디스플레이수단을 적어도 구비하는 통신단말장치의 상기 촬상수단 통해 촬영되는 사용자의 제스처 및 표정 중 적어도 어느 하나로부터 상기 사용자의 감정상태를 추출하고, 상기 추출된 감정상태에 대응하는 제스처 및 표정을 나타내는 상기 사용자의 바디 및 얼굴 중 적어도 어느 하나의 부위를 과장시켜, 상기 사용자와 영상통화 하는 상대방의 통신단말장치의 디스플레이수단에 표시하는 통신단말장치의 감정 컨텐츠 서비스 방법.Extracting the emotion state of the user from at least one of a gesture and an expression of the user photographed through the imaging means of the communication terminal device having at least an image pickup means and a display means, and extracting a gesture and facial expression corresponding to the extracted emotion state; Emotion content service method of a communication terminal device to exaggerate at least one portion of the user's body and face to be displayed on the display means of the communication terminal device of the other party having a video call with the user.
PCT/KR2011/008399 2011-08-22 2011-11-07 Apparatus and method for emotional content services on telecommunication devices, apparatus and method for emotion recognition therefor, and apparatus and method for generating and matching the emotional content using same WO2013027893A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR1020110083435A KR20130022434A (en) 2011-08-22 2011-08-22 Apparatus and method for servicing emotional contents on telecommunication devices, apparatus and method for recognizing emotion thereof, apparatus and method for generating and matching the emotional contents using the same
KR10-2011-0083435 2011-08-22

Publications (1)

Publication Number Publication Date
WO2013027893A1 true WO2013027893A1 (en) 2013-02-28

Family

ID=47746615

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2011/008399 WO2013027893A1 (en) 2011-08-22 2011-11-07 Apparatus and method for emotional content services on telecommunication devices, apparatus and method for emotion recognition therefor, and apparatus and method for generating and matching the emotional content using same

Country Status (2)

Country Link
KR (1) KR20130022434A (en)
WO (1) WO2013027893A1 (en)

Cited By (181)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2014182052A1 (en) * 2013-05-09 2014-11-13 Samsung Electronics Co., Ltd. Method and apparatus for providing contents including augmented reality information
US9269374B1 (en) 2014-10-27 2016-02-23 Mattersight Corporation Predictive video analytics system and methods
WO2016077578A1 (en) * 2014-11-13 2016-05-19 Intel Corporation System and method for feature-based authentication
KR101652486B1 (en) * 2015-04-05 2016-08-30 주식회사 큐버 Sentiment communication system based on multiple multimodal agents
CN106127829A (en) * 2016-06-28 2016-11-16 广东欧珀移动通信有限公司 The processing method of a kind of augmented reality, device and terminal
CN106127828A (en) * 2016-06-28 2016-11-16 广东欧珀移动通信有限公司 The processing method of a kind of augmented reality, device and mobile terminal
CN106157363A (en) * 2016-06-28 2016-11-23 广东欧珀移动通信有限公司 A kind of photographic method based on augmented reality, device and mobile terminal
CN106157262A (en) * 2016-06-28 2016-11-23 广东欧珀移动通信有限公司 The processing method of a kind of augmented reality, device and mobile terminal
KR101743763B1 (en) * 2015-06-29 2017-06-05 (주)참빛솔루션 Method for providng smart learning education based on sensitivity avatar emoticon, and smart learning education device for the same
GB2529037B (en) * 2014-06-10 2018-05-23 2Mee Ltd Augmented reality apparatus and method
US10043406B1 (en) 2017-03-10 2018-08-07 Intel Corporation Augmented emotion display for austistic persons
CN108830917A (en) * 2018-05-29 2018-11-16 努比亚技术有限公司 A kind of information generating method, terminal and computer readable storage medium
CN108961431A (en) * 2018-07-03 2018-12-07 百度在线网络技术(北京)有限公司 Generation method, device and the terminal device of facial expression
CN109727303A (en) * 2018-12-29 2019-05-07 广州华多网络科技有限公司 Video display method, system, computer equipment, storage medium and terminal
CN109840009A (en) * 2017-11-28 2019-06-04 浙江思考者科技有限公司 A kind of intelligence true man's advertisement screen interactive system and implementation method
CN110046336A (en) * 2019-04-15 2019-07-23 南京孜博汇信息科技有限公司 Position encoded sheet disposal method and system
WO2019204464A1 (en) * 2018-04-18 2019-10-24 Snap Inc. Augmented expression system
CN110431838A (en) * 2017-03-22 2019-11-08 韩国斯诺有限公司 The method and system of the dynamic content of recognition of face video camera are provided
CN110705356A (en) * 2019-08-31 2020-01-17 深圳市大拿科技有限公司 Function control method and related equipment
US10554698B2 (en) 2017-12-28 2020-02-04 Hyperconnect, Inc. Terminal and server providing video call service
CN110874137A (en) * 2018-08-31 2020-03-10 阿里巴巴集团控股有限公司 Interaction method and device
CN111183455A (en) * 2017-08-29 2020-05-19 互曼人工智能科技(上海)有限公司 Image data processing system and method
CN111191564A (en) * 2019-12-26 2020-05-22 三盟科技股份有限公司 Multi-pose face emotion recognition method and system based on multi-angle neural network
CN111353842A (en) * 2018-12-24 2020-06-30 阿里巴巴集团控股有限公司 Processing method and system of push information
CN111773676A (en) * 2020-07-23 2020-10-16 网易(杭州)网络有限公司 Method and device for determining virtual role action
CN111918015A (en) * 2019-05-07 2020-11-10 阿瓦亚公司 Video call routing and management of facial emotions based on artificial intelligence determination
US10848446B1 (en) 2016-07-19 2020-11-24 Snap Inc. Displaying customized electronic messaging graphics
US10852918B1 (en) 2019-03-08 2020-12-01 Snap Inc. Contextual information in chat
US10861170B1 (en) 2018-11-30 2020-12-08 Snap Inc. Efficient human pose tracking in videos
US10872451B2 (en) 2018-10-31 2020-12-22 Snap Inc. 3D avatar rendering
US10880246B2 (en) 2016-10-24 2020-12-29 Snap Inc. Generating and displaying customized avatars in electronic messages
US10893385B1 (en) 2019-06-07 2021-01-12 Snap Inc. Detection of a physical collision between two client devices in a location sharing system
CN112215929A (en) * 2020-10-10 2021-01-12 珠海格力电器股份有限公司 Virtual social data processing method, device and system
US10896534B1 (en) 2018-09-19 2021-01-19 Snap Inc. Avatar style transformation using neural networks
US10895964B1 (en) 2018-09-25 2021-01-19 Snap Inc. Interface to display shared user groups
US10902661B1 (en) 2018-11-28 2021-01-26 Snap Inc. Dynamic composite user identifier
US10904181B2 (en) 2018-09-28 2021-01-26 Snap Inc. Generating customized graphics having reactions to electronic message content
US10911387B1 (en) 2019-08-12 2021-02-02 Snap Inc. Message reminder interface
US10939246B1 (en) 2019-01-16 2021-03-02 Snap Inc. Location-based context information sharing in a messaging system
US10936157B2 (en) 2017-11-29 2021-03-02 Snap Inc. Selectable item including a customized graphic for an electronic messaging application
US10936066B1 (en) 2019-02-13 2021-03-02 Snap Inc. Sleep detection in a location sharing system
US10951562B2 (en) 2017-01-18 2021-03-16 Snap. Inc. Customized contextual media content item generation
US10952013B1 (en) 2017-04-27 2021-03-16 Snap Inc. Selective location-based identity communication
US10949648B1 (en) 2018-01-23 2021-03-16 Snap Inc. Region-based stabilized face tracking
US10964082B2 (en) 2019-02-26 2021-03-30 Snap Inc. Avatar based on weather
US10963529B1 (en) 2017-04-27 2021-03-30 Snap Inc. Location-based search mechanism in a graphical user interface
US10979752B1 (en) 2018-02-28 2021-04-13 Snap Inc. Generating media content items based on location information
USD916871S1 (en) 2019-05-28 2021-04-20 Snap Inc. Display screen or portion thereof with a transitional graphical user interface
USD916811S1 (en) 2019-05-28 2021-04-20 Snap Inc. Display screen or portion thereof with a transitional graphical user interface
US10984569B2 (en) 2016-06-30 2021-04-20 Snap Inc. Avatar based ideogram generation
USD916872S1 (en) 2019-05-28 2021-04-20 Snap Inc. Display screen or portion thereof with a graphical user interface
US10984575B2 (en) 2019-02-06 2021-04-20 Snap Inc. Body pose estimation
USD916809S1 (en) 2019-05-28 2021-04-20 Snap Inc. Display screen or portion thereof with a transitional graphical user interface
USD916810S1 (en) 2019-05-28 2021-04-20 Snap Inc. Display screen or portion thereof with a graphical user interface
US10991395B1 (en) 2014-02-05 2021-04-27 Snap Inc. Method for real time video processing involving changing a color of an object on a human face in a video
US10992619B2 (en) 2019-04-30 2021-04-27 Snap Inc. Messaging system with avatar generation
US11010022B2 (en) 2019-02-06 2021-05-18 Snap Inc. Global event-based avatar
US11030789B2 (en) 2017-10-30 2021-06-08 Snap Inc. Animated chat presence
US11030813B2 (en) 2018-08-30 2021-06-08 Snap Inc. Video clip object tracking
US11032670B1 (en) 2019-01-14 2021-06-08 Snap Inc. Destination sharing in location sharing system
US11036781B1 (en) 2020-01-30 2021-06-15 Snap Inc. Video generation system to render frames on demand using a fleet of servers
US11036989B1 (en) 2019-12-11 2021-06-15 Snap Inc. Skeletal tracking using previous frames
US11039270B2 (en) 2019-03-28 2021-06-15 Snap Inc. Points of interest in a location sharing system
WO2021114710A1 (en) * 2019-12-09 2021-06-17 上海幻电信息科技有限公司 Live streaming video interaction method and apparatus, and computer device
CN113014471A (en) * 2021-01-18 2021-06-22 腾讯科技(深圳)有限公司 Session processing method, device, terminal and storage medium
US11048916B2 (en) 2016-03-31 2021-06-29 Snap Inc. Automated avatar generation
US11055514B1 (en) 2018-12-14 2021-07-06 Snap Inc. Image face manipulation
CN113099150A (en) * 2020-01-08 2021-07-09 华为技术有限公司 Image processing method, device and system
US11063891B2 (en) 2019-12-03 2021-07-13 Snap Inc. Personalized avatar notification
US11069103B1 (en) 2017-04-20 2021-07-20 Snap Inc. Customized user interface for electronic communications
US11074675B2 (en) 2018-07-31 2021-07-27 Snap Inc. Eye texture inpainting
US11080917B2 (en) 2019-09-30 2021-08-03 Snap Inc. Dynamic parameterized user avatar stories
US11100311B2 (en) 2016-10-19 2021-08-24 Snap Inc. Neural networks for facial modeling
US11103795B1 (en) 2018-10-31 2021-08-31 Snap Inc. Game drawer
US11120601B2 (en) 2018-02-28 2021-09-14 Snap Inc. Animated expressive icon
US11122094B2 (en) 2017-07-28 2021-09-14 Snap Inc. Software application manager for messaging applications
US11120597B2 (en) 2017-10-26 2021-09-14 Snap Inc. Joint audio-video facial animation system
US11128715B1 (en) 2019-12-30 2021-09-21 Snap Inc. Physical friend proximity in chat
US11128586B2 (en) 2019-12-09 2021-09-21 Snap Inc. Context sensitive avatar captions
US11140515B1 (en) 2019-12-30 2021-10-05 Snap Inc. Interfaces for relative device positioning
US11166123B1 (en) 2019-03-28 2021-11-02 Snap Inc. Grouped transmission of location data in a location sharing system
US11169658B2 (en) 2019-12-31 2021-11-09 Snap Inc. Combined map icon with action indicator
US11176737B2 (en) 2018-11-27 2021-11-16 Snap Inc. Textured mesh building
US11188190B2 (en) 2019-06-28 2021-11-30 Snap Inc. Generating animation overlays in a communication session
US11189098B2 (en) 2019-06-28 2021-11-30 Snap Inc. 3D object camera customization system
US11189070B2 (en) 2018-09-28 2021-11-30 Snap Inc. System and method of generating targeted user lists using customizable avatar characteristics
US11199957B1 (en) 2018-11-30 2021-12-14 Snap Inc. Generating customized avatars based on location information
US11218838B2 (en) 2019-10-31 2022-01-04 Snap Inc. Focused map-based context information surfacing
US11217020B2 (en) 2020-03-16 2022-01-04 Snap Inc. 3D cutout image modification
US11227442B1 (en) 2019-12-19 2022-01-18 Snap Inc. 3D captions with semantic graphical elements
US11229849B2 (en) 2012-05-08 2022-01-25 Snap Inc. System and method for generating and displaying avatars
US11245658B2 (en) 2018-09-28 2022-02-08 Snap Inc. System and method of generating private notifications between users in a communication session
US11263817B1 (en) 2019-12-19 2022-03-01 Snap Inc. 3D captions with face tracking
US11284144B2 (en) 2020-01-30 2022-03-22 Snap Inc. Video generation system to render frames on demand using a fleet of GPUs
US11294936B1 (en) 2019-01-30 2022-04-05 Snap Inc. Adaptive spatial density based clustering
US11310176B2 (en) 2018-04-13 2022-04-19 Snap Inc. Content suggestion system
US11307747B2 (en) 2019-07-11 2022-04-19 Snap Inc. Edge gesture interface with smart interactions
US11320969B2 (en) 2019-09-16 2022-05-03 Snap Inc. Messaging system with battery level sharing
US11356720B2 (en) 2020-01-30 2022-06-07 Snap Inc. Video generation system to render frames on demand
US11360733B2 (en) 2020-09-10 2022-06-14 Snap Inc. Colocated shared augmented reality without shared backend
CN114630135A (en) * 2020-12-11 2022-06-14 北京字跳网络技术有限公司 Live broadcast interaction method and device
WO2022143128A1 (en) * 2020-12-29 2022-07-07 华为技术有限公司 Video call method and apparatus based on avatar, and terminal
US11411895B2 (en) 2017-11-29 2022-08-09 Snap Inc. Generating aggregated media content items for a group of users in an electronic messaging application
US11425062B2 (en) 2019-09-27 2022-08-23 Snap Inc. Recommended content viewed by friends
US11425068B2 (en) 2009-02-03 2022-08-23 Snap Inc. Interactive avatar in messaging environment
US11438341B1 (en) 2016-10-10 2022-09-06 Snap Inc. Social media post subscribe requests for buffer user accounts
US11450051B2 (en) 2020-11-18 2022-09-20 Snap Inc. Personalized avatar real-time motion capture
US11452939B2 (en) 2020-09-21 2022-09-27 Snap Inc. Graphical marker generation system for synchronizing users
US11455082B2 (en) 2018-09-28 2022-09-27 Snap Inc. Collaborative achievement interface
US11455081B2 (en) 2019-08-05 2022-09-27 Snap Inc. Message thread prioritization interface
US11460974B1 (en) 2017-11-28 2022-10-04 Snap Inc. Content discovery refresh
US11516173B1 (en) 2018-12-26 2022-11-29 Snap Inc. Message composition interface
US11543939B2 (en) 2020-06-08 2023-01-03 Snap Inc. Encoded image based messaging system
US11544885B2 (en) 2021-03-19 2023-01-03 Snap Inc. Augmented reality experience based on physical items
US11544883B1 (en) 2017-01-16 2023-01-03 Snap Inc. Coded vision system
US11562548B2 (en) 2021-03-22 2023-01-24 Snap Inc. True size eyewear in real time
US11580700B2 (en) 2016-10-24 2023-02-14 Snap Inc. Augmented reality object manipulation
US11580682B1 (en) 2020-06-30 2023-02-14 Snap Inc. Messaging system with augmented reality makeup
US11616745B2 (en) 2017-01-09 2023-03-28 Snap Inc. Contextual generation and selection of customized media content
US11615592B2 (en) 2020-10-27 2023-03-28 Snap Inc. Side-by-side character animation from realtime 3D body motion capture
US11619501B2 (en) 2020-03-11 2023-04-04 Snap Inc. Avatar based on trip
US11625873B2 (en) 2020-03-30 2023-04-11 Snap Inc. Personalized media overlay recommendation
US11636662B2 (en) 2021-09-30 2023-04-25 Snap Inc. Body normal network light and rendering control
US11636654B2 (en) 2021-05-19 2023-04-25 Snap Inc. AR-based connected portal shopping
US11651572B2 (en) 2021-10-11 2023-05-16 Snap Inc. Light and rendering of garments
US11651539B2 (en) 2020-01-30 2023-05-16 Snap Inc. System for generating media content items on demand
US11660022B2 (en) 2020-10-27 2023-05-30 Snap Inc. Adaptive skeletal joint smoothing
US11663792B2 (en) 2021-09-08 2023-05-30 Snap Inc. Body fitted accessory with physics simulation
US11662900B2 (en) 2016-05-31 2023-05-30 Snap Inc. Application control using a gesture based trigger
US11670059B2 (en) 2021-09-01 2023-06-06 Snap Inc. Controlling interactive fashion based on body gestures
US11676199B2 (en) 2019-06-28 2023-06-13 Snap Inc. Generating customizable avatar outfits
US11673054B2 (en) 2021-09-07 2023-06-13 Snap Inc. Controlling AR games on fashion items
US11683280B2 (en) 2020-06-10 2023-06-20 Snap Inc. Messaging system including an external-resource dock and drawer
US11704878B2 (en) 2017-01-09 2023-07-18 Snap Inc. Surface aware lens
US11734866B2 (en) 2021-09-13 2023-08-22 Snap Inc. Controlling interactive fashion based on voice
US11734894B2 (en) 2020-11-18 2023-08-22 Snap Inc. Real-time motion transfer for prosthetic limbs
US11734959B2 (en) 2021-03-16 2023-08-22 Snap Inc. Activating hands-free mode on mirroring device
US11748958B2 (en) 2021-12-07 2023-09-05 Snap Inc. Augmented reality unboxing experience
US11748931B2 (en) 2020-11-18 2023-09-05 Snap Inc. Body animation sharing and remixing
US11763481B2 (en) 2021-10-20 2023-09-19 Snap Inc. Mirror-based augmented reality experience
US11790531B2 (en) 2021-02-24 2023-10-17 Snap Inc. Whole body segmentation
US11790614B2 (en) 2021-10-11 2023-10-17 Snap Inc. Inferring intent from pose and speech input
US11798238B2 (en) 2021-09-14 2023-10-24 Snap Inc. Blending body mesh into external mesh
US11798201B2 (en) 2021-03-16 2023-10-24 Snap Inc. Mirroring device with whole-body outfits
US11809633B2 (en) 2021-03-16 2023-11-07 Snap Inc. Mirroring device with pointing based navigation
US11818286B2 (en) 2020-03-30 2023-11-14 Snap Inc. Avatar recommendation and reply
US11823346B2 (en) 2022-01-17 2023-11-21 Snap Inc. AR body part tracking system
US11830209B2 (en) 2017-05-26 2023-11-28 Snap Inc. Neural network-based image stream modification
US11836862B2 (en) 2021-10-11 2023-12-05 Snap Inc. External mesh with vertex attributes
US11836866B2 (en) 2021-09-20 2023-12-05 Snap Inc. Deforming real-world object using an external mesh
US11842411B2 (en) 2017-04-27 2023-12-12 Snap Inc. Location-based virtual avatars
US11852554B1 (en) 2019-03-21 2023-12-26 Snap Inc. Barometer calibration in a location sharing system
US11854069B2 (en) 2021-07-16 2023-12-26 Snap Inc. Personalized try-on ads
US11863513B2 (en) 2020-08-31 2024-01-02 Snap Inc. Media content playback and comments management
US11870743B1 (en) 2017-01-23 2024-01-09 Snap Inc. Customized digital avatar accessories
US11868414B1 (en) 2019-03-14 2024-01-09 Snap Inc. Graph-based prediction for contact suggestion in a location sharing system
US11870745B1 (en) 2022-06-28 2024-01-09 Snap Inc. Media gallery sharing and management
US11880947B2 (en) 2021-12-21 2024-01-23 Snap Inc. Real-time upper-body garment exchange
US11888795B2 (en) 2020-09-21 2024-01-30 Snap Inc. Chats with micro sound clips
US11887260B2 (en) 2021-12-30 2024-01-30 Snap Inc. AR position indicator
US11893166B1 (en) 2022-11-08 2024-02-06 Snap Inc. User avatar movement control using an augmented reality eyewear device
US11900506B2 (en) 2021-09-09 2024-02-13 Snap Inc. Controlling interactive fashion based on facial expressions
US11910269B2 (en) 2020-09-25 2024-02-20 Snap Inc. Augmented reality content items including user avatar to share location
US11908083B2 (en) 2021-08-31 2024-02-20 Snap Inc. Deforming custom mesh based on body mesh
US11908243B2 (en) 2021-03-16 2024-02-20 Snap Inc. Menu hierarchy navigation on electronic mirroring devices
US11922010B2 (en) 2020-06-08 2024-03-05 Snap Inc. Providing contextual information with keyboard interface for messaging system
US11928783B2 (en) 2021-12-30 2024-03-12 Snap Inc. AR position and orientation along a plane
US11941227B2 (en) 2021-06-30 2024-03-26 Snap Inc. Hybrid search system for customizable media
US11954762B2 (en) 2022-01-19 2024-04-09 Snap Inc. Object replacement system
US11956190B2 (en) 2020-05-08 2024-04-09 Snap Inc. Messaging system with a carousel of related entities
US11960784B2 (en) 2021-12-07 2024-04-16 Snap Inc. Shared augmented reality unboxing experience
US11969075B2 (en) 2020-03-31 2024-04-30 Snap Inc. Augmented reality beauty product tutorials
US11978283B2 (en) 2021-03-16 2024-05-07 Snap Inc. Mirroring device with a hands-free mode
US11983826B2 (en) 2021-09-30 2024-05-14 Snap Inc. 3D upper garment tracking
US11983462B2 (en) 2021-08-31 2024-05-14 Snap Inc. Conversation guided augmented reality experience
US11991419B2 (en) 2020-01-30 2024-05-21 Snap Inc. Selecting avatars to be included in the video being generated on demand
US11996113B2 (en) 2021-10-29 2024-05-28 Snap Inc. Voice notes with changing effects
US11995757B2 (en) 2021-10-29 2024-05-28 Snap Inc. Customized animation from video
US12002146B2 (en) 2022-03-28 2024-06-04 Snap Inc. 3D modeling based on neural light field
US12008811B2 (en) 2020-12-30 2024-06-11 Snap Inc. Machine learning-based selection of a representative video frame within a messaging application
US12020384B2 (en) 2022-06-21 2024-06-25 Snap Inc. Integrating augmented reality experiences with other components

Families Citing this family (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9516259B2 (en) * 2013-10-22 2016-12-06 Google Inc. Capturing media content in accordance with a viewer expression
CN103945161B (en) * 2014-04-14 2017-06-27 联想(北京)有限公司 A kind of information processing method and electronic equipment
KR101535574B1 (en) * 2014-07-18 2015-07-10 오용운 System and method for providing social network emoticon using 3d character
KR101681501B1 (en) * 2016-06-28 2016-12-01 (주) 키글 System and method for creating face avatar
KR102616172B1 (en) * 2016-08-12 2023-12-19 주식회사 케이티 System for character providing and information gathering method using same
KR102120871B1 (en) * 2017-11-08 2020-06-09 주식회사 하이퍼커넥트 Terminal and server providing a video call service
WO2019103484A1 (en) * 2017-11-24 2019-05-31 주식회사 제네시스랩 Multi-modal emotion recognition device, method and storage medium using artificial intelligence
US11012389B2 (en) 2018-05-07 2021-05-18 Apple Inc. Modifying images with supplemental content for messaging
US10681310B2 (en) 2018-05-07 2020-06-09 Apple Inc. Modifying video streams with supplemental content for video conferencing
KR102647656B1 (en) * 2018-09-04 2024-03-15 삼성전자주식회사 Electronic device displaying additional object on augmented reality image and method for driving the electronic device
KR102611458B1 (en) * 2018-09-06 2023-12-11 주식회사 아이앤나 Method for Providing Augmented Reality by Baby's Emotional Sate using Baby's Peripheral Region
KR102648993B1 (en) * 2018-12-21 2024-03-20 삼성전자주식회사 Electronic device for providing avatar based on emotion state of user and method thereof
CN109831638B (en) * 2019-01-23 2021-01-08 广州视源电子科技股份有限公司 Video image transmission method and device, interactive intelligent panel and storage medium
JP6581742B1 (en) * 2019-03-27 2019-09-25 株式会社ドワンゴ VR live broadcast distribution system, distribution server, distribution server control method, distribution server program, and VR raw photo data structure
KR102236718B1 (en) * 2019-07-25 2021-04-06 주식회사 모두커뮤니케이션 Apparatus and method for creating personalized objects with emotion reflected
KR102114457B1 (en) * 2019-10-21 2020-05-22 (주)부즈 Method and apparatus for processing real-time character streaming contents
KR102260022B1 (en) * 2020-05-25 2021-06-02 전남대학교산학협력단 System and method for object classification in image based on deep running
KR102637373B1 (en) * 2021-01-26 2024-02-19 주식회사 플랫팜 Apparatus and method for generating emoticon
KR20240077627A (en) * 2022-11-24 2024-06-03 주식회사 피씨엔 User emotion interaction method and system for extended reality based on non-verbal elements

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2008004844A1 (en) * 2006-07-06 2008-01-10 Ktfreetel Co., Ltd. Method and system for providing voice analysis service, and apparatus therefor
KR20080057030A (en) * 2006-12-19 2008-06-24 엘지전자 주식회사 Apparatus and method for image communication inserting emoticon
KR100868638B1 (en) * 2007-08-07 2008-11-12 에스케이 텔레콤주식회사 System and method for balloon providing during video communication
KR20110025721A (en) * 2009-09-05 2011-03-11 에스케이텔레콤 주식회사 System and method for delivering feeling during video call
US20110122219A1 (en) * 2009-11-23 2011-05-26 Samsung Electronics Co. Ltd. Method and apparatus for video call in a mobile terminal

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2008004844A1 (en) * 2006-07-06 2008-01-10 Ktfreetel Co., Ltd. Method and system for providing voice analysis service, and apparatus therefor
KR20080057030A (en) * 2006-12-19 2008-06-24 엘지전자 주식회사 Apparatus and method for image communication inserting emoticon
KR100868638B1 (en) * 2007-08-07 2008-11-12 에스케이 텔레콤주식회사 System and method for balloon providing during video communication
KR20110025721A (en) * 2009-09-05 2011-03-11 에스케이텔레콤 주식회사 System and method for delivering feeling during video call
US20110122219A1 (en) * 2009-11-23 2011-05-26 Samsung Electronics Co. Ltd. Method and apparatus for video call in a mobile terminal

Cited By (296)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11425068B2 (en) 2009-02-03 2022-08-23 Snap Inc. Interactive avatar in messaging environment
US11607616B2 (en) 2012-05-08 2023-03-21 Snap Inc. System and method for generating and displaying avatars
US11925869B2 (en) 2012-05-08 2024-03-12 Snap Inc. System and method for generating and displaying avatars
US11229849B2 (en) 2012-05-08 2022-01-25 Snap Inc. System and method for generating and displaying avatars
US9710970B2 (en) 2013-05-09 2017-07-18 Samsung Electronics Co., Ltd. Method and apparatus for providing contents including augmented reality information
WO2014182052A1 (en) * 2013-05-09 2014-11-13 Samsung Electronics Co., Ltd. Method and apparatus for providing contents including augmented reality information
US11651797B2 (en) 2014-02-05 2023-05-16 Snap Inc. Real time video processing for changing proportions of an object in the video
US10991395B1 (en) 2014-02-05 2021-04-27 Snap Inc. Method for real time video processing involving changing a color of an object on a human face in a video
US11443772B2 (en) 2014-02-05 2022-09-13 Snap Inc. Method for triggering events in a video
GB2529037B (en) * 2014-06-10 2018-05-23 2Mee Ltd Augmented reality apparatus and method
US10262195B2 (en) 2014-10-27 2019-04-16 Mattersight Corporation Predictive and responsive video analytics system and methods
US9437215B2 (en) 2014-10-27 2016-09-06 Mattersight Corporation Predictive video analytics system and methods
US9269374B1 (en) 2014-10-27 2016-02-23 Mattersight Corporation Predictive video analytics system and methods
WO2016077578A1 (en) * 2014-11-13 2016-05-19 Intel Corporation System and method for feature-based authentication
US9811649B2 (en) 2014-11-13 2017-11-07 Intel Corporation System and method for feature-based authentication
WO2016163565A1 (en) * 2015-04-05 2016-10-13 한신대학교 산학협력단 Multi-modal multi-agent-based emotional communication system
KR101652486B1 (en) * 2015-04-05 2016-08-30 주식회사 큐버 Sentiment communication system based on multiple multimodal agents
KR101743763B1 (en) * 2015-06-29 2017-06-05 (주)참빛솔루션 Method for providng smart learning education based on sensitivity avatar emoticon, and smart learning education device for the same
US11631276B2 (en) 2016-03-31 2023-04-18 Snap Inc. Automated avatar generation
US11048916B2 (en) 2016-03-31 2021-06-29 Snap Inc. Automated avatar generation
US11662900B2 (en) 2016-05-31 2023-05-30 Snap Inc. Application control using a gesture based trigger
CN106157262A (en) * 2016-06-28 2016-11-23 广东欧珀移动通信有限公司 The processing method of a kind of augmented reality, device and mobile terminal
CN106157262B (en) * 2016-06-28 2020-04-17 Oppo广东移动通信有限公司 Augmented reality processing method and device and mobile terminal
CN106157363A (en) * 2016-06-28 2016-11-23 广东欧珀移动通信有限公司 A kind of photographic method based on augmented reality, device and mobile terminal
CN106127828A (en) * 2016-06-28 2016-11-16 广东欧珀移动通信有限公司 The processing method of a kind of augmented reality, device and mobile terminal
CN106127829B (en) * 2016-06-28 2020-06-30 Oppo广东移动通信有限公司 Augmented reality processing method and device and terminal
CN106127829A (en) * 2016-06-28 2016-11-16 广东欧珀移动通信有限公司 The processing method of a kind of augmented reality, device and terminal
US10984569B2 (en) 2016-06-30 2021-04-20 Snap Inc. Avatar based ideogram generation
US11438288B2 (en) 2016-07-19 2022-09-06 Snap Inc. Displaying customized electronic messaging graphics
US10848446B1 (en) 2016-07-19 2020-11-24 Snap Inc. Displaying customized electronic messaging graphics
US11509615B2 (en) 2016-07-19 2022-11-22 Snap Inc. Generating customized electronic messaging graphics
US11418470B2 (en) 2016-07-19 2022-08-16 Snap Inc. Displaying customized electronic messaging graphics
US10855632B2 (en) 2016-07-19 2020-12-01 Snap Inc. Displaying customized electronic messaging graphics
US11962598B2 (en) 2016-10-10 2024-04-16 Snap Inc. Social media post subscribe requests for buffer user accounts
US11438341B1 (en) 2016-10-10 2022-09-06 Snap Inc. Social media post subscribe requests for buffer user accounts
US11100311B2 (en) 2016-10-19 2021-08-24 Snap Inc. Neural networks for facial modeling
US10880246B2 (en) 2016-10-24 2020-12-29 Snap Inc. Generating and displaying customized avatars in electronic messages
US11580700B2 (en) 2016-10-24 2023-02-14 Snap Inc. Augmented reality object manipulation
US11218433B2 (en) 2016-10-24 2022-01-04 Snap Inc. Generating and displaying customized avatars in electronic messages
US10938758B2 (en) 2016-10-24 2021-03-02 Snap Inc. Generating and displaying customized avatars in media overlays
US11876762B1 (en) 2016-10-24 2024-01-16 Snap Inc. Generating and displaying customized avatars in media overlays
US11843456B2 (en) 2016-10-24 2023-12-12 Snap Inc. Generating and displaying customized avatars in media overlays
US11616745B2 (en) 2017-01-09 2023-03-28 Snap Inc. Contextual generation and selection of customized media content
US11704878B2 (en) 2017-01-09 2023-07-18 Snap Inc. Surface aware lens
US11989809B2 (en) 2017-01-16 2024-05-21 Snap Inc. Coded vision system
US11544883B1 (en) 2017-01-16 2023-01-03 Snap Inc. Coded vision system
US11991130B2 (en) 2017-01-18 2024-05-21 Snap Inc. Customized contextual media content item generation
US10951562B2 (en) 2017-01-18 2021-03-16 Snap. Inc. Customized contextual media content item generation
US11870743B1 (en) 2017-01-23 2024-01-09 Snap Inc. Customized digital avatar accessories
US10043406B1 (en) 2017-03-10 2018-08-07 Intel Corporation Augmented emotion display for austistic persons
CN110431838B (en) * 2017-03-22 2022-03-29 韩国斯诺有限公司 Method and system for providing dynamic content of face recognition camera
CN110431838A (en) * 2017-03-22 2019-11-08 韩国斯诺有限公司 The method and system of the dynamic content of recognition of face video camera are provided
US11593980B2 (en) 2017-04-20 2023-02-28 Snap Inc. Customized user interface for electronic communications
US11069103B1 (en) 2017-04-20 2021-07-20 Snap Inc. Customized user interface for electronic communications
US11392264B1 (en) 2017-04-27 2022-07-19 Snap Inc. Map-based graphical user interface for multi-type social media galleries
US11842411B2 (en) 2017-04-27 2023-12-12 Snap Inc. Location-based virtual avatars
US10963529B1 (en) 2017-04-27 2021-03-30 Snap Inc. Location-based search mechanism in a graphical user interface
US10952013B1 (en) 2017-04-27 2021-03-16 Snap Inc. Selective location-based identity communication
US11385763B2 (en) 2017-04-27 2022-07-12 Snap Inc. Map-based graphical user interface indicating geospatial activity metrics
US11451956B1 (en) 2017-04-27 2022-09-20 Snap Inc. Location privacy management on map-based social media platforms
US11418906B2 (en) 2017-04-27 2022-08-16 Snap Inc. Selective location-based identity communication
US11782574B2 (en) 2017-04-27 2023-10-10 Snap Inc. Map-based graphical user interface indicating geospatial activity metrics
US11474663B2 (en) 2017-04-27 2022-10-18 Snap Inc. Location-based search mechanism in a graphical user interface
US11893647B2 (en) 2017-04-27 2024-02-06 Snap Inc. Location-based virtual avatars
US11995288B2 (en) 2017-04-27 2024-05-28 Snap Inc. Location-based search mechanism in a graphical user interface
US11830209B2 (en) 2017-05-26 2023-11-28 Snap Inc. Neural network-based image stream modification
US11882162B2 (en) 2017-07-28 2024-01-23 Snap Inc. Software application manager for messaging applications
US11122094B2 (en) 2017-07-28 2021-09-14 Snap Inc. Software application manager for messaging applications
US11659014B2 (en) 2017-07-28 2023-05-23 Snap Inc. Software application manager for messaging applications
CN111183455A (en) * 2017-08-29 2020-05-19 互曼人工智能科技(上海)有限公司 Image data processing system and method
US11610354B2 (en) 2017-10-26 2023-03-21 Snap Inc. Joint audio-video facial animation system
US11120597B2 (en) 2017-10-26 2021-09-14 Snap Inc. Joint audio-video facial animation system
US11930055B2 (en) 2017-10-30 2024-03-12 Snap Inc. Animated chat presence
US11030789B2 (en) 2017-10-30 2021-06-08 Snap Inc. Animated chat presence
US11354843B2 (en) 2017-10-30 2022-06-07 Snap Inc. Animated chat presence
US11706267B2 (en) 2017-10-30 2023-07-18 Snap Inc. Animated chat presence
US11460974B1 (en) 2017-11-28 2022-10-04 Snap Inc. Content discovery refresh
CN109840009A (en) * 2017-11-28 2019-06-04 浙江思考者科技有限公司 A kind of intelligence true man's advertisement screen interactive system and implementation method
US10936157B2 (en) 2017-11-29 2021-03-02 Snap Inc. Selectable item including a customized graphic for an electronic messaging application
US11411895B2 (en) 2017-11-29 2022-08-09 Snap Inc. Generating aggregated media content items for a group of users in an electronic messaging application
US10554698B2 (en) 2017-12-28 2020-02-04 Hyperconnect, Inc. Terminal and server providing video call service
US11769259B2 (en) 2018-01-23 2023-09-26 Snap Inc. Region-based stabilized face tracking
US10949648B1 (en) 2018-01-23 2021-03-16 Snap Inc. Region-based stabilized face tracking
US11688119B2 (en) 2018-02-28 2023-06-27 Snap Inc. Animated expressive icon
US11120601B2 (en) 2018-02-28 2021-09-14 Snap Inc. Animated expressive icon
US10979752B1 (en) 2018-02-28 2021-04-13 Snap Inc. Generating media content items based on location information
US11523159B2 (en) 2018-02-28 2022-12-06 Snap Inc. Generating media content items based on location information
US11468618B2 (en) 2018-02-28 2022-10-11 Snap Inc. Animated expressive icon
US11880923B2 (en) 2018-02-28 2024-01-23 Snap Inc. Animated expressive icon
US11310176B2 (en) 2018-04-13 2022-04-19 Snap Inc. Content suggestion system
WO2019204464A1 (en) * 2018-04-18 2019-10-24 Snap Inc. Augmented expression system
US10719968B2 (en) 2018-04-18 2020-07-21 Snap Inc. Augmented expression system
US11875439B2 (en) 2018-04-18 2024-01-16 Snap Inc. Augmented expression system
KR20240027845A (en) 2018-04-18 2024-03-04 스냅 인코포레이티드 Augmented expression system
CN108830917A (en) * 2018-05-29 2018-11-16 努比亚技术有限公司 A kind of information generating method, terminal and computer readable storage medium
CN108830917B (en) * 2018-05-29 2023-04-18 努比亚技术有限公司 Information generation method, terminal and computer readable storage medium
CN108961431A (en) * 2018-07-03 2018-12-07 百度在线网络技术(北京)有限公司 Generation method, device and the terminal device of facial expression
US11074675B2 (en) 2018-07-31 2021-07-27 Snap Inc. Eye texture inpainting
US11030813B2 (en) 2018-08-30 2021-06-08 Snap Inc. Video clip object tracking
US11715268B2 (en) 2018-08-30 2023-08-01 Snap Inc. Video clip object tracking
CN110874137B (en) * 2018-08-31 2023-06-13 阿里巴巴集团控股有限公司 Interaction method and device
CN110874137A (en) * 2018-08-31 2020-03-10 阿里巴巴集团控股有限公司 Interaction method and device
US10896534B1 (en) 2018-09-19 2021-01-19 Snap Inc. Avatar style transformation using neural networks
US11348301B2 (en) 2018-09-19 2022-05-31 Snap Inc. Avatar style transformation using neural networks
US11868590B2 (en) 2018-09-25 2024-01-09 Snap Inc. Interface to display shared user groups
US10895964B1 (en) 2018-09-25 2021-01-19 Snap Inc. Interface to display shared user groups
US11294545B2 (en) 2018-09-25 2022-04-05 Snap Inc. Interface to display shared user groups
US11477149B2 (en) 2018-09-28 2022-10-18 Snap Inc. Generating customized graphics having reactions to electronic message content
US11610357B2 (en) 2018-09-28 2023-03-21 Snap Inc. System and method of generating targeted user lists using customizable avatar characteristics
US11704005B2 (en) 2018-09-28 2023-07-18 Snap Inc. Collaborative achievement interface
US11455082B2 (en) 2018-09-28 2022-09-27 Snap Inc. Collaborative achievement interface
US10904181B2 (en) 2018-09-28 2021-01-26 Snap Inc. Generating customized graphics having reactions to electronic message content
US11824822B2 (en) 2018-09-28 2023-11-21 Snap Inc. Generating customized graphics having reactions to electronic message content
US11171902B2 (en) 2018-09-28 2021-11-09 Snap Inc. Generating customized graphics having reactions to electronic message content
US11189070B2 (en) 2018-09-28 2021-11-30 Snap Inc. System and method of generating targeted user lists using customizable avatar characteristics
US11245658B2 (en) 2018-09-28 2022-02-08 Snap Inc. System and method of generating private notifications between users in a communication session
US10872451B2 (en) 2018-10-31 2020-12-22 Snap Inc. 3D avatar rendering
US11103795B1 (en) 2018-10-31 2021-08-31 Snap Inc. Game drawer
US11321896B2 (en) 2018-10-31 2022-05-03 Snap Inc. 3D avatar rendering
US11836859B2 (en) 2018-11-27 2023-12-05 Snap Inc. Textured mesh building
US11176737B2 (en) 2018-11-27 2021-11-16 Snap Inc. Textured mesh building
US20220044479A1 (en) 2018-11-27 2022-02-10 Snap Inc. Textured mesh building
US11620791B2 (en) 2018-11-27 2023-04-04 Snap Inc. Rendering 3D captions within real-world environments
US11887237B2 (en) 2018-11-28 2024-01-30 Snap Inc. Dynamic composite user identifier
US10902661B1 (en) 2018-11-28 2021-01-26 Snap Inc. Dynamic composite user identifier
US11199957B1 (en) 2018-11-30 2021-12-14 Snap Inc. Generating customized avatars based on location information
US11315259B2 (en) 2018-11-30 2022-04-26 Snap Inc. Efficient human pose tracking in videos
US11698722B2 (en) 2018-11-30 2023-07-11 Snap Inc. Generating customized avatars based on location information
US10861170B1 (en) 2018-11-30 2020-12-08 Snap Inc. Efficient human pose tracking in videos
US11783494B2 (en) 2018-11-30 2023-10-10 Snap Inc. Efficient human pose tracking in videos
US11055514B1 (en) 2018-12-14 2021-07-06 Snap Inc. Image face manipulation
US11798261B2 (en) 2018-12-14 2023-10-24 Snap Inc. Image face manipulation
CN111353842A (en) * 2018-12-24 2020-06-30 阿里巴巴集团控股有限公司 Processing method and system of push information
US11516173B1 (en) 2018-12-26 2022-11-29 Snap Inc. Message composition interface
CN109727303B (en) * 2018-12-29 2023-07-25 广州方硅信息技术有限公司 Video display method, system, computer equipment, storage medium and terminal
CN109727303A (en) * 2018-12-29 2019-05-07 广州华多网络科技有限公司 Video display method, system, computer equipment, storage medium and terminal
US11032670B1 (en) 2019-01-14 2021-06-08 Snap Inc. Destination sharing in location sharing system
US11877211B2 (en) 2019-01-14 2024-01-16 Snap Inc. Destination sharing in location sharing system
US10939246B1 (en) 2019-01-16 2021-03-02 Snap Inc. Location-based context information sharing in a messaging system
US11751015B2 (en) 2019-01-16 2023-09-05 Snap Inc. Location-based context information sharing in a messaging system
US10945098B2 (en) 2019-01-16 2021-03-09 Snap Inc. Location-based context information sharing in a messaging system
US11693887B2 (en) 2019-01-30 2023-07-04 Snap Inc. Adaptive spatial density based clustering
US11294936B1 (en) 2019-01-30 2022-04-05 Snap Inc. Adaptive spatial density based clustering
US11557075B2 (en) 2019-02-06 2023-01-17 Snap Inc. Body pose estimation
US10984575B2 (en) 2019-02-06 2021-04-20 Snap Inc. Body pose estimation
US11714524B2 (en) 2019-02-06 2023-08-01 Snap Inc. Global event-based avatar
US11010022B2 (en) 2019-02-06 2021-05-18 Snap Inc. Global event-based avatar
US11275439B2 (en) 2019-02-13 2022-03-15 Snap Inc. Sleep detection in a location sharing system
US10936066B1 (en) 2019-02-13 2021-03-02 Snap Inc. Sleep detection in a location sharing system
US11809624B2 (en) 2019-02-13 2023-11-07 Snap Inc. Sleep detection in a location sharing system
US10964082B2 (en) 2019-02-26 2021-03-30 Snap Inc. Avatar based on weather
US11574431B2 (en) 2019-02-26 2023-02-07 Snap Inc. Avatar based on weather
US10852918B1 (en) 2019-03-08 2020-12-01 Snap Inc. Contextual information in chat
US11301117B2 (en) 2019-03-08 2022-04-12 Snap Inc. Contextual information in chat
US11868414B1 (en) 2019-03-14 2024-01-09 Snap Inc. Graph-based prediction for contact suggestion in a location sharing system
US11852554B1 (en) 2019-03-21 2023-12-26 Snap Inc. Barometer calibration in a location sharing system
US11638115B2 (en) 2019-03-28 2023-04-25 Snap Inc. Points of interest in a location sharing system
US11166123B1 (en) 2019-03-28 2021-11-02 Snap Inc. Grouped transmission of location data in a location sharing system
US11039270B2 (en) 2019-03-28 2021-06-15 Snap Inc. Points of interest in a location sharing system
CN110046336A (en) * 2019-04-15 2019-07-23 南京孜博汇信息科技有限公司 Position encoded sheet disposal method and system
US11973732B2 (en) 2019-04-30 2024-04-30 Snap Inc. Messaging system with avatar generation
US10992619B2 (en) 2019-04-30 2021-04-27 Snap Inc. Messaging system with avatar generation
CN111918015A (en) * 2019-05-07 2020-11-10 阿瓦亚公司 Video call routing and management of facial emotions based on artificial intelligence determination
USD916810S1 (en) 2019-05-28 2021-04-20 Snap Inc. Display screen or portion thereof with a graphical user interface
USD916811S1 (en) 2019-05-28 2021-04-20 Snap Inc. Display screen or portion thereof with a transitional graphical user interface
USD916809S1 (en) 2019-05-28 2021-04-20 Snap Inc. Display screen or portion thereof with a transitional graphical user interface
USD916872S1 (en) 2019-05-28 2021-04-20 Snap Inc. Display screen or portion thereof with a graphical user interface
USD916871S1 (en) 2019-05-28 2021-04-20 Snap Inc. Display screen or portion thereof with a transitional graphical user interface
US11917495B2 (en) 2019-06-07 2024-02-27 Snap Inc. Detection of a physical collision between two client devices in a location sharing system
US11601783B2 (en) 2019-06-07 2023-03-07 Snap Inc. Detection of a physical collision between two client devices in a location sharing system
US10893385B1 (en) 2019-06-07 2021-01-12 Snap Inc. Detection of a physical collision between two client devices in a location sharing system
US11188190B2 (en) 2019-06-28 2021-11-30 Snap Inc. Generating animation overlays in a communication session
US11443491B2 (en) 2019-06-28 2022-09-13 Snap Inc. 3D object camera customization system
US11676199B2 (en) 2019-06-28 2023-06-13 Snap Inc. Generating customizable avatar outfits
US11189098B2 (en) 2019-06-28 2021-11-30 Snap Inc. 3D object camera customization system
US11823341B2 (en) 2019-06-28 2023-11-21 Snap Inc. 3D object camera customization system
US11714535B2 (en) 2019-07-11 2023-08-01 Snap Inc. Edge gesture interface with smart interactions
US11307747B2 (en) 2019-07-11 2022-04-19 Snap Inc. Edge gesture interface with smart interactions
US11455081B2 (en) 2019-08-05 2022-09-27 Snap Inc. Message thread prioritization interface
US10911387B1 (en) 2019-08-12 2021-02-02 Snap Inc. Message reminder interface
US11588772B2 (en) 2019-08-12 2023-02-21 Snap Inc. Message reminder interface
US11956192B2 (en) 2019-08-12 2024-04-09 Snap Inc. Message reminder interface
CN110705356A (en) * 2019-08-31 2020-01-17 深圳市大拿科技有限公司 Function control method and related equipment
CN110705356B (en) * 2019-08-31 2023-12-29 深圳市大拿科技有限公司 Function control method and related equipment
US11662890B2 (en) 2019-09-16 2023-05-30 Snap Inc. Messaging system with battery level sharing
US11320969B2 (en) 2019-09-16 2022-05-03 Snap Inc. Messaging system with battery level sharing
US11822774B2 (en) 2019-09-16 2023-11-21 Snap Inc. Messaging system with battery level sharing
US11425062B2 (en) 2019-09-27 2022-08-23 Snap Inc. Recommended content viewed by friends
US11270491B2 (en) 2019-09-30 2022-03-08 Snap Inc. Dynamic parameterized user avatar stories
US11676320B2 (en) 2019-09-30 2023-06-13 Snap Inc. Dynamic media collection generation
US11080917B2 (en) 2019-09-30 2021-08-03 Snap Inc. Dynamic parameterized user avatar stories
US11218838B2 (en) 2019-10-31 2022-01-04 Snap Inc. Focused map-based context information surfacing
US11063891B2 (en) 2019-12-03 2021-07-13 Snap Inc. Personalized avatar notification
US11563702B2 (en) 2019-12-03 2023-01-24 Snap Inc. Personalized avatar notification
WO2021114710A1 (en) * 2019-12-09 2021-06-17 上海幻电信息科技有限公司 Live streaming video interaction method and apparatus, and computer device
US11582176B2 (en) 2019-12-09 2023-02-14 Snap Inc. Context sensitive avatar captions
US11778263B2 (en) 2019-12-09 2023-10-03 Shanghai Hode Information Technology Co., Ltd. Live streaming video interaction method and apparatus, and computer device
US11128586B2 (en) 2019-12-09 2021-09-21 Snap Inc. Context sensitive avatar captions
US11594025B2 (en) 2019-12-11 2023-02-28 Snap Inc. Skeletal tracking using previous frames
US11036989B1 (en) 2019-12-11 2021-06-15 Snap Inc. Skeletal tracking using previous frames
US11227442B1 (en) 2019-12-19 2022-01-18 Snap Inc. 3D captions with semantic graphical elements
US11810220B2 (en) 2019-12-19 2023-11-07 Snap Inc. 3D captions with face tracking
US11908093B2 (en) 2019-12-19 2024-02-20 Snap Inc. 3D captions with semantic graphical elements
US11263817B1 (en) 2019-12-19 2022-03-01 Snap Inc. 3D captions with face tracking
US11636657B2 (en) 2019-12-19 2023-04-25 Snap Inc. 3D captions with semantic graphical elements
CN111191564A (en) * 2019-12-26 2020-05-22 三盟科技股份有限公司 Multi-pose face emotion recognition method and system based on multi-angle neural network
US11128715B1 (en) 2019-12-30 2021-09-21 Snap Inc. Physical friend proximity in chat
US11140515B1 (en) 2019-12-30 2021-10-05 Snap Inc. Interfaces for relative device positioning
US11169658B2 (en) 2019-12-31 2021-11-09 Snap Inc. Combined map icon with action indicator
US11893208B2 (en) 2019-12-31 2024-02-06 Snap Inc. Combined map icon with action indicator
CN113099150A (en) * 2020-01-08 2021-07-09 华为技术有限公司 Image processing method, device and system
US11263254B2 (en) 2020-01-30 2022-03-01 Snap Inc. Video generation system to render frames on demand using a fleet of servers
US11651539B2 (en) 2020-01-30 2023-05-16 Snap Inc. System for generating media content items on demand
US11651022B2 (en) 2020-01-30 2023-05-16 Snap Inc. Video generation system to render frames on demand using a fleet of servers
US11036781B1 (en) 2020-01-30 2021-06-15 Snap Inc. Video generation system to render frames on demand using a fleet of servers
US11284144B2 (en) 2020-01-30 2022-03-22 Snap Inc. Video generation system to render frames on demand using a fleet of GPUs
US11729441B2 (en) 2020-01-30 2023-08-15 Snap Inc. Video generation system to render frames on demand
US11831937B2 (en) 2020-01-30 2023-11-28 Snap Inc. Video generation system to render frames on demand using a fleet of GPUS
US11991419B2 (en) 2020-01-30 2024-05-21 Snap Inc. Selecting avatars to be included in the video being generated on demand
US11356720B2 (en) 2020-01-30 2022-06-07 Snap Inc. Video generation system to render frames on demand
US11619501B2 (en) 2020-03-11 2023-04-04 Snap Inc. Avatar based on trip
US11775165B2 (en) 2020-03-16 2023-10-03 Snap Inc. 3D cutout image modification
US11217020B2 (en) 2020-03-16 2022-01-04 Snap Inc. 3D cutout image modification
US11625873B2 (en) 2020-03-30 2023-04-11 Snap Inc. Personalized media overlay recommendation
US11818286B2 (en) 2020-03-30 2023-11-14 Snap Inc. Avatar recommendation and reply
US11978140B2 (en) 2020-03-30 2024-05-07 Snap Inc. Personalized media overlay recommendation
US11969075B2 (en) 2020-03-31 2024-04-30 Snap Inc. Augmented reality beauty product tutorials
US11956190B2 (en) 2020-05-08 2024-04-09 Snap Inc. Messaging system with a carousel of related entities
US11822766B2 (en) 2020-06-08 2023-11-21 Snap Inc. Encoded image based messaging system
US11543939B2 (en) 2020-06-08 2023-01-03 Snap Inc. Encoded image based messaging system
US11922010B2 (en) 2020-06-08 2024-03-05 Snap Inc. Providing contextual information with keyboard interface for messaging system
US11683280B2 (en) 2020-06-10 2023-06-20 Snap Inc. Messaging system including an external-resource dock and drawer
US11580682B1 (en) 2020-06-30 2023-02-14 Snap Inc. Messaging system with augmented reality makeup
CN111773676A (en) * 2020-07-23 2020-10-16 网易(杭州)网络有限公司 Method and device for determining virtual role action
US11863513B2 (en) 2020-08-31 2024-01-02 Snap Inc. Media content playback and comments management
US11360733B2 (en) 2020-09-10 2022-06-14 Snap Inc. Colocated shared augmented reality without shared backend
US11893301B2 (en) 2020-09-10 2024-02-06 Snap Inc. Colocated shared augmented reality without shared backend
US11452939B2 (en) 2020-09-21 2022-09-27 Snap Inc. Graphical marker generation system for synchronizing users
US11888795B2 (en) 2020-09-21 2024-01-30 Snap Inc. Chats with micro sound clips
US11833427B2 (en) 2020-09-21 2023-12-05 Snap Inc. Graphical marker generation system for synchronizing users
US11910269B2 (en) 2020-09-25 2024-02-20 Snap Inc. Augmented reality content items including user avatar to share location
CN112215929A (en) * 2020-10-10 2021-01-12 珠海格力电器股份有限公司 Virtual social data processing method, device and system
US11660022B2 (en) 2020-10-27 2023-05-30 Snap Inc. Adaptive skeletal joint smoothing
US11615592B2 (en) 2020-10-27 2023-03-28 Snap Inc. Side-by-side character animation from realtime 3D body motion capture
US12002175B2 (en) 2020-11-18 2024-06-04 Snap Inc. Real-time motion transfer for prosthetic limbs
US11748931B2 (en) 2020-11-18 2023-09-05 Snap Inc. Body animation sharing and remixing
US11734894B2 (en) 2020-11-18 2023-08-22 Snap Inc. Real-time motion transfer for prosthetic limbs
US11450051B2 (en) 2020-11-18 2022-09-20 Snap Inc. Personalized avatar real-time motion capture
CN114630135A (en) * 2020-12-11 2022-06-14 北京字跳网络技术有限公司 Live broadcast interaction method and device
WO2022143128A1 (en) * 2020-12-29 2022-07-07 华为技术有限公司 Video call method and apparatus based on avatar, and terminal
US12008811B2 (en) 2020-12-30 2024-06-11 Snap Inc. Machine learning-based selection of a representative video frame within a messaging application
CN113014471A (en) * 2021-01-18 2021-06-22 腾讯科技(深圳)有限公司 Session processing method, device, terminal and storage medium
CN113014471B (en) * 2021-01-18 2022-08-19 腾讯科技(深圳)有限公司 Session processing method, device, terminal and storage medium
US11790531B2 (en) 2021-02-24 2023-10-17 Snap Inc. Whole body segmentation
US11809633B2 (en) 2021-03-16 2023-11-07 Snap Inc. Mirroring device with pointing based navigation
US11978283B2 (en) 2021-03-16 2024-05-07 Snap Inc. Mirroring device with a hands-free mode
US11798201B2 (en) 2021-03-16 2023-10-24 Snap Inc. Mirroring device with whole-body outfits
US11734959B2 (en) 2021-03-16 2023-08-22 Snap Inc. Activating hands-free mode on mirroring device
US11908243B2 (en) 2021-03-16 2024-02-20 Snap Inc. Menu hierarchy navigation on electronic mirroring devices
US11544885B2 (en) 2021-03-19 2023-01-03 Snap Inc. Augmented reality experience based on physical items
US11562548B2 (en) 2021-03-22 2023-01-24 Snap Inc. True size eyewear in real time
US11636654B2 (en) 2021-05-19 2023-04-25 Snap Inc. AR-based connected portal shopping
US11941767B2 (en) 2021-05-19 2024-03-26 Snap Inc. AR-based connected portal shopping
US11941227B2 (en) 2021-06-30 2024-03-26 Snap Inc. Hybrid search system for customizable media
US11854069B2 (en) 2021-07-16 2023-12-26 Snap Inc. Personalized try-on ads
US11908083B2 (en) 2021-08-31 2024-02-20 Snap Inc. Deforming custom mesh based on body mesh
US11983462B2 (en) 2021-08-31 2024-05-14 Snap Inc. Conversation guided augmented reality experience
US11670059B2 (en) 2021-09-01 2023-06-06 Snap Inc. Controlling interactive fashion based on body gestures
US11673054B2 (en) 2021-09-07 2023-06-13 Snap Inc. Controlling AR games on fashion items
US11663792B2 (en) 2021-09-08 2023-05-30 Snap Inc. Body fitted accessory with physics simulation
US11900506B2 (en) 2021-09-09 2024-02-13 Snap Inc. Controlling interactive fashion based on facial expressions
US11734866B2 (en) 2021-09-13 2023-08-22 Snap Inc. Controlling interactive fashion based on voice
US11798238B2 (en) 2021-09-14 2023-10-24 Snap Inc. Blending body mesh into external mesh
US11836866B2 (en) 2021-09-20 2023-12-05 Snap Inc. Deforming real-world object using an external mesh
US11636662B2 (en) 2021-09-30 2023-04-25 Snap Inc. Body normal network light and rendering control
US11983826B2 (en) 2021-09-30 2024-05-14 Snap Inc. 3D upper garment tracking
US11651572B2 (en) 2021-10-11 2023-05-16 Snap Inc. Light and rendering of garments
US11836862B2 (en) 2021-10-11 2023-12-05 Snap Inc. External mesh with vertex attributes
US11790614B2 (en) 2021-10-11 2023-10-17 Snap Inc. Inferring intent from pose and speech input
US11763481B2 (en) 2021-10-20 2023-09-19 Snap Inc. Mirror-based augmented reality experience
US11995757B2 (en) 2021-10-29 2024-05-28 Snap Inc. Customized animation from video
US12020358B2 (en) 2021-10-29 2024-06-25 Snap Inc. Animated custom sticker creation
US11996113B2 (en) 2021-10-29 2024-05-28 Snap Inc. Voice notes with changing effects
US11960784B2 (en) 2021-12-07 2024-04-16 Snap Inc. Shared augmented reality unboxing experience
US11748958B2 (en) 2021-12-07 2023-09-05 Snap Inc. Augmented reality unboxing experience
US11880947B2 (en) 2021-12-21 2024-01-23 Snap Inc. Real-time upper-body garment exchange
US11928783B2 (en) 2021-12-30 2024-03-12 Snap Inc. AR position and orientation along a plane
US11887260B2 (en) 2021-12-30 2024-01-30 Snap Inc. AR position indicator
US11823346B2 (en) 2022-01-17 2023-11-21 Snap Inc. AR body part tracking system
US11954762B2 (en) 2022-01-19 2024-04-09 Snap Inc. Object replacement system
US12002146B2 (en) 2022-03-28 2024-06-04 Snap Inc. 3D modeling based on neural light field
US12020384B2 (en) 2022-06-21 2024-06-25 Snap Inc. Integrating augmented reality experiences with other components
US12020386B2 (en) 2022-06-23 2024-06-25 Snap Inc. Applying pregenerated virtual experiences in new location
US11870745B1 (en) 2022-06-28 2024-01-09 Snap Inc. Media gallery sharing and management
US11893166B1 (en) 2022-11-08 2024-02-06 Snap Inc. User avatar movement control using an augmented reality eyewear device
US12020377B2 (en) 2023-05-09 2024-06-25 Snap Inc. Textured mesh building

Also Published As

Publication number Publication date
KR20130022434A (en) 2013-03-07

Similar Documents

Publication Publication Date Title
WO2013027893A1 (en) Apparatus and method for emotional content services on telecommunication devices, apparatus and method for emotion recognition therefor, and apparatus and method for generating and matching the emotional content using same
US11736756B2 (en) Producing realistic body movement using body images
US20230283748A1 (en) Communication using interactive avatars
WO2020204000A1 (en) Communication assistance system, communication assistance method, communication assistance program, and image control program
US20190197755A1 (en) Producing realistic talking Face with Expression using Images text and voice
JP6616288B2 (en) Method, user terminal, and server for information exchange in communication
CN116797694A (en) Emotion symbol doll
US20190222806A1 (en) Communication system and method
US20090016617A1 (en) Sender dependent messaging viewer
CN109691054A (en) Animation user identifier
CN108874114B (en) Method and device for realizing emotion expression of virtual object, computer equipment and storage medium
US11151796B2 (en) Systems and methods for providing real-time composite video from multiple source devices featuring augmented reality elements
CN112199016B (en) Image processing method, image processing device, electronic equipment and computer readable storage medium
KR102148151B1 (en) Intelligent chat based on digital communication network
US11423627B2 (en) Systems and methods for providing real-time composite video from multiple source devices featuring augmented reality elements
KR20120018479A (en) Server and method for providing avatar using facial expression and gesture recognition
US11553009B2 (en) Information processing device, information processing method, and computer program for switching between communications performed in real space and virtual space
US20220291752A1 (en) Distributed Application Platform Projected on a Secondary Display for Entertainment, Gaming and Learning with Intelligent Gesture Interactions and Complex Input Composition for Control
KR20130082693A (en) Apparatus and method for video chatting using avatar
CN115396390B (en) Interaction method, system and device based on video chat and electronic equipment
JP2023099309A (en) Method, computer device, and computer program for interpreting voice of video into sign language through avatar
JP5894505B2 (en) Image communication system, image generation apparatus, and program
KR100736541B1 (en) System for unification personal character in online network
Jang et al. Mobile video communication based on augmented reality
CN111461005A (en) Gesture recognition method and device, computer equipment and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11871115

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 11871115

Country of ref document: EP

Kind code of ref document: A1