KR101802062B1 - Method and system to measure stereoscopic 3d content-induced visual discomfort based facial expression recognition for sensitivity tv - Google Patents
Method and system to measure stereoscopic 3d content-induced visual discomfort based facial expression recognition for sensitivity tv Download PDFInfo
- Publication number
- KR101802062B1 KR101802062B1 KR1020160030246A KR20160030246A KR101802062B1 KR 101802062 B1 KR101802062 B1 KR 101802062B1 KR 1020160030246 A KR1020160030246 A KR 1020160030246A KR 20160030246 A KR20160030246 A KR 20160030246A KR 101802062 B1 KR101802062 B1 KR 101802062B1
- Authority
- KR
- South Korea
- Prior art keywords
- viewer
- face
- content
- discomfort
- face image
- Prior art date
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/16—Devices for psychotechnics; Testing reaction times ; Devices for evaluating the psychological state
- A61B5/165—Evaluating the state of mind, e.g. depression, anxiety
-
- G06F15/18—
-
- G06K9/00288—
-
- G06K9/00302—
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Engineering & Computer Science (AREA)
- Animal Behavior & Ethology (AREA)
- Heart & Thoracic Surgery (AREA)
- Veterinary Medicine (AREA)
- Public Health (AREA)
- General Health & Medical Sciences (AREA)
- Psychiatry (AREA)
- Physics & Mathematics (AREA)
- Surgery (AREA)
- Biophysics (AREA)
- Pathology (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Medical Informatics (AREA)
- Developmental Disabilities (AREA)
- Child & Adolescent Psychology (AREA)
- Social Psychology (AREA)
- Psychology (AREA)
- Hospice & Palliative Care (AREA)
- Educational Technology (AREA)
- Dentistry (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Image Analysis (AREA)
Abstract
A method of measuring visual fatigue due to stereoscopic 3D content based on facial expression recognition for emotional TV comprises the steps of: acquiring a face image of the viewer when the 3D content is provided to a viewer; Performing face area detection / normalization on the face image; Extracting a facial feature from the normalized facial image based on at least one of a facial geometry or an appearance; And applying a machine learning algorithm to the facial feature to determine whether the viewer is uncomfortable.
Description
The following embodiments relate to a system and method for measuring visual fatigue due to stereoscopic 3D content based on facial expression recognition for emotional TV, and more particularly, to a system and method for measuring visual fatigue caused by a viewer's discomfort Quot ;. < / RTI >
Recently, there is an increasing demand for emotional TV technology for analyzing emotional information (for example, gestures, facial expressions, or melancholy) of viewers to search for / recommend suitable media or automatically change content episodes.
Accordingly, the conventional emotional TV technology mainly aims at classifying the basic emotions (for example, happiness, surprise, nausea, anger, fear, or sadness) caused by the experience of viewing 2D contents.
However, as content provided on TV changes from 2D to 3D, basic emotion classification for 3D contents is required. Particularly, stereoscopic 3D (S3D) TV can improve the perceived viewing quality by providing better presence than 2D TV, but it can also cause negative feeling of visual inconvenience. Therefore, A technique for measuring the visual discomfort due to watching the S3D contents is required.
Conventional discomfort measurement technology uses various sensing technologies such as eye movement, brain wave, and fMRI to measure the viewer's visual discomfort about 3D contents. However, considering the convenience of the viewer and the cost of the equipment, there is a limit to apply to the TV viewing environment.
Therefore, the following embodiments propose a camera image-based visual fatigue measurement technique which is relatively unrestricted.
In particular, since the correlation between the change in the facial expression of the viewer and the inconvenience is not revealed by the existing literature and the invention, it is difficult to immediately apply the technique for recognizing the change in the facial expression in the past. Accordingly, since the conventional techniques can not achieve the objects to be achieved by the following embodiments, the following embodiments provide a new discomfort analysis technique applying a technique of recognizing a facial expression change.
The following embodiments provide a method and system for measuring visual fatigue attributable to 3D content by analyzing a change in facial expression of a viewer based on a camera image.
Specifically, in the following embodiments, a facial feature of a viewer is extracted based on a face image of a viewer obtained through a camera, and a visual fatigue Measuring method and system.
According to an embodiment, there is provided a method of measuring visual fatigue due to stereoscopic 3D content based on facial expression recognition for emotional TV, comprising: obtaining a face image of the viewer when the 3D content is provided to a viewer; Performing face area detection / normalization on the face image; Extracting a facial feature from the normalized facial image based on at least one of a facial geometry or an appearance; And applying a machine learning algorithm to the facial feature to determine whether the viewer is uncomfortable.
Wherein extracting the facial feature from the normalized facial image comprises: extracting a face landmark from the normalized facial image; Calculating a displacement between the extracted face landmark and a reference face landmark previously extracted from a face image of the viewer in a normal time; And obtaining the displacement with the facial feature.
The step of determining whether the viewer is uncomfortable by applying the machine learning algorithm to the facial feature may include determining whether the viewer is uncomfortable by performing the machine learning algorithm with the displacement input.
Wherein the step of determining whether the viewer is uncomfortable by performing the machine learning algorithm based on the displacement is characterized in that when the deviation is greater than or equal to a preset reference value, And judging that there is a presence.
The step of performing face area detection / normalization on the face image includes: detecting a face area from the face image using an object detection algorithm; And performing normalization by performing scale / rotation compensation on the detected face area.
The step of applying the machine learning algorithm to the facial feature to determine whether the viewer is uncomfortable may further include adjusting a depth sense of the 3D content based on the determination result.
The adjusting the depth of the 3D content based on the determination result may include performing either disparity scaling or disparity shifting on the 3D content.
Wherein the step of determining whether the viewer feels uncomfortable by applying the machine learning algorithm to the facial feature comprises: collecting features of the 3D content when it is determined that the viewer's discomfort exists; Generating a discomfort pattern of the viewer by using the feature of the 3D content; And recommending at least one 3D content of the plurality of 3D contents to the viewer based on the discomfort pattern of the viewer.
According to one embodiment, a system for measuring visual fatigue based on facial expression recognition based stereoscopic 3D content for emotional TV includes a face image obtaining unit for obtaining a face image of the viewer when the 3D content is provided to a viewer; A face detection / normalization unit for performing face detection / normalization on the face image; A facial feature extraction unit for extracting a facial feature from the normalized facial image based on at least one of facial geometry and appearance; And a discomfort determination unit for determining whether the viewer feels uncomfortable by applying a machine learning algorithm to the facial feature.
The facial feature extraction unit extracts a face landmark from the normalized face image and calculates a displacement between the extracted face landmark and a reference face landmark extracted in advance from the normal facial image of the viewer , The displacement can be obtained with the facial feature.
The discomfort judgment unit may determine whether the viewer is uncomfortable by performing the machine learning algorithm with the displacement input.
The discomfort judgment unit may determine that the discomfort of the viewer is present when the defect value on which the machine learning algorithm is performed is equal to or greater than a preset reference value.
The face detection / normalization unit may detect a face region from the face image using an object detection algorithm, and perform normalization by performing scale / rotation compensation on the detected face region.
The discomfort determination unit may adjust the depth sense of the 3D content based on the determination result.
Wherein the discomfort determination unit collects the 3D feature information and generates a discomfort pattern of the viewer using the 3D feature if the discomfort of the viewer is determined as a result of the determination, It is possible to recommend at least one of the plurality of 3D contents to the viewer based on the discomfort pattern of the viewer.
The embodiments described below can provide a method and system for measuring visual fatigue attributable to 3D contents by analyzing changes in facial expression of a viewer based on a camera image.
Specifically, in the following embodiments, a facial feature of a viewer is extracted based on a face image of a viewer obtained through a camera, and a visual fatigue A measurement method and a system can be provided.
Therefore, the following embodiments are different from the discomfort measuring technique using various sensing technologies such as eye movements, brain waves, and fMRI, so that the embodiments do not use additional equipment, so that they are easy to implement, costly, A method and system for measuring visual fatigue can be provided.
Such a visual fatigue measurement method and system can be utilized as an emotional TV technology that enables an auto-interaction between 3D content and viewer's discomfort.
1 is a view for explaining the principle of a method of measuring visual fatigue according to an embodiment.
2 is a view for explaining a method of measuring visual fatigue according to an embodiment.
FIGS. 3A and 3B are views for explaining a process of extracting the facial feature shown in FIG. 2 and a process of determining whether the facial feature is uncomfortable.
4 is a view illustrating a result of performing a visual fatigue measurement method according to an embodiment.
5 is a diagram illustrating an emotional TV technique based on a visual fatigue measurement method according to an embodiment.
FIG. 6 is a diagram illustrating an emotional TV technique based on a visual fatigue measurement method according to another embodiment.
FIG. 7 is a flowchart illustrating a method of measuring visual fatigue according to an embodiment.
8 is a block diagram illustrating a visual fatigue measurement system according to an embodiment.
Hereinafter, embodiments according to the present invention will be described in detail with reference to the accompanying drawings. However, the present invention is not limited to or limited by the embodiments. In addition, the same reference numerals shown in the drawings denote the same members.
Also, terminologies used herein are terms used to properly represent preferred embodiments of the present invention, which may vary depending on the viewer, the intention of the operator, or the custom in the field to which the present invention belongs. Therefore, the definitions of these terms should be based on the contents throughout this specification.
1 is a view for explaining the principle of a method of measuring visual fatigue according to an embodiment.
Referring to FIG. 1, a method for measuring visual fatigue according to an embodiment is performed based on a face image of a viewer when 3D content is provided to a viewer.
For example, in the
Accordingly, the visual fatigue measurement method according to an embodiment analyzes the facial expression change of the viewer based on the
2 is a view for explaining a method of measuring visual fatigue according to an embodiment.
Referring to FIG. 2, a method for measuring visual fatigue according to an embodiment is performed by the visual
The visual
Next, the visual
The visual
That is, the facial expression
Thereafter, the visual
In this way, the visual
FIGS. 3A and 3B are views for explaining a process of extracting the facial feature shown in FIG. 2 and a process of determining whether the facial feature is uncomfortable.
3A, the facial feature extracting unit included in the visual
Specifically, the facial feature extraction unit extracts a face landmark from the face image on which the normalization has been performed (311), and calculates a displacement between the extracted face landmark and the reference face landmark extracted in advance from the normal face image of the viewer By calculating (312), the calculated displacement can be obtained as a facial feature.
3B, in general, a change in the facial expression of a viewer (particularly, a change in facial expression when there is an uncomfortable feeling) is detected by a face landmark 320 (see FIG. 3B), which is an eyeball, eye, nose, As shown in FIG. Thus, the change between the viewer's normal facial expression and the viewer's particular facial expression when the 3D content is presented is usually expressed as the displacement between the viewer's facial landmark and the viewer's facial landmark when the 3D content is presented .
Accordingly, the facial expression feature extraction unit uses the displacement between the reference face landmark of the viewer and the face landmark of the viewer when the 3D content is provided, as a facial feature to determine whether or not the viewer is uncomfortable.
For example, the facial feature extraction unit may calculate the displacement d by calculating the displacement d between the face landmark extracted from the normalized facial image and the reference face landmark as shown in Formula 1 (312) It can be used as a facial feature in judging process.
<
In
At this time, in the process of extracting the face landmark from the normalized face image, when the viewer wears the 3D glasses, the occlusion effect can not be ignored. Accordingly, the facial expression feature extraction unit may extract the face landmark from the normalized facial image using at least one of face geometry and shape so that the occlusion effect is minimized (311). In addition, the facial expression feature extraction unit may perform additional image processing operations such as histogram adjustment on the normalized facial image so that the occlusion effect is minimized.
In other words, if the machine learning algorithm is performed by inputting the displacement between the face landmark extracted by the facial feature extraction unit and the reference face landmark, it can be determined whether or not the viewer's discomfort exists (330).
Accordingly, the inconvenience determination unit included in the visual
In addition, the discomfort judgment unit not only determines whether the viewer is uncomfortable according to the defect value in which the machine learning algorithm is performed, by inputting the displacement between the face landmark extracted from the normalized face image and the reference face landmark, It is also possible to measure the degree of discomfort of the viewer (the parameter in which the viewer feels the discomfort scale).
4 is a view illustrating a result of performing a visual fatigue measurement method according to an embodiment.
Referring to FIG. 4, the visual fatigue system according to an exemplary embodiment of the present invention may be realized by inputting a displacement between a face landmark extracted from a normalized face image and a reference face landmark as input, as described with reference to FIGS. 3A through 3B. If the
In addition, the visual fatigue system inputs the displacement between the face landmark extracted from the face image subjected to the normalization and the reference face landmark, and then, according to the variation of the
Here, the visual fatigue system can use the
5 is a diagram illustrating an emotional TV technique based on a visual fatigue measurement method according to an embodiment.
Referring to FIG. 5, the visual
For example, the visual
At this time, the inconvenience determination unit may use various rendering techniques such as depth image-based rendering (DIBR) or warping-based image rendering for
In
As described above, the visual
FIG. 6 is a diagram illustrating an emotional TV technique based on a visual fatigue measurement method according to another embodiment.
Referring to FIG. 6, the visual
In general, susceptibility to visual discomfort per person for the same 3D content may be different. Accordingly, the visual
Accordingly, the inconvenience determination unit stores and maintains the generated discomfort pattern of the viewer in the personal inconvenience database, thereby extracting at least one 3D content suitable for the discomfort pattern of the viewer among the plurality of 3D contents from the 3D content database , And may be recommended to the viewer (640).
As described above, the visual
FIG. 7 is a flowchart illustrating a method of measuring visual fatigue according to an embodiment.
Referring to FIG. 7, a visual fatigue measurement system according to an embodiment obtains a face image of a viewer when 3D content is provided to a viewer (710).
Next, the visual fatigue measurement system performs face area detection / normalization on the face image (720).
For example, in
Next, the visual fatigue measurement system extracts the facial feature from the normalized facial image based on at least one of the facial geometry and the appearance (730).
Specifically, in
Thereafter, the visual fatigue measurement system applies a machine learning algorithm to the facial feature to determine whether the viewer is uncomfortable (740).
For example, in
At this time, the visual fatigue measurement system can determine that there is a feeling of discomfort of the viewer when the defect value in which the machine learning algorithm is performed by inputting the displacement is equal to or greater than a preset reference value. In addition, the visual fatigue measurement system can use the displacement value as input, and the result of the machine learning algorithm is used as the degree of inconvenience of the viewer.
Also, although not shown in the figure, in
Although it is not shown in the drawing, if the visual fatigue measurement system determines in
8 is a block diagram illustrating a visual fatigue measurement system according to an embodiment.
8, the system for measuring visual fatigue includes a face
The face
The face detection /
The facial expression
Specifically, the facial
The
At this time, the
The
If it is determined that the viewer's discomfort is present as a result of the determination, the
The apparatus described above may be implemented as a hardware component, a software component, and / or a combination of hardware components and software components. For example, the apparatus and components described in the embodiments may be implemented within a computer system, such as, for example, a processor, a controller, an arithmetic logic unit (ALU), a digital signal processor, a microcomputer, a field programmable array (FPA) A programmable logic unit (PLU), a microprocessor, or any other device capable of executing and responding to instructions. The processing device may execute an operating system (OS) and one or more software applications running on the operating system. The processing device may also access, store, manipulate, process, and generate data in response to execution of the software. For ease of understanding, the processing apparatus may be described as being used singly, but those skilled in the art will recognize that the processing apparatus may have a plurality of processing elements and / As shown in FIG. For example, the processing unit may comprise a plurality of processors or one processor and one controller. Other processing configurations are also possible, such as a parallel processor.
The software may include a computer program, code, instructions, or a combination of one or more of the foregoing, and may be configured to configure the processing device to operate as desired or to process it collectively or collectively Device can be commanded. The software and / or data may be in the form of any type of machine, component, physical device, virtual equipment, computer storage media, or device , Or may be permanently or temporarily embodied in a transmitted signal wave. The software may be distributed over a networked computer system and stored or executed in a distributed manner. The software and data may be stored on one or more computer readable recording media.
The method according to an embodiment may be implemented in the form of a program command that can be executed through various computer means and recorded in a computer-readable medium. The computer-readable medium may include program instructions, data files, data structures, and the like, alone or in combination. The program instructions to be recorded on the medium may be those specially designed and configured for the embodiments or may be available to those skilled in the art of computer software. Examples of computer-readable recording media include magnetic media such as hard disks, floppy disks and magnetic tape, optical recording media such as CD-ROMs and DVDs, magnetic recording media such as floptical disks, Includes hardware devices that are specially configured to store and execute program instructions such as magneto-OTPical media and ROM, RAM, flash memory, and the like. Examples of program instructions include machine language code such as those produced by a compiler, as well as high-level language code that can be executed by a computer using an interpreter or the like. The hardware devices described above may be configured to operate as one or more software modules to perform the operations of the embodiments, and vice versa.
While the present invention has been particularly shown and described with reference to exemplary embodiments thereof, it is to be understood that the invention is not limited to the disclosed exemplary embodiments. For example, it is to be understood that the techniques described may be performed in a different order than the described methods, and / or that components of the described systems, structures, devices, circuits, Lt; / RTI > or equivalents, even if it is replaced or replaced.
Therefore, other implementations, other embodiments, and equivalents to the claims are also within the scope of the following claims.
Claims (15)
Obtaining a face image of the viewer when the 3D content is provided to a viewer;
Performing face region detection and normalization on the face image;
Extracting a face landmark from the normalized face image based on at least one of face geometry or appearance;
Extracting a facial feature by calculating a displacement between the extracted face landmark and a reference face landmark previously extracted from the facial image of the viewer; And
Determining whether the viewer feels uncomfortable by applying a machine learning algorithm that inputs the displacement to the facial feature;
Lt; / RTI >
The step of determining whether the viewer is uncomfortable by applying the machine learning algorithm
Determining that there is an uncomfortable feeling of the viewer if the result value to which the machine learning algorithm is applied is greater than or equal to a preset reference value,
When the result of applying the machine learning algorithm is equal to or greater than a preset reference value, the step of determining that the viewer's discomfort exists
Measuring a degree of discomfort of the viewer according to a result value to which the machine learning algorithm is applied; measuring a degree of discomfort of the viewer based on a measure of discomfort when the viewer views the 3D content;
Determining a degree of controlling a depth sense of the 3D content provided to the viewer based on the degree of discomfort; And
Adjusting a depth sense of the 3D content according to a degree of controlling the depth feeling of the 3D content
Lt; / RTI >
Collecting characteristics of the 3D content when it is determined that the viewer's discomfort exists;
Generating a discomfort pattern of the viewer by using the feature of the 3D content; And
And recommending at least one of the plurality of 3D contents to the viewer based on the discomfort pattern of the viewer
And measuring the visual fatigue.
Wherein the step of performing face region detection and normalization on the face image comprises:
Detecting a face region from the face image using an object detection algorithm; And
Performing a scale on the detected face area, performing normalization by performing rotation compensation
Wherein the visual fatigue measuring method comprises the steps of:
The step of adjusting the depth of the 3D content
Performing either disparity scaling or disparity shifting on the 3D content;
Wherein the visual fatigue measuring method comprises the steps of:
A face image obtaining unit that obtains a face image of the viewer when the 3D content is provided to a viewer;
A face detection / normalization unit for performing face detection and normalization on the face image;
Extracting a face landmark from the normalized face image based on at least one of face geometry and appearance and extracting a face landmark extracted from the extracted face landmark and a face image A facial feature extracting unit for extracting a facial feature by calculating a displacement between reference face landmarks; And
A discomfort judgment unit for determining whether the viewer feels uncomfortable by applying a machine learning algorithm that inputs the displacement to the facial feature,
Lt; / RTI >
The discomfort judgment unit
When the result of applying the machine learning algorithm is equal to or greater than a preset reference value, it is determined that the viewer's discomfort is present,
A degree of discomfort of the viewer according to a result value to which the machine learning algorithm is applied, and a degree of the discomfort is a parameter that is a numerical value of a measure in which the viewer feels discomfort when watching the 3D content, And adjusting a depth sense of the 3D content according to a degree of adjusting a depth sense of the 3D content by determining a degree of adjusting a depth sense of the 3D content provided to the viewer,
Collecting a feature of the 3D content, generating a discomfort pattern of the viewer by using a feature of the 3D content when the discomfort of the viewer is judged to exist, The 3D content of at least one of the 3D contents of the 3D content is recommended to the viewer.
The face detection / normalization unit
A visual fatigue measurement system for detecting a face region from a face image using an object detection algorithm, performing scale on the detected face region, or performing normalization by performing rotation compensation.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020160030246A KR101802062B1 (en) | 2016-03-14 | 2016-03-14 | Method and system to measure stereoscopic 3d content-induced visual discomfort based facial expression recognition for sensitivity tv |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020160030246A KR101802062B1 (en) | 2016-03-14 | 2016-03-14 | Method and system to measure stereoscopic 3d content-induced visual discomfort based facial expression recognition for sensitivity tv |
Publications (2)
Publication Number | Publication Date |
---|---|
KR20170106755A KR20170106755A (en) | 2017-09-22 |
KR101802062B1 true KR101802062B1 (en) | 2017-11-27 |
Family
ID=60034811
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020160030246A KR101802062B1 (en) | 2016-03-14 | 2016-03-14 | Method and system to measure stereoscopic 3d content-induced visual discomfort based facial expression recognition for sensitivity tv |
Country Status (1)
Country | Link |
---|---|
KR (1) | KR101802062B1 (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108545080A (en) * | 2018-03-20 | 2018-09-18 | 北京理工大学 | Driver Fatigue Detection and system |
KR102141740B1 (en) * | 2018-12-06 | 2020-08-05 | 연세대학교 산학협력단 | Apparatus and method for predicting virtual reality sickness |
CN109919131A (en) * | 2019-03-22 | 2019-06-21 | 杭州电子科技大学 | A kind of fatigue detection method based on image processing techniques |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2010117948A (en) * | 2008-11-13 | 2010-05-27 | Canon Inc | Facial expression determination device, control method thereof, imaging device and program |
-
2016
- 2016-03-14 KR KR1020160030246A patent/KR101802062B1/en active IP Right Grant
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2010117948A (en) * | 2008-11-13 | 2010-05-27 | Canon Inc | Facial expression determination device, control method thereof, imaging device and program |
Also Published As
Publication number | Publication date |
---|---|
KR20170106755A (en) | 2017-09-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11227158B2 (en) | Detailed eye shape model for robust biometric applications | |
KR102596897B1 (en) | Method of motion vector and feature vector based fake face detection and apparatus for the same | |
US9235751B2 (en) | Method and apparatus for image detection and correction | |
EP3674852B1 (en) | Method and apparatus with gaze estimation | |
KR102209595B1 (en) | Detailed eye shape model for robust biometric applications | |
EP2923306B1 (en) | Method and apparatus for facial image processing | |
KR102357340B1 (en) | Method and apparatus for face recognition | |
US20140254891A1 (en) | Method and apparatus for registering face images, and apparatus for inducing pose change, and apparatus for recognizing faces | |
US11790494B2 (en) | Facial verification method and apparatus based on three-dimensional (3D) image | |
EP2182469B1 (en) | System and method for sensing facial gesture | |
JP5225870B2 (en) | Emotion analyzer | |
CN114627543A (en) | Method and apparatus for face recognition | |
US20210089760A1 (en) | Facial expression image processing method and apparatus | |
KR101802062B1 (en) | Method and system to measure stereoscopic 3d content-induced visual discomfort based facial expression recognition for sensitivity tv | |
CN105095853B (en) | Image processing apparatus and image processing method | |
JP6822482B2 (en) | Line-of-sight estimation device, line-of-sight estimation method, and program recording medium | |
JP2007163864A (en) | Display control apparatus, display control method, display control program, and display control program recording medium | |
JP6349448B1 (en) | Information processing apparatus, information processing program, and information processing method | |
US9786030B1 (en) | Providing focal length adjustments | |
JP2019040592A (en) | Information processing device, information processing program, and information processing method | |
Fang et al. | Visual attention modeling for stereoscopic video | |
Hersidon | Eye tracking by pattern matching using codebook generation and machine learning | |
KR20230114686A (en) | Electronic device and method for user's hands determination in the video | |
Revathi | Sign Language Recognition Using Principal Component Analysis | |
KR20150068005A (en) | Method for detecting profile line and device for detecting profile line |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
A201 | Request for examination | ||
E902 | Notification of reason for refusal | ||
AMND | Amendment | ||
E601 | Decision to refuse application | ||
AMND | Amendment | ||
X701 | Decision to grant (after re-examination) | ||
GRNT | Written decision to grant |