WO2021210652A1 - Information processing device, system, method, and program - Google Patents

Information processing device, system, method, and program Download PDF

Info

Publication number
WO2021210652A1
WO2021210652A1 PCT/JP2021/015615 JP2021015615W WO2021210652A1 WO 2021210652 A1 WO2021210652 A1 WO 2021210652A1 JP 2021015615 W JP2021015615 W JP 2021015615W WO 2021210652 A1 WO2021210652 A1 WO 2021210652A1
Authority
WO
WIPO (PCT)
Prior art keywords
viewer
video content
degree
eyes
satisfaction
Prior art date
Application number
PCT/JP2021/015615
Other languages
French (fr)
Japanese (ja)
Inventor
壮太郎 五十嵐
Original Assignee
株式会社Theater Guild
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社Theater Guild filed Critical 株式会社Theater Guild
Publication of WO2021210652A1 publication Critical patent/WO2021210652A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N17/00Diagnosis, testing or measuring for television systems or their details
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/258Client or end-user data management, e.g. managing client capabilities, user preferences or demographics, processing of multiple end-users preferences to derive collaborative data

Definitions

  • This disclosure relates to information processing devices, systems, methods and programs.
  • Patent Document 1 discloses a social platform for evaluating a video by so-called throwing money for a video in a video sharing service on the Internet.
  • a payment means for settlement of the amount of money corresponding to the item image is provided, and a money throwing function is implemented.
  • an information processing device including a control unit.
  • the control unit detects the viewer's eyes from the image data of the viewer viewing the video content and acquires the viewer's eye opening degree, and the viewer's eye from the image data. It is provided with a blinking number acquisition unit that detects the number of blinks of the viewer and acquires the number of blinks of the viewer, and a calculation unit that calculates the satisfaction level of the viewer with respect to the video content by using the degree of opening and the number of blinks as parameters.
  • a system including a camera and an information processing device.
  • the camera captures a viewer viewing the video content, generates image data and causes the information processing device to acquire the image data
  • the information processing device includes a control unit, and the control unit captures the viewer viewing the video content.
  • the opening degree acquisition unit that detects the viewer's eyes and acquires the viewer's eye opening degree from the image data, and detects the viewer's eyes from the image data and acquires the number of blinks of the viewer. It is provided with a blinking count acquisition unit and a calculation unit for calculating the viewer's satisfaction with the video content using the degree of opening and the number of blinks as parameters.
  • a method for being executed by a computer equipped with a processor is provided.
  • the method is a step in which the processor detects the viewer's eyes from the image data of the viewer viewing the video content and obtains the degree of opening of the viewer's eyes, and the viewer's eyes are obtained from the image data.
  • a step of detecting and acquiring the number of blinks of the viewer and a step of calculating the satisfaction level of the viewer with respect to the video content are executed by using the degree of opening and the number of blinks as parameters.
  • a program for causing a computer having a processor to execute the program is provided.
  • the program tells the processor a step of detecting the viewer's eyes from the image data of the viewer viewing the video content and acquiring the degree of opening of the viewer's eyes, and the viewer's eyes from the image data.
  • a step of detecting and acquiring the number of blinks of the viewer and a step of calculating the satisfaction level of the viewer with respect to the video content are executed by using the degree of opening and the number of blinks as parameters.
  • the degree of opening of the viewer's eyes and the number of blinks are acquired from the image data of the viewer viewing the video content, and the satisfaction level of the viewer with the video content is calculated based on these. do. Therefore, it is possible to acquire information indicating the reaction of the viewer to the video content, for example, for each scene. This makes it possible to analyze the evaluation of the video content.
  • FIG. 1 It is a figure which shows the whole structure of the satisfaction degree calculation system 1. It is a bird's-eye view which shows the example of the appearance of the facility where the satisfaction calculation system 1 is installed. It is a perspective view which shows the example of the appearance of the voice output device 30. It is a block diagram which shows the functional configuration of the server 10 which comprises the satisfaction calculation system 1 of Embodiment 1.
  • FIG. It is a schematic diagram which shows the specific method of calculating the degree of opening of eyes in the degree of opening degree acquisition module 1034 of FIG. It is a figure which shows the functional configuration of the audio output device 30 which comprises the satisfaction calculation system 1 of Embodiment 1.
  • FIG. It is a figure which shows the data structure of the video content database 181 and the viewer database 184 stored in a server 10.
  • FIG. It is a flowchart which shows an example of the flow which performs the video content output processing by the satisfaction calculation system 1 of Embodiment 1.
  • FIG. It is a flowchart which shows an example of the flow which performs the satisfaction calculation processing by the satisfaction calculation system 1 of Embodiment 1.
  • FIG. It is a figure which shows the screen example of the server 10 which displayed the transition of each parameter in the satisfaction degree calculation process.
  • the satisfaction calculation system provides the viewer with an opportunity to display (screen) a video content such as a movie, and at that time, captures the state in which the viewer is watching the video content and also watches the video content. It is a system that acquires the state of a person with a sensor, analyzes the viewer from the captured image data and the detection signal of the sensor, and calculates the degree of satisfaction.
  • the degree of satisfaction is a numerical value of the degree of satisfaction with the video content, which is estimated from the biological reaction of the viewer who viewed the video content.
  • the degree of satisfaction is a numerical value of the degree of satisfaction with the video content, which is estimated from the biological reaction of the viewer who viewed the video content.
  • the eyes eyelids
  • the number of blinks decreases, and the movement of the face. It is empirically known that (number of times and magnitude of movement) decreases and pulse rate and body temperature increase. Therefore, when the viewer who is viewing the video content shows the above reaction, it is considered that the viewer is concentrating on viewing the video content.
  • the satisfaction calculation system of the present disclosure the state in which the viewer is viewing the video content is photographed with a camera, the face of the viewer is identified from the photographed image data, and the eyes of each viewer are seen. Detect. From the movement of the viewer's eyes in the image data, the degree of eye opening and the number of blinks are acquired. In addition, the movement of the viewer's face is acquired by the detection signal of the sensor that detects the movement of the viewer's face and the biological information, and the biological information of the viewer is acquired. Based on the acquired degree of eye opening, number of blinks, face movement, and biometric information, the degree of satisfaction indicating the degree of satisfaction with the video content is calculated. With such a configuration, it is possible to surely acquire the satisfaction level of the viewer's video content based on the biological reaction of the viewer.
  • an audio output device such as headphones or earphones is attached, and the audio is heard from the audio output device.
  • the above-mentioned sensor is arranged in the audio output device, and the viewer can detect the movement of the viewer's face and the biological information of the viewer by wearing the audio output device. With this configuration, it is possible to reliably acquire the movement of the viewer's face and the biological information of the viewer.
  • each viewer who is viewing the video content is identified, and the satisfaction is calculated for each viewer. This makes it possible to acquire satisfaction for each attribute of the viewer and analyze the evaluation of the video content.
  • FIG. 1 is a diagram showing the overall configuration of the satisfaction calculation system 1.
  • the satisfaction calculation system is a system for displaying (screening) video content such as a movie, providing the viewer with an opportunity to view the content, analyzing the viewer at that time, and calculating the satisfaction.
  • the satisfaction calculation system 1 shows a server 10, a video output device 20, and a plurality of audio output devices (in FIG. 1, the audio output device 30A and the audio output device 30B are shown below. Collectively, it may be referred to as "audio output device 30") and the camera 40.
  • the server 10, the video output device 20, the audio output device 30, and the camera 40 are connected to each other so as to be able to communicate with each other via the network 80.
  • the network 80 is composed of a wired or wireless network.
  • the server 10 is a device that manages video content to be viewed by the viewer, image data obtained by capturing the viewer, a viewer detection signal by a sensor, and information on satisfaction with the video content for each viewer.
  • the server 10 transmits the video data of the video content to the video output device 20, and transmits the audio data to the audio output device 30.
  • the server 10 acquires the image data of the viewer and the detection signal of the sensor, and calculates the satisfaction level of the viewer.
  • the server 10 is communicably connected to the video output device 20, the audio output device 30, and the camera 40 via the network 80.
  • the server 10 supports wireless base stations 81 that support communication standards such as 4G, 5G, and LTE (Long Term Evolution), and wireless LAN (Local Area Network) standards such as IEEE (Institute of Electrical and Electronics Engineers) 802.11. It is connected to the network 80 by communicating with a communication device such as the wireless LAN router 82.
  • each device may be connected wirelessly or by wire, and when connecting wirelessly, the communication protocols include, for example, Z-Wave (registered trademark), ZigBee (registered trademark), and Bluetooth (registered).
  • a trademark) or the like may be used, and when connecting by wire, a USB (Universal Serial Bus) cable or the like may be used for connection.
  • the server 10 is a computer connected to the network 80.
  • the server 10 includes a communication IF (Interface) 12, an input / output IF 13, a memory 15, a storage 16, and a processor 19.
  • IF Interface
  • the communication IF 12 is an interface for inputting / outputting signals because the server 10 communicates with an external device.
  • the input / output IF 13 is an interface with an input device (for example, a pointing device such as a mouse, a keyboard) for receiving an input operation from a user, and an output device (display, a speaker, etc.) for presenting information to the user.
  • the memory 15 is for temporarily storing a program, data processed by the program or the like, and is a volatile memory such as a DRAM (Dynamic Random Access Memory).
  • the storage 16 is a storage device for storing data, and is, for example, a flash memory, an HDD (Hard Disc Drive), or an SSD (Solid State Drive).
  • the processor 19 is hardware for executing an instruction set described in a program, and is composed of an arithmetic unit, registers, peripheral circuits, and the like.
  • the video output device 20 is a device that receives video data of video content and displays (screens) it as a video according to an instruction from the server 10.
  • an LED panel in which LED elements are arranged on the edge of a frame or an organic device. It is composed of an EL display or a liquid crystal display.
  • the video output device 20 is provided in a screening facility such as a movie theater.
  • the video output device 20 may be configured such that a plurality of LED panels are arranged in parallel to form a large-scale vision of about several hundred inches so that a large number of viewers can view the video at a time.
  • the video output device 20 is not limited to one, and may be configured by a plurality of video output devices 20, so that a facility provided with the plurality of video output devices 20 can be remotely controlled by one server 10, for example. It may be configured.
  • the audio output device 30 is a device that outputs audio data of video content as audio in response to an instruction from the server 10, and the audio output from the audio output device 30 synchronizes with the video displayed on the video output device 20. It is configured as follows. In addition, when a plurality of video output devices 20 are provided as described above and different video contents are screened, the sound corresponding to the screened video content is configured to be output, and the plurality of sounds are output. It may be configured to be selectable from the output channel.
  • the audio output device 30 is mounted so as to cover the ears of the viewer who views the video content, and is realized by wireless headphones or the like that receive the audio of the video content by wireless communication and output it as audio under the control of the server 10. ..
  • the audio output device 30 is not limited to, for example, headphones connected by wire or those that cover the ears of the viewer, and may be earphones connected wirelessly or by wire.
  • the voice output device 30 includes a communication IF 32, a sensor unit 33, a speaker 34, a memory 35, a storage unit 36, and a processor 39.
  • the communication IF 32 is an interface for inputting / outputting signals because the voice output device 30 communicates with an external device.
  • the sensor unit 33 is a device that acquires the facial movement and biological information of the viewer who is viewing the video content. For example, a gyro sensor that detects the facial movement of the viewer as an angular velocity and a pulse or body temperature are detected. It is composed of an electrocardiographic sensor such as a temperature sensor and an optical sensor that detect it.
  • the sensor unit 33 may be configured by MEMS (Micro Electro Mechanical Systems) in which a sensor and other parts are integrated. That is, the sensor unit 33 has functions such as vibration detection, a thermometer, an electroencephalograph, and a heart rate monitor.
  • the speaker 34 is an output device for converting a voice signal into voice and outputting information as voice to a user.
  • the memory 35 is for temporarily storing a program, data processed by the program or the like, and is a volatile memory such as a DRAM (Dynamic Random Access Memory).
  • the storage unit 36 is a storage device for storing data, and is, for example, a flash memory, an HDD (Hard Disc Drive), or an SSD (Solid State Drive).
  • the processor 39 is hardware for executing an instruction set described in a program, and is composed of an arithmetic unit, registers, peripheral circuits, and the like.
  • the camera 40 is a device that captures a viewer who is viewing video content according to an instruction from the server 10.
  • an image pickup device such as a CCD (Charge Coupled Device) image sensor and the captured image are captured from a digital signal. It is composed of a digital camera or the like equipped with an A / D conversion device for converting the image data.
  • the camera 40 may be configured to change the shooting direction under the control of the server 10, or may be configured to appropriately move the shooting direction in order to enable shooting in a wide range.
  • the camera 40 shoots a viewer viewing the video content as a moving image, or shoots as a still image at predetermined time intervals, and converts the moving image or the still image into image data. Further, the camera 40 transmits the captured image data to the server 10.
  • the number of cameras 40 is not limited to one, and a plurality of cameras 40 may be provided in one facility. Further, the camera 40 is not limited to a single device, and may be configured by, for example, a camera function provided in a mobile terminal such as a smartphone or tablet compatible with a mobile communication system.
  • FIG. 2 is a bird's-eye view showing an example of the appearance of the facility in which the satisfaction calculation system 1 is installed.
  • the satisfaction calculation system 1 shown in FIG. 2 shows an example in which the satisfaction calculation system 1 is installed in a space for showing video contents, such as a movie theater.
  • the satisfaction calculation system 1 is configured so that the viewer wears an audio output device 30 such as headphones to acquire audio, it is possible to eat, drink, and perform other work other than viewing the video content. It is a facility that constitutes a multipurpose space with a so-called cafe space. Further, the satisfaction calculation system 1 may be configured at, for example, the viewer's home.
  • the server 10 is composed of a cloud server or the like, and the server 10 outputs video to a video output device 20 composed of a television or a display used by the viewer by a service such as on-demand, and headphones used by the viewer.
  • the audio output device 30 configured by the above may output the audio, the viewer may be photographed by the camera 40 configured by the smartphone or the like used by the viewer, and the image data may be transmitted to the server 10.
  • the satisfaction calculation system 1 shown in FIG. 2 is provided on the floor F in which the screening space for screening the video content and the cafe space are provided, and the video is displayed at a predetermined location (upper right direction shown in FIG. 2) on the floor F.
  • An output device 20 and a camera 40 are installed.
  • the viewer W wears the audio output device 30 when viewing the video content as described above.
  • the viewer W can concentrate on viewing the video content without worrying about the surrounding situation, and can enjoy the video content like a user of the cafe space. Those who are not watching can eat, drink, and perform other actions without worrying about the audio of the video content.
  • the video output device 20 is composed of an LED panel or the like, it irradiates the viewer W side with light having a constant brightness. Therefore, the camera 40 can shoot the viewer W at a constant resolution by the light.
  • the facility in which the satisfaction calculation system 1 is installed is not limited to a closed space such as a movie theater, but is a space of a commercial facility such as an eat-in corner of a convenience store or an event space. You may. Further, the satisfaction calculation system 1 may be installed in a space where a certain level of quietness is ensured, such as in a restaurant, a hotel lounge, or a shared space of an apartment house such as an apartment. Further, the satisfaction calculation system 1 may be installed in a game facility such as a pachinko or pachislot store, or in a facility with a certain amount of noise such as a game center or an amusement facility.
  • FIG. 3 is a perspective view showing an example of the appearance of the audio output device 30.
  • the audio output device 30 is composed of headphones worn so that both ears are sandwiched from the head in order for the viewer viewing the video content to hear the audio, and the pair of ears accommodating each covering both ears. It is composed of a portion 37 and an arm portion 38 that supports the pair of ear accommodating portions 37 and attaches the ear accommodating portion 37 to the viewer's ear by pressing the viewer's head with an appropriate pressure.
  • the audio output device 30 is not limited to such a configuration, and may be configured such that the ear accommodating portion 37 is mounted so as to be locked to the earlobe of the viewer without providing the arm portion 38. It may be a canal type earphone that is inserted into the ear hole of the viewer.
  • a sensor unit 33 is provided at a position inside the pair of ear accommodating units 37 in contact with the skin around the viewer's ear.
  • the position in the ear accommodating unit 37 is not limited to a specific position, and a plurality of positions may be provided in the ear accommodating unit 37. Further, depending on the type of biological information acquired by the sensor unit 33, it may be arranged so as to be in direct contact with the skin of the viewer, or may be arranged so as to be separated from the skin of the viewer. Since the audio output device 30 has such a configuration, it is desirable to use the audio output device 30 without attaching the ear pads, but it does not prevent the ear pads from being attached.
  • FIG. 4 is a diagram showing a functional configuration of the server 10 constituting the satisfaction calculation system 1 of the first embodiment. As shown in FIG. 4, the server 10 functions as a communication unit 101, a storage unit 102, and a control unit 103.
  • the communication unit 101 performs processing for the server 10 to communicate with an external device.
  • the storage unit 102 stores data and programs used by the server 10.
  • the storage unit 102 stores the video content database 181, the sensor detection signal database 182, the camera-captured image database 183, the viewer database 184, and the like.
  • the video content database 181 is a database for holding information on the video content and its screening time provided by the satisfaction calculation system 1, that is, the screening schedule of the video output device 20. Details will be described later.
  • the sensor detection signal database 182 is a database for holding the detection signals from the sensor unit 33 of the voice output device 30 in chronological order, which is worn by the viewer who uses the satisfaction calculation system 1. Since the detection signal of the sensor unit 33 is transmitted from the voice output device 30 to the server 10 with the date and time information added, the detection signal is stored.
  • the camera-captured image database 183 is a database for holding the image data of the camera 40 in chronological order, which captures the viewers who use the satisfaction calculation system 1. Since the image data of the camera 40 is transmitted from the camera 40 to the server 10 with date and time information added, the detection signal is stored.
  • the viewer database 184 is a database for holding information on viewers who use the satisfaction calculation system 1 and information on the satisfaction of the video content viewed for each viewer. Details will be described later.
  • the viewer database 184 may store member information (including attribute information) registered in the facility.
  • the control unit 103 performs processing according to a program by the processor of the server 10, and as various modules, the reception control module 1031, the transmission control module 1032, the video content output control module 1033, the opening degree acquisition module (opening degree acquisition unit) 1034, The functions shown in the blink count acquisition module (blink frequency acquisition unit) 1035, face movement acquisition module (movement acquisition unit) 1036, biometric information acquisition module (biological information acquisition unit) 1037, and satisfaction calculation module (calculation unit) 1038 are exhibited. do.
  • the reception control module 1031 controls the process in which the server 10 receives a signal from an external device according to the communication protocol.
  • the transmission control module 1032 controls a process in which the server 10 transmits a signal to an external device according to a communication protocol.
  • the video content output control module 1033 controls the process of transmitting the video data of the video content to the video output device 20 and transmitting the audio data to the audio output device 30 in order to allow the viewer to view the video content.
  • the video content output control module 1033 refers to the video content database 181 and controls the video content to be screened while transmitting the image data and the audio data of the video content based on the schedule information indicating the date and time when the video content is screened. It transmits signals and also transmits information such as the data capacity, data format, and resolution of video content.
  • the video content output control module 1033 may collectively transmit and screen the image data and the audio data to the video output device 20 and the audio output device 30, and sequentially transmit and screen the image data and the audio data in a streaming format. May be good.
  • the video content output control module 1033 may transmit credit information such as the screening time of the video content and the name or name of the creator.
  • the opening degree acquisition module 1034 controls a process of acquiring image data taken by the camera 40, analyzing the image data, recognizing and detecting the viewer's eyes, and acquiring the viewer's eye opening degree.
  • the opening degree acquisition module 1034 uses an image recognition technique such as machine learning, which is a known technique, as a method for identifying the eyes of the viewer.
  • the opening degree acquisition module 1034 analyzes the image data to detect the viewer's eyes, calculates the area of the portion of the image data in which the viewer's eyes are opened, and acquires this area as the degree of opening of the eyes. .. Further, the opening degree acquisition module 1034 analyzes the image data to detect the viewer's eyes, and the contact point between the upper eyelid and the lower eyelid (the left and right end points of the eyes) in the portion of the image data where the viewer's eyes are opened. ) And the upper end point of the upper eyelid, or the contact point between the iris (so-called black eye) and the upper eyelid, and the lower end point of the lower eyelid, or the position information of the contact point between the iris and the lower eyelid. Based on the information, it is acquired as the degree of eyelid opening of the viewer.
  • FIG. 5 is a schematic diagram showing a specific method for calculating the degree of opening of the eyes in the degree of opening degree acquisition module 1034 of FIG. The method for calculating the degree of eye opening will be described with reference to FIG.
  • FIG. 5 shows eyes E schematically showing the eyes of the viewer, and the eyes E show the upper eyelid L1, the lower eyelid L2, and the iris L3.
  • FIG. 5A is an example in which the iris L3 is in contact with the upper eyelid L1 and the lower eyelid L2
  • FIG. 5B is an example in which the iris L3 is in contact with the upper eyelid L1 and the lower eyelid L2. This is an example when there is no such thing.
  • the contact points P1 and P2 between the upper eyelid L1 and the lower eyelid L2 and the iris are analyzed by analyzing the image data. It is possible to recognize the positions of the contacts P3 and P4 between L3 and the upper eyelid L1 and the contacts P5 and P6 between the iris L3 and the lower eyelid L2. Therefore, the coordinates of the contacts P1 to P6 in the image data of the viewer's eyes E are calculated, and the degree of opening of the viewer's eyes is acquired based on the position information of the coordinates.
  • the contact points P1 and P2 between the upper eyelid L1 and the lower eyelid L2 are obtained by analyzing the image data. It is possible to recognize the positions of the upper end point P7 of the upper eyelid L1 and the lower end point P8 of the lower eyelid L2. Therefore, the coordinates of the contacts P1, P2 and the end points P7 and P8 in the image data of the viewer's eyes E are calculated, and the degree of opening of the viewer's eyes is acquired based on the position information of the coordinates.
  • the contact points P3 and P4 are recognized for the upper eyelid L1 by the method shown in FIG. 5A.
  • the end point P8 may be recognized by the method shown in FIG. 5 (b), and vice versa.
  • the blink count acquisition module 1035 controls a process of acquiring image data taken by the camera 40, analyzing the image data, recognizing and detecting the viewer's eyes, and acquiring the number of blinks in the viewer's eyes. .. Similar to the opening degree acquisition module 1034, the blink count acquisition module 1035 uses an image recognition technique such as machine learning, which is a known technique, as a method for identifying the eyes of the viewer. In addition, in order to acquire the blink of a person's eyes as image data, a frame rate of 50 fps (frame per second) or more is generally required. Therefore, the camera 40 shoots at a frame rate of 50 fps or more. It is said.
  • the blink count acquisition module 1035 analyzes the image data to detect the viewer's eyes and acquires the number of blinks of the viewer in a predetermined time (for example, 1 minute). Further, in the blink count acquisition module 1035, when the viewer's eyes are closed and the viewer's eyes are opened within a predetermined time (for example, 1 second) from that state, the viewer Recognizes that you have blinked and counts it as the number of blinks. In such a configuration, if the viewer's eyes are closed and the eyes are not opened for a certain period of time, it cannot be said that the viewer is watching the video content during that period, and it is considered that the viewer is in a sleeping state, for example. Therefore, it is set to distinguish it from the sleeping state.
  • a predetermined time for example, 1 minute
  • the face movement acquisition module 1036 acquires a detection signal detected and output by the sensor unit 33 from the audio output device 30 worn when the viewer is viewing the video content, and analyzes the detection signal. Controls the process of acquiring the movement of the viewer's face.
  • the face movement acquisition module 1036 acquires the movement of the viewer's face from the angular velocity detected by the gyro sensor constituting the sensor unit 33 and the time of the movement.
  • the viewer tilts the face back and forth and left and right, or the viewer moves the direction of the face based on the angular velocity detected by the gyro sensor constituting the sensor unit 33 and the time of the movement. Acquires the movement (movement rotated in the left-right direction).
  • the movement in which the viewer tilts his / her face may be a movement by moving the neck or a movement of the entire upper body by moving the trunk portion.
  • a specific example of the movement of the viewer's face to be acquired may be the number of times the viewer moves the face beyond the predetermined movement range within a predetermined time (for example, 10 minutes), and the angle or distance of movement of the face may be obtained. But it may be.
  • the biological information acquisition module 1037 acquires a detection signal detected and output by the sensor unit 33 from the audio output device 30 worn when the viewer is viewing the video content, and analyzes the detection signal. Controls the process of acquiring the viewer's biometric information.
  • the biological information acquisition module 1037 acquires the viewer's pulse, blood pressure, and body temperature detected by the electrocardiographic sensor constituting the sensor unit 33.
  • the satisfaction calculation module 1038 includes the degree of eye opening acquired by the degree of opening acquisition module 1034, the number of blinks acquired by the number of blinks acquisition module 1035, the movement of the face acquired by the face movement acquisition module 1036, and the biometric information acquisition module 1037. Controls the process of calculating the viewer's satisfaction with the video content using the biometric information acquired by The satisfaction calculation module 1038 may calculate the satisfaction only from the degree of eye opening and the number of blinks as parameters, may add facial movement as a parameter, or may add biometric information as a parameter. ..
  • Satisfaction is a time-series quantification of the degree of satisfaction with the video content, which is estimated from the biological reaction of the viewer who viewed the video content as described above, at predetermined time intervals (for example, 1 minute). Yes, it is calculated as the satisfaction level of the viewer for each viewer, or the satisfaction level of a plurality of viewers is calculated as a total value or an average value and used as an evaluation index of the video content. Further, this satisfaction level may be acquired as a time-series transition at predetermined time intervals.
  • the transition of individual satisfaction of viewers it is possible to group viewers who have similar transitions of satisfaction with a certain video content, or to analyze the attributes of viewers for each group. It may be used as an index value for marketing video content.
  • the video content can be selected. You may recommend it.
  • the satisfaction calculation module 1038 sets, for example, the eyes, face, and physical condition of the viewer first acquired as initial values (reference values), and calculates the satisfaction from the subsequent changes in the respective values or the rate of change. Satisfaction may be calculated from the change in each value within a predetermined time or the rate of change, and the difference from the eyes, face, and physical condition (for example, average value) of a general viewer is satisfied. The degree may be calculated.
  • the satisfaction calculation module 1038 performs predetermined calculations with the parameters of the degree of eye opening, the number of blinks, the movement of the face, and the value of biometric information quantified by the above example, and after the calculation, each of them performs a predetermined calculation. Satisfaction is calculated by adding the values of.
  • the function J for calculating the degree of opening of the eyes is a function of the acquired degree of opening value j.
  • the function K for calculating the number of blinks be a function of the value k of the acquired number of blinks.
  • the function L for calculating the movement of the face be a function of the acquired value l of the movement of the face.
  • the function M for calculating the biological information be the value m of the acquired biological information.
  • each function may use each value as it is.
  • the satisfaction calculation module 1038 multiplies the degree of eye opening, the number of blinks, the movement of the face, and the function of biometric information by different predetermined coefficients as weights according to the above example, and after multiplication. Satisfaction is calculated by adding each value of.
  • a, b, c, and d are predetermined coefficients, respectively.
  • the degree of satisfaction may be calculated by using the coefficients a and d as positive coefficient and the coefficients b and c as negative coefficients. Further, it is considered that the value and the rate of change of each of the above values differ depending on the genre of the video content being viewed. Therefore, the weighting of the coefficients a, b, c, and d may be changed for each genre of the video content. Further, it is considered that a time lag may occur until the viewer reacts to each of the above values depending on the genre of the video content being viewed and the emotion of the viewer for each scene. Therefore, each value may be acquired by correcting the time lag.
  • the satisfaction calculation module 1038 adds the functions of the degree of opening of the eyes and the number of blinks according to the above example to obtain the degree of concentration, and obtains the degree of satisfaction by performing a predetermined calculation on the degree of concentration. It may be calculated.
  • R may be a function or value obtained by performing a predetermined operation on the function of facial movement and biometric information, or may be a function or value calculated using other values as parameters.
  • FIG. 6 is a diagram showing a functional configuration of the audio output device 30 constituting the satisfaction calculation system 1 of the first embodiment.
  • the voice output device 30 exerts functions as a communication unit 301, a detection unit 302, a storage unit 303, and a control unit 304.
  • the communication unit 301 performs processing for the voice output device 30 to communicate with an external device.
  • the detection unit 302 performs processing for the voice output device 30 to acquire the movement of the viewer's face and biometric information. It is composed of the above-mentioned gyro sensor and electrocardiographic sensor.
  • the storage unit 303 stores the data and the program used by the voice output device 30.
  • the storage unit 303 stores the gyro sensor detection signal database 381, the biosensor detection signal database 382, and the like.
  • the gyro sensor detection signal database 381 is a database for holding the detection signals of the gyro sensors constituting the detection unit 302 in chronological order.
  • the biosensor detection signal database 382 is a database for holding the detection signals of the electrocardiographic sensors constituting the detection unit 302 in chronological order.
  • the control unit 304 performs processing according to a program by the processor of the voice output device 30, and causes the reception control module 3031, the transmission control module 3032, the voice output module 3033, the gyro sensor control module 3034, and the biosensor control module 3035 as various modules. Demonstrate the function shown.
  • the reception control module 3031 controls a process in which the voice output device 30 receives a signal from an external device according to a communication protocol.
  • the transmission control module 3032 controls a process in which the voice output device 30 transmits a signal to an external device according to a communication protocol.
  • the audio output module 3033 controls a process of outputting audio data transmitted from the server 10 as audio of video content from the speaker 34.
  • the gyro sensor control module 3034 controls the operation processing of the gyro sensor constituting the detection unit 302. Further, the gyro sensor control module 3034 transmits a detection signal to the server 10.
  • the biosensor control module 3035 controls the operation processing of the electrocardiographic sensor constituting the detection unit 302. Further, the biosensor control module 3035 transmits a detection signal to the server 10.
  • FIG. 7 is a diagram showing the data structures of the video content database 181 and the viewer database 184 stored in the server 10.
  • each record in the video content database 181 includes an item "facility ID”, an item “facility name”, an item “content screening information”, and the like.
  • the item "facility ID” is information that identifies each facility in which the satisfaction calculation system 1 is installed. When a plurality of video output devices 20 are installed in one facility and different video contents are screened, the item “facility ID” having a different value is assigned and managed separately.
  • the item "facility name” is information indicating the name of the facility in which the satisfaction calculation system 1 is installed.
  • the item “content screening information” is information on the screening schedule of video content such as a movie to be screened at a facility using the satisfaction calculation system 1. Specifically, the item “start date and time” and the item “end time”. , And the item “content name” and the like. As shown in FIG. 7, one or a plurality of items “content screening information” are stored in chronological order for one item “facility ID”.
  • start date and time is information indicating the start date and start time of the video content to be screened at the facility.
  • the item "end time” is information indicating the end time of the video content to be screened at the facility.
  • the item "content name” is information indicating the name of the video content to be screened at the facility.
  • the server 10 updates the video content database 181 in response to receiving input of video content screening information from the facility manager.
  • Each of the records in the viewer database 184 includes a "viewer ID”, an item “viewer name”, an item “viewing content”, an item “satisfaction information”, and the like.
  • the item "viewer ID” is information that identifies each viewer who has viewed the video content screened at the facility that uses the satisfaction calculation system 1.
  • the item "viewer name” is information indicating the name of the viewer who has viewed the video content screened at the facility using the satisfaction calculation system 1. If the viewer's name information cannot be obtained at the facility, the item “viewer ID” shown in FIG. 7 does not have to be stored as in the record shown by "# 0202".
  • the item "viewing content” is information indicating the name of the video content viewed by the viewer, and corresponds to the item "content name” of the video content database 281.
  • the item "satisfaction information” is information on the satisfaction of the viewer who has viewed the video content screened at the facility using the satisfaction calculation system 1, and specifically, the item “elapsed time” and the item “elapsed time”. Includes “satisfaction” and the like. As shown in FIG. 7, one or a plurality of items “satisfaction information” are stored in a time series for each predetermined time (for example, one minute) for one item “viewer ID”.
  • the item "elapsed time” is information indicating the elapsed time for each predetermined time from the start of the video content.
  • the item "satisfaction” is information indicating the satisfaction of the viewer who has viewed the video content at the time of the item "elapsed time”, and is the value of the satisfaction calculated by the satisfaction calculation module 1038.
  • the satisfaction calculation module 1038 of the server 10 updates the viewer database 184 every time the satisfaction of the viewer who has viewed the video content is calculated.
  • FIG. 8 is a flowchart showing an example of a flow of performing video content output processing by the satisfaction calculation system 1 of the first embodiment.
  • step S111 the video content output control module 1033 of the server 10 refers to the video content database 181 and manages the date and time when the video content is screened.
  • the video content output control module 1033 proceeds to the process of step S112 when it is time to transmit the video content in order to screen it.
  • step S112 the video content output control module 1033 of the server 10 transmits the video data of the video content to the video output device 20 and the audio data to the audio output device 30 in order to allow the viewer to view the video content.
  • step S122 the video output device 20 receives the video data of the video content transmitted by the server 10.
  • the audio output device 30 receives audio data of video content transmitted by the server 10.
  • step S123 the video output device 20 screens the video data of the received video content as a video.
  • the audio output device 30 outputs the audio data of the received video content as audio.
  • the satisfaction calculation system 1 transmits the video data of the video content to the video output device 20 and the audio data to the audio output device 30 at a predetermined time.
  • the video output device 20 screens video data as video, and the audio output device 30 outputs audio data as audio.
  • the video content is screened, and the viewer wears the audio output device 30 for viewing.
  • FIG. 9 is a flowchart showing an example of a flow of performing satisfaction calculation processing by the satisfaction calculation system 1 of the first embodiment.
  • step S211 the gyro sensor control module 3034 of the voice output device 30 controls the operation of the gyro sensors constituting the detection unit 302, and transmits the detection signal of the gyro sensor to the server 10 via the communication unit 301. Further, the biosensor control module 3035 controls the operation of the biosensors constituting the detection unit 302, and transmits the detection signal of the biosensor to the server 10 via the communication unit 301.
  • step S231 the camera 40 transmits the captured image data to the server 10.
  • step S221 the server 10 receives the detection signal transmitted from the voice output device 30 via the communication unit 101. Further, the server 10 receives the image data transmitted from the camera 40 via the communication unit 101. The server 10 stores the received detection signal in the sensor detection signal database 182 and the received image data in the camera-captured image database 183.
  • step S222 the open degree acquisition module 1034 of the server 10 acquires the image data stored in the camera-captured image database 183.
  • the opening degree acquisition module 1034 analyzes the image data, recognizes and detects the eyes of the viewer, and acquires the degree of opening of the eyes in the eyes of the viewer.
  • step S223 the blink count acquisition module 1035 of the server 10 acquires the image data stored in the camera-captured image database 183.
  • the blink count acquisition module 1035 analyzes the image data, recognizes and detects the viewer's eyes, and acquires the number of blinks of the viewer.
  • step S224 the face movement acquisition module 1036 of the server 10 acquires the detection signal stored in the sensor detection signal database 182.
  • the face movement acquisition module 1036 analyzes the detection signal of the gyro sensor and acquires the movement of the viewer's face from the angular velocity of the detection signal and the time of the movement.
  • the biometric information acquisition module 1037 of the server 10 acquires the detection signal stored in the sensor detection signal database 182.
  • the biological information acquisition module 1037 analyzes the detection signal of the electrocardiographic sensor and acquires the biological information of the viewer (the pulse and body temperature of the viewer).
  • step S226 the satisfaction calculation module 1038 of the server 10 is based on the degree of eye opening acquired in step S222, the number of blinks acquired in step S223, the facial movement acquired in step S224, and the biometric information acquired in step S225. Based on this, the degree of satisfaction of the viewer with the video content is calculated.
  • step S227 the satisfaction calculation module 1038 of the server 10 stores the calculated satisfaction in the viewer database 184 and updates the database.
  • FIG. 10 is a diagram showing a screen example of the server 10 displaying the transition of each parameter in the satisfaction calculation process.
  • the screen example of FIG. 10 shows a screen example in which the degree of eye opening, the number of blinks, the movement of the face, and the time-series transition of biological information (pulse) are displayed on the input / output IF13 (display) of the server 10. It corresponds to the calculation result in step S226 of FIG.
  • a graph 1031a showing the degree of opening of the eyes of the viewer, the number of blinks, the movement of the face, and the time-series transition of biological information (pulse). Is displayed as an example.
  • the degree of eye opening, the number of blinks, the movement of the face, and the pulse at the start of the video content are set to 100, and the rate of change of each parameter with the subsequent passage is calculated and graphed.
  • the screening time of the video content is 2 hours.
  • the degree of eye opening and the pulse continue to decrease after the start of the video content, reverse to increase at 30 minutes after the start, and decrease after peaking at 90 minutes after the start. Subsequently, the value is around 100 at the end of the video content. In addition, the number of blinks and the movement of the face continue to increase after the start of the video content, peak at 30 minutes after the start and reverse to decrease, and then reverse to increase at 90 minutes after the start of the video content. At the end, the value is around 100.
  • the coefficient for calculating the degree of satisfaction is set so that the degree of eye opening and the degree of satisfaction are high when the pulse is large (positive coefficient), and when the number of blinks and the movement of the face are small. Set so that the satisfaction level is high (negative coefficient).
  • the image data obtained by capturing the viewer is analyzed, and the degree of eye opening and the number of blinks are acquired. Face movement and biometric information are acquired from the detection signal detected by the sensor. Based on this information, the degree of satisfaction with the video content is calculated. Therefore, it is possible to acquire information indicating the reaction of the viewer to the video content. This makes it possible to analyze the evaluation of the video content.
  • the degree of eye opening, the number of blinks, the movement of the face, and the biological information are acquired at predetermined time intervals, and the satisfaction level is calculated at predetermined time intervals.
  • the satisfaction level is calculated at predetermined time intervals.
  • An information processing device including a control unit, in which the control unit detects the viewer's eyes from image data obtained by capturing a viewer viewing the content and acquires the degree of opening of the viewer's eyes.
  • An information processing device including a calculation unit for calculating satisfaction.
  • the control unit further includes a motion acquisition unit that acquires the movement of the viewer's face from image data or from a detection signal output from a sensor provided in the audio output device worn by the viewer.
  • the information processing device according to (Appendix 1), wherein the calculation unit calculates the degree of satisfaction of the viewer with the video content using the movement of the face as a parameter.
  • the motion acquisition unit acquires the motion of the viewer's face from the detection signal output from the sensor provided in the audio output device worn by the viewer on his / her ear, according to (Appendix 2). Information processing device.
  • the motion acquisition unit acquires a motion in which the viewer's face is tilted or a motion in which the viewer's face is moved from the detection signal output from the gyro sensor provided in the voice output device. , (Appendix 2) or (Appendix 3).
  • the control unit further includes a biometric information acquisition unit that acquires the biometric information of the viewer from the detection signal output from the sensor provided in the audio output device worn by the viewer, and the calculation unit includes a biometric information acquisition unit.
  • the information processing device according to any one of (Appendix 1) to (Appendix 4), which calculates the degree of satisfaction of the viewer with the video content using biological information as a parameter.
  • the biometric information acquisition unit acquires the viewer's pulse or blood pressure information as biometric information from the detection signal output from the pulse sensor provided in the audio output device worn by the viewer (Appendix 5). ).
  • the biometric information acquisition unit acquires the viewer's body temperature information as biometric information from the detection signal output from the temperature sensor provided in the audio output device worn by the viewer (Appendix 5) or.
  • the information processing apparatus according to (Appendix 6).
  • the calculation unit determines whether the difference between each value of the parameter and the reference value at a predetermined time, the rate of change from the reference value, the value of change within a predetermined time, or the rate of change within a predetermined time.
  • the information processing apparatus according to any one of (Appendix 1) to (Appendix 7), which calculates satisfaction based on one or more of them.
  • the calculation unit performs an operation by a predetermined function for each value of the parameter, and adds each value after the operation to calculate the satisfaction level, from (Appendix 1) to (Appendix 8).
  • the information processing device according to any one.
  • the calculation unit calculates satisfaction by multiplying each value after the parameter or calculation by a predetermined coefficient and adding each value after the multiplication.
  • Information processing equipment
  • the calculation unit multiplies each predetermined value after the parameter or calculation by a coefficient of a positive value, and the coefficient of a negative value with respect to each predetermined value after the parameter or calculation.
  • (Appendix 12) The information processing device according to any one of (Appendix 1) to (Appendix 11), wherein the calculation unit identifies one or a plurality of viewers, and calculates satisfaction for each viewer.
  • the opening degree acquisition unit calculates the area of the portion where the viewer's eyes are opened and acquires it as the viewer's eye opening degree, according to any one of (Appendix 1) to (Appendix 12). Information processing device.
  • the degree of opening acquisition unit is the contact point between the upper eyelid and the lower eyelid and the upper end point of the upper eyelid, or the contact point between the iris and the upper eyelid, and the lower eyelid in the portion where the viewer's eyes are opened. Any of (Appendix 1) to (Appendix 12) that calculates the position information of the lower end point or the contact point between the iris and the lower eyelid and acquires it as the degree of opening of the viewer's eyes based on the position information. Information processing device described in the eyelids.
  • (Appendix 15) The information processing device according to any one of (Appendix 1) to (Appendix 14), wherein the blink count acquisition unit acquires the number of blinks of the viewer in a predetermined time of the image data obtained by photographing the viewer.
  • the camera photographs a viewer who views video content, generates image data, and causes the information processing device to acquire the image data.
  • the control unit includes an opening degree acquisition unit that detects the eyes of the viewer from the image data of the viewer viewing the video content and acquires the opening degree of the viewer's eyes, and image data. From, a blinking count acquisition unit that detects the viewer's eyes and acquires the number of blinks of the viewer, and a calculation unit that calculates the satisfaction level of the viewer with the video content using the degree of opening and the number of blinks as parameters. , A system.
  • Appendix 18 A method for being executed by a computer equipped with a processor, wherein the processor detects the eyes of the viewer from the image data of the viewer who views the video content, and the viewer's eyes are detected.
  • 10 server 20 video output device, 30 audio output device, 40 camera, 80 network, 12 communication IF, 13 input / output IF, 15 memory, 16 storage, 19 processor, 33 sensor unit, 101 communication unit, 102 storage unit, 181 Video content database, 182 sensor detection signal database, 183 camera shot image database, 184 viewer database, 103 control unit, 1033 video content output control module, 1034 opening degree acquisition module (opening degree acquisition unit), 1035 blink count acquisition module ( Blink count acquisition unit), 1036 Face movement acquisition module (movement acquisition unit), 1037 Biometric information acquisition module (biological information acquisition unit), 1038 Satisfaction calculation module (calculation unit)

Abstract

A server 10 of a satisfaction level calculation system 1 is provided with: an opening degree acquisition module 1034 capable of acquiring and analyzing image data captured by a camera, and acquiring an opening degree of an eye of a viewer; a number-of-blinks acquisition module 1035 for acquiring the number of blinks of the eye of the viewer by analyzing the image data; and a satisfaction level calculation module 1038 which, using the opening degree of the eye and the number of blinks as parameters, calculates a satisfaction level of the viewer with respect to video content.

Description

情報処理装置、システム、方法及びプログラムInformation processing equipment, systems, methods and programs
 本開示は、情報処理装置、システム、方法及びプログラムに関する。 This disclosure relates to information processing devices, systems, methods and programs.
 近年、映像コンテンツを制作するためのコストを確実に回収するため、または映像コンテンツの趣向を分析するためのマーケティングを目的として、映像コンテンツを視聴する機会を提供し、当該映像コンテンツに対する評価情報を取得する手法が検討されている。映像コンテンツ自体がデジタル化により容易に編集することが可能になったこともあり、例えばA/Bテストのようなマーケティング手法を用いて、映像コンテンツに対する評価を分析することが考えられている。 In recent years, in order to reliably recover the cost of producing video content, or for marketing purposes to analyze the taste of video content, we have provided an opportunity to view video content and obtained evaluation information for the video content. The method of doing this is being studied. Since the video content itself can be easily edited by digitization, it is considered to analyze the evaluation of the video content by using a marketing method such as an A / B test.
 インターネット上の動画に対する評価手法については、いくつかの実施例がある。特許文献1には、インターネット上の動画共有サービスにおける動画に対する、いわゆる投げ銭により、動画を評価するソーシャルプラットフォームが開示されている。特許文献1に記載のソーシャルプラットフォームでは、投げ銭を行うためのアイテム画像を表示させると、そのアイテム画像に対応する金額の投げ銭の決済が行われる決済手段が設けられ、投げ銭機能が実装されている。 There are some examples of evaluation methods for videos on the Internet. Patent Document 1 discloses a social platform for evaluating a video by so-called throwing money for a video in a video sharing service on the Internet. In the social platform described in Patent Document 1, when an item image for making money is displayed, a payment means for settlement of the amount of money corresponding to the item image is provided, and a money throwing function is implemented.
特開2012-120098号公報Japanese Unexamined Patent Publication No. 2012-121998
 しかしながら、未公開の映像コンテンツ、例えば映画のような映像コンテンツに対する評価を得るためには、試写会のような場所でアンケートを取るような手法しかないのが現状であり、このような映像コンテンツに対する視聴者の反応を示す情報を、例えばシーンごとに取得することが可能な技術が望まれていた。 However, in order to obtain an evaluation for unpublished video content, for example, video content such as a movie, the current situation is that there is only a method of taking a questionnaire at a place such as a preview, and for such video content. There has been a demand for a technique capable of acquiring information indicating a viewer's reaction for each scene, for example.
 そこで、本開示では、映像コンテンツを視聴する視聴者の満足度を算出することが可能な技術について説明する。 Therefore, in this disclosure, a technology capable of calculating the satisfaction level of a viewer who views video content will be described.
 本開示の一実施形態によると、制御部を備える情報処理装置が提供される。制御部は、映像コンテンツを視聴する視聴者を撮影した画像データから、視聴者の目を検知し、視聴者の目の開き度合いを取得する開き度合い取得部と、画像データから、視聴者の目を検知し、視聴者のまばたきの回数を取得するまばたき回数取得部と、開き度合い、及びまばたきの回数をパラメータとして、視聴者の映像コンテンツに対する満足度を算出する算出部と、を備える。 According to one embodiment of the present disclosure, an information processing device including a control unit is provided. The control unit detects the viewer's eyes from the image data of the viewer viewing the video content and acquires the viewer's eye opening degree, and the viewer's eye from the image data. It is provided with a blinking number acquisition unit that detects the number of blinks of the viewer and acquires the number of blinks of the viewer, and a calculation unit that calculates the satisfaction level of the viewer with respect to the video content by using the degree of opening and the number of blinks as parameters.
 本開示の一実施形態によると、カメラと、情報処理装置と、を備えるシステムが提供される。カメラは、映像コンテンツを視聴する視聴者を撮影し、画像データを生成して情報処理装置に取得させ、情報処理装置は、制御部を備え、制御部は、映像コンテンツを視聴する視聴者を撮影した画像データから、視聴者の目を検知し、視聴者の目の開き度合いを取得する開き度合い取得部と、画像データから、視聴者の目を検知し、視聴者のまばたきの回数を取得するまばたき回数取得部と、開き度合い、及びまばたきの回数をパラメータとして、視聴者の映像コンテンツに対する満足度を算出する算出部と、を備える。 According to one embodiment of the present disclosure, a system including a camera and an information processing device is provided. The camera captures a viewer viewing the video content, generates image data and causes the information processing device to acquire the image data, the information processing device includes a control unit, and the control unit captures the viewer viewing the video content. The opening degree acquisition unit that detects the viewer's eyes and acquires the viewer's eye opening degree from the image data, and detects the viewer's eyes from the image data and acquires the number of blinks of the viewer. It is provided with a blinking count acquisition unit and a calculation unit for calculating the viewer's satisfaction with the video content using the degree of opening and the number of blinks as parameters.
 一実施形態によると、プロセッサを備えるコンピュータにより実行されるための方法が提供される。方法は、プロセッサが、映像コンテンツを視聴する視聴者を撮影した画像データから、視聴者の目を検知し、視聴者の目の開き度合いを取得するステップと、画像データから、視聴者の目を検知し、視聴者のまばたきの回数を取得するステップと、開き度合い、及びまばたきの回数をパラメータとして、視聴者の映像コンテンツに対する満足度を算出するステップと、を実行する。 According to one embodiment, a method for being executed by a computer equipped with a processor is provided. The method is a step in which the processor detects the viewer's eyes from the image data of the viewer viewing the video content and obtains the degree of opening of the viewer's eyes, and the viewer's eyes are obtained from the image data. A step of detecting and acquiring the number of blinks of the viewer and a step of calculating the satisfaction level of the viewer with respect to the video content are executed by using the degree of opening and the number of blinks as parameters.
 また、一実施形態によると、プロセッサを備えるコンピュータに実行させるためのプログラムが提供される。プログラムは、プロセッサに、映像コンテンツを視聴する視聴者を撮影した画像データから、視聴者の目を検知し、視聴者の目の開き度合いを取得するステップと、画像データから、視聴者の目を検知し、視聴者のまばたきの回数を取得するステップと、開き度合い、及びまばたきの回数をパラメータとして、視聴者の映像コンテンツに対する満足度を算出するステップと、を実行させる。 Further, according to one embodiment, a program for causing a computer having a processor to execute the program is provided. The program tells the processor a step of detecting the viewer's eyes from the image data of the viewer viewing the video content and acquiring the degree of opening of the viewer's eyes, and the viewer's eyes from the image data. A step of detecting and acquiring the number of blinks of the viewer and a step of calculating the satisfaction level of the viewer with respect to the video content are executed by using the degree of opening and the number of blinks as parameters.
 本開示によれば、映像コンテンツを視聴する視聴者を撮影した画像データから、視聴者の目の開き度合い、及びまばたきの回数を取得し、これらに基づき、視聴者の映像コンテンツに対する満足度を算出する。そのため、映像コンテンツに対する視聴者の反応を示す情報を、例えばシーンごとに取得することが可能である。これにより、映像コンテンツに対する評価を分析することが可能になる。 According to the present disclosure, the degree of opening of the viewer's eyes and the number of blinks are acquired from the image data of the viewer viewing the video content, and the satisfaction level of the viewer with the video content is calculated based on these. do. Therefore, it is possible to acquire information indicating the reaction of the viewer to the video content, for example, for each scene. This makes it possible to analyze the evaluation of the video content.
満足度算出システム1の全体の構成を示す図である。It is a figure which shows the whole structure of the satisfaction degree calculation system 1. 満足度算出システム1が設置された施設の外観の例を示す鳥瞰図である。It is a bird's-eye view which shows the example of the appearance of the facility where the satisfaction calculation system 1 is installed. 音声出力装置30の外観の例を示す斜視図である。It is a perspective view which shows the example of the appearance of the voice output device 30. 実施の形態1の満足度算出システム1を構成するサーバ10の機能的な構成を示すブロック図である。It is a block diagram which shows the functional configuration of the server 10 which comprises the satisfaction calculation system 1 of Embodiment 1. FIG. 図4の開き度合い取得モジュール1034における目の開き度合いを算出する具体的手法を示す模式図である。It is a schematic diagram which shows the specific method of calculating the degree of opening of eyes in the degree of opening degree acquisition module 1034 of FIG. 実施の形態1の満足度算出システム1を構成する音声出力装置30の機能的な構成を示す図である。It is a figure which shows the functional configuration of the audio output device 30 which comprises the satisfaction calculation system 1 of Embodiment 1. FIG. サーバ10が記憶する映像コンテンツデータベース181、視聴者データベース184のデータ構造を示す図である。It is a figure which shows the data structure of the video content database 181 and the viewer database 184 stored in a server 10. 実施の形態1の満足度算出システム1による映像コンテンツ出力処理を行う流れの一例を示すフローチャートである。It is a flowchart which shows an example of the flow which performs the video content output processing by the satisfaction calculation system 1 of Embodiment 1. FIG. 実施の形態1の満足度算出システム1による満足度算出処理を行う流れの一例を示すフローチャートである。It is a flowchart which shows an example of the flow which performs the satisfaction calculation processing by the satisfaction calculation system 1 of Embodiment 1. FIG. 満足度算出処理における各パラメータの推移を表示したサーバ10の画面例を示す図である。It is a figure which shows the screen example of the server 10 which displayed the transition of each parameter in the satisfaction degree calculation process.
 以下、図面を参照しつつ、本開示の実施の形態について説明する。以下の説明では、同一の部品には同一の符号を付してある。それらの名称及び機能も同じである。従って、それらについての詳細な説明は繰り返さない。 Hereinafter, embodiments of the present disclosure will be described with reference to the drawings. In the following description, the same parts are designated by the same reference numerals. Their names and functions are the same. Therefore, the detailed description of them will not be repeated.
 <概要>
 以下、満足度算出システムにおける、映像コンテンツを視聴した視聴者の満足度の算出処理について説明する。満足度算出システムは、映画のような映像コンテンツを表示(上映)して視聴者に視聴する機会を提供し、そのときに視聴者が当該映像コンテンツを視聴している状態を撮影するとともに、視聴者の状態をセンサで取得し、撮影した画像データ及びセンサの検知信号から、視聴者の分析を行い、満足度を算出するシステムである。
<Overview>
Hereinafter, the calculation process of the satisfaction of the viewer who has viewed the video content in the satisfaction calculation system will be described. The satisfaction calculation system provides the viewer with an opportunity to display (screen) a video content such as a movie, and at that time, captures the state in which the viewer is watching the video content and also watches the video content. It is a system that acquires the state of a person with a sensor, analyzes the viewer from the captured image data and the detection signal of the sensor, and calculates the degree of satisfaction.
 ここで、満足度とは、映像コンテンツを視聴した視聴者の生体反応等から推測される、映像コンテンツに対する満足の度合いを数値化したものである。一般に、(映像コンテンツに限らず)何かを視聴している者がその対象について集中して視聴している場合、目(瞼)の開きが大きくなり、まばたきの回数が減少し、顔の動き(回数及び動きの大きさ)が少なくなり、脈拍数及び体温が増加することが経験的に知られている。そのため、映像コンテンツを視聴している視聴者が、上記のような反応を示している場合、当該視聴者はその映像コンテンツの視聴に集中していると考えられる。 Here, the degree of satisfaction is a numerical value of the degree of satisfaction with the video content, which is estimated from the biological reaction of the viewer who viewed the video content. In general, when a person watching something (not limited to video content) concentrates on the subject, the eyes (eyelids) open wider, the number of blinks decreases, and the movement of the face. It is empirically known that (number of times and magnitude of movement) decreases and pulse rate and body temperature increase. Therefore, when the viewer who is viewing the video content shows the above reaction, it is considered that the viewer is concentrating on viewing the video content.
 また、映像コンテンツを視聴している視聴者は、その映像コンテンツを面白い、楽しいと感じている場合、集中力が上昇すると考えられ、その映像コンテンツをつまらない、退屈なものだと感じている場合、集中力が散漫になって低下すると考えられる。以上より、視聴者の目の開き、まばたきの回数、顔の動き、脈拍数及び体温と、視聴した映像コンテンツの満足度との間に因果関係があると考えられる。 Also, if the viewer who is watching the video content finds the video content interesting and fun, he / she is considered to be more focused, and if he / she finds the video content boring or boring. It is thought that concentration becomes distracted and decreases. From the above, it is considered that there is a causal relationship between the viewer's eye opening, blinking frequency, face movement, pulse rate and body temperature, and the satisfaction level of the viewed video content.
 そこで、本開示の満足度算出システムでは、視聴者が当該映像コンテンツを視聴している状態をカメラで撮影し、撮影した画像データから、視聴者の顔を識別してそれぞれの視聴者の目を検知する。画像データにおける視聴者の目の動きから、目の開き度合いと、まばたきの回数を取得する。また、視聴者の顔の動き、及び生体情報を検知するセンサの検知信号により、視聴者の顔の動きを取得し、視聴者の生体情報を取得する。取得した目の開き度合い、まばたきの回数、顔の動き、及び生体情報に基づき、映像コンテンツに対する満足の度合いを示す満足度を算出する。このように構成することで、視聴者の生体反応に基づく、視聴者の映像コンテンツに対する満足度を確実に取得することを可能にしている。 Therefore, in the satisfaction calculation system of the present disclosure, the state in which the viewer is viewing the video content is photographed with a camera, the face of the viewer is identified from the photographed image data, and the eyes of each viewer are seen. Detect. From the movement of the viewer's eyes in the image data, the degree of eye opening and the number of blinks are acquired. In addition, the movement of the viewer's face is acquired by the detection signal of the sensor that detects the movement of the viewer's face and the biological information, and the biological information of the viewer is acquired. Based on the acquired degree of eye opening, number of blinks, face movement, and biometric information, the degree of satisfaction indicating the degree of satisfaction with the video content is calculated. With such a configuration, it is possible to surely acquire the satisfaction level of the viewer's video content based on the biological reaction of the viewer.
 また、本開示の満足度算出システムでは、映像コンテンツを視聴する際に、ヘッドホンやイヤホンのような音声出力装置を装着し、音声出力装置から音声を聞くように構成されている。その音声出力装置には、上記のセンサが配置され、視聴者が音声出力装置を装着することで、視聴者の顔の動き及び視聴者の生体情報を検知できるようになっている。このように構成することで、視聴者の顔の動き、及び視聴者の生体情報を確実に取得することを可能にしている。 Further, in the satisfaction calculation system of the present disclosure, when viewing video content, an audio output device such as headphones or earphones is attached, and the audio is heard from the audio output device. The above-mentioned sensor is arranged in the audio output device, and the viewer can detect the movement of the viewer's face and the biological information of the viewer by wearing the audio output device. With this configuration, it is possible to reliably acquire the movement of the viewer's face and the biological information of the viewer.
 さらに、本開示の満足度算出システムでは、映像コンテンツを視聴している視聴者をそれぞれ識別し、視聴者ごとに満足度を算出している。これにより、視聴者の属性ごとに満足度を取得し、映像コンテンツに対する評価を分析することを可能にしている。 Further, in the satisfaction calculation system of the present disclosure, each viewer who is viewing the video content is identified, and the satisfaction is calculated for each viewer. This makes it possible to acquire satisfaction for each attribute of the viewer and analyze the evaluation of the video content.
 <第1の実施の形態>
 以下、満足度算出システム1について説明する。
<First Embodiment>
Hereinafter, the satisfaction calculation system 1 will be described.
 <1 満足度算出システム1の全体構成>
 図1は、満足度算出システム1の全体の構成を示す図である。満足度算出システムは、映画のような映像コンテンツを表示(上映)して視聴者に視聴する機会を提供し、そのときの視聴者の分析を行い、満足度を算出するためのシステムである。図1に示すように、満足度算出システム1は、サーバ10と、映像出力装置20と、複数の音声出力装置(図1では、音声出力装置30A及び音声出力装置30Bを示している。以下、総称して「音声出力装置30」ということもある)と、カメラ40とを含む。サーバ10と映像出力装置20と音声出力装置30とカメラ40とは、ネットワーク80を介して相互に通信可能に接続されている。ネットワーク80は、有線または無線ネットワークにより構成される。
<1 Overall configuration of satisfaction calculation system 1>
FIG. 1 is a diagram showing the overall configuration of the satisfaction calculation system 1. The satisfaction calculation system is a system for displaying (screening) video content such as a movie, providing the viewer with an opportunity to view the content, analyzing the viewer at that time, and calculating the satisfaction. As shown in FIG. 1, the satisfaction calculation system 1 shows a server 10, a video output device 20, and a plurality of audio output devices (in FIG. 1, the audio output device 30A and the audio output device 30B are shown below. Collectively, it may be referred to as "audio output device 30") and the camera 40. The server 10, the video output device 20, the audio output device 30, and the camera 40 are connected to each other so as to be able to communicate with each other via the network 80. The network 80 is composed of a wired or wireless network.
 サーバ10は、視聴者に視聴させる映像コンテンツや、視聴者を撮影した画像データ、センサによる視聴者の検知信号、及び、視聴者ごとの映像コンテンツに対する満足度の情報を管理する装置である。サーバ10は、映像コンテンツの映像データを映像出力装置20へ送信し、音声データを音声出力装置30へ送信する。また、サーバ10は、視聴者を撮影した画像データ及びセンサの検知信号を取得し、視聴者の満足度を算出する。 The server 10 is a device that manages video content to be viewed by the viewer, image data obtained by capturing the viewer, a viewer detection signal by a sensor, and information on satisfaction with the video content for each viewer. The server 10 transmits the video data of the video content to the video output device 20, and transmits the audio data to the audio output device 30. In addition, the server 10 acquires the image data of the viewer and the detection signal of the sensor, and calculates the satisfaction level of the viewer.
 サーバ10は、ネットワーク80を介して映像出力装置20、音声出力装置30及びカメラ40と通信可能に接続される。サーバ10は、4G、5G、LTE(Long Term Evolution)等の通信規格に対応した無線基地局81、IEEE(Institute of Electrical and Electronics Engineers)802.11などの無線LAN(Local Area Network)規格に対応した無線LANルータ82等の通信機器と通信することにより、ネットワーク80に接続される。ネットワーク80のその他の構成として、各装置間を無線または有線で接続してもよく、無線で接続する場合、通信プロトコルとして例えば、Z-Wave(登録商標)、ZigBee(登録商標)、Bluetooth(登録商標)等が使用されてもよく、有線で接続する場合、USB(Universal Serial Bus)ケーブル等により接続してもよい。 The server 10 is communicably connected to the video output device 20, the audio output device 30, and the camera 40 via the network 80. The server 10 supports wireless base stations 81 that support communication standards such as 4G, 5G, and LTE (Long Term Evolution), and wireless LAN (Local Area Network) standards such as IEEE (Institute of Electrical and Electronics Engineers) 802.11. It is connected to the network 80 by communicating with a communication device such as the wireless LAN router 82. As another configuration of the network 80, each device may be connected wirelessly or by wire, and when connecting wirelessly, the communication protocols include, for example, Z-Wave (registered trademark), ZigBee (registered trademark), and Bluetooth (registered). A trademark) or the like may be used, and when connecting by wire, a USB (Universal Serial Bus) cable or the like may be used for connection.
 サーバ10は、ネットワーク80に接続されたコンピュータである。サーバ10は、通信IF(Interface)12と、入出力IF13と、メモリ15と、ストレージ16と、プロセッサ19とを備える。 The server 10 is a computer connected to the network 80. The server 10 includes a communication IF (Interface) 12, an input / output IF 13, a memory 15, a storage 16, and a processor 19.
 通信IF12は、サーバ10が外部の装置と通信するため、信号を入出力するためのインタフェースである。入出力IF13は、ユーザからの入力操作を受け付けるための入力装置(例えば、マウス等のポインティングデバイス、キーボード)、及び、ユーザに対し情報を提示するための出力装置(ディスプレイ、スピーカ等)とのインタフェースとして機能する。メモリ15は、プログラム、及び、プログラム等で処理されるデータ等を一時的に記憶するためのものであり、例えばDRAM(Dynamic Random Access Memory)等の揮発性のメモリである。ストレージ16は、データを保存するための記憶装置であり、例えばフラッシュメモリ、HDD(Hard Disc Drive)、SSD(Solid State Drive)である。プロセッサ19は、プログラムに記述された命令セットを実行するためのハードウェアであり、演算装置、レジスタ、周辺回路などにより構成される。 The communication IF 12 is an interface for inputting / outputting signals because the server 10 communicates with an external device. The input / output IF 13 is an interface with an input device (for example, a pointing device such as a mouse, a keyboard) for receiving an input operation from a user, and an output device (display, a speaker, etc.) for presenting information to the user. Functions as. The memory 15 is for temporarily storing a program, data processed by the program or the like, and is a volatile memory such as a DRAM (Dynamic Random Access Memory). The storage 16 is a storage device for storing data, and is, for example, a flash memory, an HDD (Hard Disc Drive), or an SSD (Solid State Drive). The processor 19 is hardware for executing an instruction set described in a program, and is composed of an arithmetic unit, registers, peripheral circuits, and the like.
 映像出力装置20は、サーバ10からの指示により、映像コンテンツの映像データを受信して映像として表示(上映)を行う装置であり、例えば、フレームの縁にLED素子を配列したLEDパネルや、有機ELディスプレイ、または液晶ディスプレイにより構成されている。この映像出力装置20は、映画館のような上映施設等に設けられている。なお、映像出力装置20は、例えば複数のLEDパネルを並列に配置して数100インチ程度の大型ビジョンとして構成し、一度に大人数の視聴者により視聴が可能に構成してもよい。さらに、映像出力装置20は1つに限られず、複数の映像出力装置20により構成してもよく、複数の映像出力装置20を備えた施設を1のサーバ10で、例えば遠隔により制御するように構成してもよい。 The video output device 20 is a device that receives video data of video content and displays (screens) it as a video according to an instruction from the server 10. For example, an LED panel in which LED elements are arranged on the edge of a frame or an organic device. It is composed of an EL display or a liquid crystal display. The video output device 20 is provided in a screening facility such as a movie theater. The video output device 20 may be configured such that a plurality of LED panels are arranged in parallel to form a large-scale vision of about several hundred inches so that a large number of viewers can view the video at a time. Further, the video output device 20 is not limited to one, and may be configured by a plurality of video output devices 20, so that a facility provided with the plurality of video output devices 20 can be remotely controlled by one server 10, for example. It may be configured.
 音声出力装置30は、サーバ10からの指示により、映像コンテンツの音声データを音声として出力する装置であり、音声出力装置30から出力される音声は、映像出力装置20に上映される映像と同期するように構成されている。なお、上記のように複数の映像出力装置20が設けられ、それぞれ異なる映像コンテンツが上映されている場合、上映されている映像コンテンツに対応する音声が出力されるように構成し、複数の音声が出力されるチャネルから選択可能に構成してもよい。 The audio output device 30 is a device that outputs audio data of video content as audio in response to an instruction from the server 10, and the audio output from the audio output device 30 synchronizes with the video displayed on the video output device 20. It is configured as follows. In addition, when a plurality of video output devices 20 are provided as described above and different video contents are screened, the sound corresponding to the screened video content is configured to be output, and the plurality of sounds are output. It may be configured to be selectable from the output channel.
 音声出力装置30は、映像コンテンツを視聴する視聴者の耳を覆うように装着され、サーバ10の制御により、映像コンテンツの音声を無線通信により受信して音声として出力するワイヤレスヘッドホン等により実現される。この他、音声出力装置30は、例えば有線で接続されるヘッドホンや、視聴者の耳を覆うものに限られず、無線または有線で接続されるイヤホンであるとしてもよい。 The audio output device 30 is mounted so as to cover the ears of the viewer who views the video content, and is realized by wireless headphones or the like that receive the audio of the video content by wireless communication and output it as audio under the control of the server 10. .. In addition, the audio output device 30 is not limited to, for example, headphones connected by wire or those that cover the ears of the viewer, and may be earphones connected wirelessly or by wire.
 図1に音声出力装置30Bとして示すように、音声出力装置30は、通信IF32と、センサ部33と、スピーカ34と、メモリ35と、記憶部36と、プロセッサ39とを備える。 As shown as the voice output device 30B in FIG. 1, the voice output device 30 includes a communication IF 32, a sensor unit 33, a speaker 34, a memory 35, a storage unit 36, and a processor 39.
 通信IF32は、音声出力装置30が外部の装置と通信するため、信号を入出力するためのインタフェースである。センサ部33は、映像コンテンツを視聴している視聴者の顔の動きと生体情報とを取得する装置であり、例えば、視聴者の顔の動きを角速度として検知するジャイロセンサと、脈拍や体温を検知する温度センサや光センサ等の心電センサにより構成されている。センサ部33は、センサ及びその他の部品を集積化したMEMS(Micro Electro Mechanical Systems)により構成してもよい。すなわち、センサ部33は、振動検知、体温計、脳波計、心拍計のような機能を備えている。 The communication IF 32 is an interface for inputting / outputting signals because the voice output device 30 communicates with an external device. The sensor unit 33 is a device that acquires the facial movement and biological information of the viewer who is viewing the video content. For example, a gyro sensor that detects the facial movement of the viewer as an angular velocity and a pulse or body temperature are detected. It is composed of an electrocardiographic sensor such as a temperature sensor and an optical sensor that detect it. The sensor unit 33 may be configured by MEMS (Micro Electro Mechanical Systems) in which a sensor and other parts are integrated. That is, the sensor unit 33 has functions such as vibration detection, a thermometer, an electroencephalograph, and a heart rate monitor.
 スピーカ34は、音声信号を音声に変換し、ユーザに対し情報を音声として出力するための出力装置である。メモリ35は、プログラム、及び、プログラム等で処理されるデータ等を一時的に記憶するためのものであり、例えばDRAM(Dynamic Random Access Memory)等の揮発性のメモリである。記憶部36は、データを保存するための記憶装置であり、例えばフラッシュメモリ、HDD(Hard Disc Drive)、SSD(Solid State Drive)である。プロセッサ39は、プログラムに記述された命令セットを実行するためのハードウェアであり、演算装置、レジスタ、周辺回路などにより構成される。 The speaker 34 is an output device for converting a voice signal into voice and outputting information as voice to a user. The memory 35 is for temporarily storing a program, data processed by the program or the like, and is a volatile memory such as a DRAM (Dynamic Random Access Memory). The storage unit 36 is a storage device for storing data, and is, for example, a flash memory, an HDD (Hard Disc Drive), or an SSD (Solid State Drive). The processor 39 is hardware for executing an instruction set described in a program, and is composed of an arithmetic unit, registers, peripheral circuits, and the like.
 カメラ40は、サーバ10からの指示により、映像コンテンツを視聴している視聴者を撮影する装置であり、例えば、CCD(Charge Coupled Device)イメージセンサ等の撮像装置と、撮影した画像をデジタル信号からなる画像データに変換するA/D変換装置とを備えるデジタルカメラ等により構成されている。カメラ40は、サーバ10の制御により撮影する方向を変更するように構成してもよく、広範囲を撮影可能にするために撮影方向を適宜動かせるように構成してもよい。 The camera 40 is a device that captures a viewer who is viewing video content according to an instruction from the server 10. For example, an image pickup device such as a CCD (Charge Coupled Device) image sensor and the captured image are captured from a digital signal. It is composed of a digital camera or the like equipped with an A / D conversion device for converting the image data. The camera 40 may be configured to change the shooting direction under the control of the server 10, or may be configured to appropriately move the shooting direction in order to enable shooting in a wide range.
 カメラ40は、映像コンテンツを視聴している視聴者を動画として撮影し、または所定の時間間隔で静止画として撮影し、動画または静止画の画像データへの変換を行う。また、カメラ40は、撮影した画像データをサーバ10へ送信する。なお、カメラ40は1つに限られず、1つの施設に複数のカメラ40が設けられてもよい。また、カメラ40は単体の装置とは限られず、例えば、移動体通信システムに対応したスマートフォン、タブレット等の携帯端末に設けられたカメラ機能により構成してもよい。 The camera 40 shoots a viewer viewing the video content as a moving image, or shoots as a still image at predetermined time intervals, and converts the moving image or the still image into image data. Further, the camera 40 transmits the captured image data to the server 10. The number of cameras 40 is not limited to one, and a plurality of cameras 40 may be provided in one facility. Further, the camera 40 is not limited to a single device, and may be configured by, for example, a camera function provided in a mobile terminal such as a smartphone or tablet compatible with a mobile communication system.
 <1.1 映像出力装置20及び音声出力装置30の構成>
 図2は、満足度算出システム1が設置された施設の外観の例を示す鳥瞰図である。図2に示す満足度算出システム1は、例えば映画館のように映像コンテンツの上映を行うスペースに設置された例を示している。また、満足度算出システム1は、視聴者がヘッドホン等の音声出力装置30を装着して音声を取得するように構成されているため、映像コンテンツを視聴する以外の、飲食やその他の作業も可能となっており、いわゆるカフェスペースも併設され、多目的スペースを構成している施設である。さらに、満足度算出システム1は、例えば視聴者の自宅等で構成してもよい。例えば、サーバ10をクラウドサーバ等で構成し、サーバ10からオンデマンド等のサービスで、視聴者が使用するテレビやディスプレイ等により構成する映像出力装置20に映像を出力させ、視聴者が使用するヘッドホン等により構成する音声出力装置30に音声を出力させ、視聴者が使用するスマートフォン等により構成するカメラ40で視聴者を撮影してその画像データをサーバ10へ送信させてもよい。
<1.1 Configuration of video output device 20 and audio output device 30>
FIG. 2 is a bird's-eye view showing an example of the appearance of the facility in which the satisfaction calculation system 1 is installed. The satisfaction calculation system 1 shown in FIG. 2 shows an example in which the satisfaction calculation system 1 is installed in a space for showing video contents, such as a movie theater. Further, since the satisfaction calculation system 1 is configured so that the viewer wears an audio output device 30 such as headphones to acquire audio, it is possible to eat, drink, and perform other work other than viewing the video content. It is a facility that constitutes a multipurpose space with a so-called cafe space. Further, the satisfaction calculation system 1 may be configured at, for example, the viewer's home. For example, the server 10 is composed of a cloud server or the like, and the server 10 outputs video to a video output device 20 composed of a television or a display used by the viewer by a service such as on-demand, and headphones used by the viewer. The audio output device 30 configured by the above may output the audio, the viewer may be photographed by the camera 40 configured by the smartphone or the like used by the viewer, and the image data may be transmitted to the server 10.
 図2に示す満足度算出システム1は、映像コンテンツの上映を行う上映スペースと、カフェスペースとが併設されたフロアFに設けられ、フロアFの所定の箇所(図2に示す右上方向)に映像出力装置20とカメラ40とが設置されている。映像出力装置20が映像コンテンツの映像情報を上映する正面側には、映像コンテンツを視聴者Wが視聴するために着座する席が設けられている。視聴者Wは、前述のように映像コンテンツを視聴する際に音声出力装置30を装着する。 The satisfaction calculation system 1 shown in FIG. 2 is provided on the floor F in which the screening space for screening the video content and the cafe space are provided, and the video is displayed at a predetermined location (upper right direction shown in FIG. 2) on the floor F. An output device 20 and a camera 40 are installed. On the front side where the video output device 20 screens the video information of the video content, a seat is provided for the viewer W to view the video content. The viewer W wears the audio output device 30 when viewing the video content as described above.
 この施設では、視聴者Wが入室するためには会員登録が必要であり、視聴者Wの氏名、性別、年齢等の属性情報を管理し、それぞれの視聴者の動きがカメラ40による撮影により把握されるようになっている。また、この音声出力装置30は、当該施設の管理者により貸与されたものである。これにより、会員登録されている視聴者Wの属性情報と、センサ部33が取得した視聴者Wの情報と、カメラ40が撮影した視聴者Wの画像データとが対応付けられるようになっており、視聴者Wの属性ごとの満足度を把握することが可能になっている。 At this facility, membership registration is required for viewer W to enter the room, attribute information such as viewer W's name, gender, and age is managed, and the movement of each viewer can be grasped by shooting with the camera 40. It is supposed to be done. The audio output device 30 is lent by the manager of the facility. As a result, the attribute information of the viewer W registered as a member, the information of the viewer W acquired by the sensor unit 33, and the image data of the viewer W taken by the camera 40 can be associated with each other. , It is possible to grasp the degree of satisfaction for each attribute of the viewer W.
 さらに、このような構成により、視聴者Wは周囲の状況を気にすることなく、映像コンテンツの視聴に集中することが可能となっており、また、カフェスペースの利用者のように映像コンテンツを視聴していない者は、映像コンテンツの音声を気にすることなく飲食やその他の行為が可能となっている。 Further, with such a configuration, the viewer W can concentrate on viewing the video content without worrying about the surrounding situation, and can enjoy the video content like a user of the cafe space. Those who are not watching can eat, drink, and perform other actions without worrying about the audio of the video content.
 また、映像出力装置20はLEDパネル等により構成されているため、一定の明るさの光を視聴者W側に照射する。そのため、カメラ40は、その光により視聴者Wを一定の解像度で撮影することが可能になっている。 Further, since the video output device 20 is composed of an LED panel or the like, it irradiates the viewer W side with light having a constant brightness. Therefore, the camera 40 can shoot the viewer W at a constant resolution by the light.
 満足度算出システム1が設置された施設は、図2に示すように、映画館のような閉鎖空間とは限られず、例えばコンビニエンスストアのイートインコーナーや、イベントスペースのような商業施設の空間であってもよい。また、飲食店内や、ホテルのラウンジ、マンションのような集合住宅の共有スペースのように、一定の静音性が確保されているスペースに満足度算出システム1を設置してもよい。さらに、パチンコやパチスロ店のような遊技施設や、ゲームセンター、アミューズメント施設のように、一定の雑音がある施設に満足度算出システム1を設置してもよい。 As shown in FIG. 2, the facility in which the satisfaction calculation system 1 is installed is not limited to a closed space such as a movie theater, but is a space of a commercial facility such as an eat-in corner of a convenience store or an event space. You may. Further, the satisfaction calculation system 1 may be installed in a space where a certain level of quietness is ensured, such as in a restaurant, a hotel lounge, or a shared space of an apartment house such as an apartment. Further, the satisfaction calculation system 1 may be installed in a game facility such as a pachinko or pachislot store, or in a facility with a certain amount of noise such as a game center or an amusement facility.
 図3は、音声出力装置30の外観の例を示す斜視図である。音声出力装置30は、図3に示すように、映像コンテンツを視聴する視聴者が音声を聞くため、頭部から両耳を挟むようにして装着するヘッドホンで構成され、両耳をそれぞれ覆う一対の耳収容部37と、一対の耳収容部37を支持すると共に視聴者の頭部を適度な圧力で押圧することで耳収容部37を視聴者の耳に装着させるアーム部38とから構成されている。なお、音声出力装置30はこのような構成に限られず、例えば、アーム部38を備えずに耳収容部37が視聴者の耳たぶに係止するように装着するような構成であってもよく、視聴者の耳の孔の中に挿入するようなカナル型イヤホンであってもよい。 FIG. 3 is a perspective view showing an example of the appearance of the audio output device 30. As shown in FIG. 3, the audio output device 30 is composed of headphones worn so that both ears are sandwiched from the head in order for the viewer viewing the video content to hear the audio, and the pair of ears accommodating each covering both ears. It is composed of a portion 37 and an arm portion 38 that supports the pair of ear accommodating portions 37 and attaches the ear accommodating portion 37 to the viewer's ear by pressing the viewer's head with an appropriate pressure. The audio output device 30 is not limited to such a configuration, and may be configured such that the ear accommodating portion 37 is mounted so as to be locked to the earlobe of the viewer without providing the arm portion 38. It may be a canal type earphone that is inserted into the ear hole of the viewer.
 図3に示すように、一対の耳収容部37の内側における、視聴者の耳の周辺の皮膚に接触する位置には、センサ部33が設けられている。センサ部33が取得する生体情報により、耳収容部37における位置は特定の位置に限られず、耳収容部37内に複数設けられてもよい。また、センサ部33が取得する生体情報の種類により、視聴者の皮膚に直接接触するように配置してもよく、視聴者の皮膚から離隔するように配置してもよい。なお、このような構成であるため、音声出力装置30にはイヤーパッドを装着しないで使用することが望ましいが、イヤーパッドを装着することを妨げるものではない。 As shown in FIG. 3, a sensor unit 33 is provided at a position inside the pair of ear accommodating units 37 in contact with the skin around the viewer's ear. According to the biological information acquired by the sensor unit 33, the position in the ear accommodating unit 37 is not limited to a specific position, and a plurality of positions may be provided in the ear accommodating unit 37. Further, depending on the type of biological information acquired by the sensor unit 33, it may be arranged so as to be in direct contact with the skin of the viewer, or may be arranged so as to be separated from the skin of the viewer. Since the audio output device 30 has such a configuration, it is desirable to use the audio output device 30 without attaching the ear pads, but it does not prevent the ear pads from being attached.
 <1.2 サーバ10の機能的な構成>
 図4は、実施の形態1の満足度算出システム1を構成するサーバ10の機能的な構成を示す図である。図4に示すように、サーバ10は、通信部101と、記憶部102と、制御部103としての機能を発揮する。
<1.2 Functional configuration of server 10>
FIG. 4 is a diagram showing a functional configuration of the server 10 constituting the satisfaction calculation system 1 of the first embodiment. As shown in FIG. 4, the server 10 functions as a communication unit 101, a storage unit 102, and a control unit 103.
 通信部101は、サーバ10が外部の装置と通信するための処理を行う。 The communication unit 101 performs processing for the server 10 to communicate with an external device.
 記憶部102は、サーバ10が使用するデータ及びプログラムを記憶する。記憶部102は、映像コンテンツデータベース181と、センサ検知信号データベース182と、カメラ撮影画像データベース183と、視聴者データベース184等を記憶する。 The storage unit 102 stores data and programs used by the server 10. The storage unit 102 stores the video content database 181, the sensor detection signal database 182, the camera-captured image database 183, the viewer database 184, and the like.
 映像コンテンツデータベース181は、満足度算出システム1において提供する映像コンテンツ及びその上映時間の情報、すなわち映像出力装置20の上映スケジュールを保持するためのデータベースである。詳細は後述する。 The video content database 181 is a database for holding information on the video content and its screening time provided by the satisfaction calculation system 1, that is, the screening schedule of the video output device 20. Details will be described later.
 センサ検知信号データベース182は、満足度算出システム1を利用する視聴者が装着する、音声出力装置30のセンサ部33からの検知信号を時系列で保持するためのデータベースである。センサ部33の検知信号は、日時情報が付加されて音声出力装置30からサーバ10へ送信されるので、その検知信号が格納される。 The sensor detection signal database 182 is a database for holding the detection signals from the sensor unit 33 of the voice output device 30 in chronological order, which is worn by the viewer who uses the satisfaction calculation system 1. Since the detection signal of the sensor unit 33 is transmitted from the voice output device 30 to the server 10 with the date and time information added, the detection signal is stored.
 カメラ撮影画像データベース183は、満足度算出システム1を利用する視聴者を撮影した、カメラ40の画像データを時系列で保持するためのデータベースである。カメラ40の画像データは、日時情報が付加されてカメラ40からサーバ10へ送信されるので、その検知信号が格納される。 The camera-captured image database 183 is a database for holding the image data of the camera 40 in chronological order, which captures the viewers who use the satisfaction calculation system 1. Since the image data of the camera 40 is transmitted from the camera 40 to the server 10 with date and time information added, the detection signal is stored.
 視聴者データベース184は、満足度算出システム1を利用する視聴者の情報と、視聴者ごとに視聴した映像コンテンツの満足度の情報を保持するためのデータベースである。詳細は後述する。なお、視聴者データベース184は、施設に登録されている会員情報(属性情報を含む)を格納してもよい。 The viewer database 184 is a database for holding information on viewers who use the satisfaction calculation system 1 and information on the satisfaction of the video content viewed for each viewer. Details will be described later. The viewer database 184 may store member information (including attribute information) registered in the facility.
 制御部103は、サーバ10のプロセッサがプログラムに従って処理を行うことにより、各種モジュールとして受信制御モジュール1031、送信制御モジュール1032、映像コンテンツ出力制御モジュール1033、開き度合い取得モジュール(開き度合い取得部)1034、まばたき回数取得モジュール(まばたき回数取得部)1035、顔動き取得モジュール(動き取得部)1036、生体情報取得モジュール(生体情報取得部)1037、及び満足度算出モジュール(算出部)1038に示す機能を発揮する。 The control unit 103 performs processing according to a program by the processor of the server 10, and as various modules, the reception control module 1031, the transmission control module 1032, the video content output control module 1033, the opening degree acquisition module (opening degree acquisition unit) 1034, The functions shown in the blink count acquisition module (blink frequency acquisition unit) 1035, face movement acquisition module (movement acquisition unit) 1036, biometric information acquisition module (biological information acquisition unit) 1037, and satisfaction calculation module (calculation unit) 1038 are exhibited. do.
 受信制御モジュール1031は、サーバ10が外部の装置から通信プロトコルに従って信号を受信する処理を制御する。 The reception control module 1031 controls the process in which the server 10 receives a signal from an external device according to the communication protocol.
 送信制御モジュール1032は、サーバ10が外部の装置に対し通信プロトコルに従って信号を送信する処理を制御する。 The transmission control module 1032 controls a process in which the server 10 transmits a signal to an external device according to a communication protocol.
 映像コンテンツ出力制御モジュール1033は、映像コンテンツを視聴者に視聴させるため、映像コンテンツの映像データを映像出力装置20へ送信し、音声データを音声出力装置30へ送信する処理を制御する。映像コンテンツ出力制御モジュール1033は、映像コンテンツデータベース181を参照し、映像コンテンツを上映させる日時を示すスケジュール情報に基づき、映像コンテンツの画像データ及び音声データを送信すると共に、映像コンテンツを上映させるように制御信号を送信し、映像コンテンツのデータ容量やデータ形式、解像度等の情報も送信する。 The video content output control module 1033 controls the process of transmitting the video data of the video content to the video output device 20 and transmitting the audio data to the audio output device 30 in order to allow the viewer to view the video content. The video content output control module 1033 refers to the video content database 181 and controls the video content to be screened while transmitting the image data and the audio data of the video content based on the schedule information indicating the date and time when the video content is screened. It transmits signals and also transmits information such as the data capacity, data format, and resolution of video content.
 このとき、映像コンテンツ出力制御モジュール1033は、映像出力装置20及び音声出力装置30に対して、画像データ及び音声データを一括送信して上映させてもよく、ストリーミング形式で順次送信して上映させてもよい。なお、映像コンテンツ出力制御モジュール1033は、映像コンテンツの上映時間やクリエイターの氏名または名称等のクレジット情報を送信してもよい。 At this time, the video content output control module 1033 may collectively transmit and screen the image data and the audio data to the video output device 20 and the audio output device 30, and sequentially transmit and screen the image data and the audio data in a streaming format. May be good. The video content output control module 1033 may transmit credit information such as the screening time of the video content and the name or name of the creator.
 開き度合い取得モジュール1034は、カメラ40が撮影した画像データを取得し、画像データを解析して視聴者の目を認識して検知し、視聴者の目の開き度合いを取得する処理を制御する。開き度合い取得モジュール1034は、視聴者の目を識別する手法として、公知技術である機械学習等による画像認識技術が用いられる。 The opening degree acquisition module 1034 controls a process of acquiring image data taken by the camera 40, analyzing the image data, recognizing and detecting the viewer's eyes, and acquiring the viewer's eye opening degree. The opening degree acquisition module 1034 uses an image recognition technique such as machine learning, which is a known technique, as a method for identifying the eyes of the viewer.
 例えば、開き度合い取得モジュール1034は、画像データを解析して視聴者の目を検知し、画像データにおける視聴者の目を開いた部分の面積を算出し、この面積を目の開き度合いとして取得する。また、開き度合い取得モジュール1034は、画像データを解析して視聴者の目を検知し、画像データにおける視聴者の目を開いた部分における、上瞼と下瞼との接点(目の左右の端点)と、上瞼の上方の端点、または虹彩(いわゆる黒目)と上瞼との接点と、下瞼の下方の端点、または虹彩と下瞼との接点との位置の情報を算出し、位置の情報に基づき、視聴者の目の開き度合いとして取得する。 For example, the opening degree acquisition module 1034 analyzes the image data to detect the viewer's eyes, calculates the area of the portion of the image data in which the viewer's eyes are opened, and acquires this area as the degree of opening of the eyes. .. Further, the opening degree acquisition module 1034 analyzes the image data to detect the viewer's eyes, and the contact point between the upper eyelid and the lower eyelid (the left and right end points of the eyes) in the portion of the image data where the viewer's eyes are opened. ) And the upper end point of the upper eyelid, or the contact point between the iris (so-called black eye) and the upper eyelid, and the lower end point of the lower eyelid, or the position information of the contact point between the iris and the lower eyelid. Based on the information, it is acquired as the degree of eyelid opening of the viewer.
 図5は、図4の開き度合い取得モジュール1034における目の開き度合いを算出する具体的手法を示す模式図である。上記の目の開き度合いを算出する手法について、図5を用いて説明する。図5には、視聴者の目を模式的に示した目Eが示されており、目Eには、上瞼L1と、下瞼L2と、虹彩L3とが示されている。ここで、図5(a)は、虹彩L3が上瞼L1及び下瞼L2と接している場合の例であり、図5(b)は、虹彩L3が上瞼L1及び下瞼L2と接していない場合の例である。 FIG. 5 is a schematic diagram showing a specific method for calculating the degree of opening of the eyes in the degree of opening degree acquisition module 1034 of FIG. The method for calculating the degree of eye opening will be described with reference to FIG. FIG. 5 shows eyes E schematically showing the eyes of the viewer, and the eyes E show the upper eyelid L1, the lower eyelid L2, and the iris L3. Here, FIG. 5A is an example in which the iris L3 is in contact with the upper eyelid L1 and the lower eyelid L2, and FIG. 5B is an example in which the iris L3 is in contact with the upper eyelid L1 and the lower eyelid L2. This is an example when there is no such thing.
 図5(a)に示すように、虹彩L3が上瞼L1及び下瞼L2と接している場合、画像データを解析することにより、上瞼L1と下瞼L2との接点P1,P2と、虹彩L3と上瞼L1との接点P3,P4と、虹彩L3と下瞼L2との接点P5,P6との位置を認識することが可能である。そこで、視聴者の目Eの画像データにおける、接点P1~P6の座標を算出し、座標の位置情報に基づき、視聴者の目の開き度合いを取得する。 As shown in FIG. 5A, when the iris L3 is in contact with the upper eyelid L1 and the lower eyelid L2, the contact points P1 and P2 between the upper eyelid L1 and the lower eyelid L2 and the iris are analyzed by analyzing the image data. It is possible to recognize the positions of the contacts P3 and P4 between L3 and the upper eyelid L1 and the contacts P5 and P6 between the iris L3 and the lower eyelid L2. Therefore, the coordinates of the contacts P1 to P6 in the image data of the viewer's eyes E are calculated, and the degree of opening of the viewer's eyes is acquired based on the position information of the coordinates.
 また、図5(b)に示すように、虹彩L3が上瞼L1及び下瞼L2と接していない場合、画像データを解析することにより、上瞼L1と下瞼L2との接点P1,P2と、上瞼L1の上方の端点P7と、下瞼L2の下方の端点P8との位置を認識することが可能である。そこで、視聴者の目Eの画像データにおける、接点P1,P2、端点P7,P8の座標を算出し、座標の位置情報に基づき、視聴者の目の開き度合いを取得する。なお、例えば虹彩L3が上瞼L1と接しているが、下瞼L2と接していないような場合には、上瞼L1については図5(a)に示す手法で接点P3,P4を認識し、下瞼L2については図5(b)に示す手法で端点P8を認識してもよく、逆の場合も同様である。 Further, as shown in FIG. 5B, when the iris L3 is not in contact with the upper eyelid L1 and the lower eyelid L2, the contact points P1 and P2 between the upper eyelid L1 and the lower eyelid L2 are obtained by analyzing the image data. It is possible to recognize the positions of the upper end point P7 of the upper eyelid L1 and the lower end point P8 of the lower eyelid L2. Therefore, the coordinates of the contacts P1, P2 and the end points P7 and P8 in the image data of the viewer's eyes E are calculated, and the degree of opening of the viewer's eyes is acquired based on the position information of the coordinates. For example, when the iris L3 is in contact with the upper eyelid L1 but not the lower eyelid L2, the contact points P3 and P4 are recognized for the upper eyelid L1 by the method shown in FIG. 5A. For the lower eyelid L2, the end point P8 may be recognized by the method shown in FIG. 5 (b), and vice versa.
 まばたき回数取得モジュール1035は、カメラ40が撮影した画像データを取得し、画像データを解析して視聴者の目を認識して検知し、視聴者の目におけるまばたきの回数を取得する処理を制御する。まばたき回数取得モジュール1035は、視聴者の目を識別する手法として、開き度合い取得モジュール1034と同様に、公知技術である機械学習等による画像認識技術が用いられる。なお、人の目のまばたきを画像データとして取得するためには、一般には50fps(frame per second)以上のフレームレートが必要とされているため、カメラ40では、50fps以上のフレームレートで撮影が行われる。 The blink count acquisition module 1035 controls a process of acquiring image data taken by the camera 40, analyzing the image data, recognizing and detecting the viewer's eyes, and acquiring the number of blinks in the viewer's eyes. .. Similar to the opening degree acquisition module 1034, the blink count acquisition module 1035 uses an image recognition technique such as machine learning, which is a known technique, as a method for identifying the eyes of the viewer. In addition, in order to acquire the blink of a person's eyes as image data, a frame rate of 50 fps (frame per second) or more is generally required. Therefore, the camera 40 shoots at a frame rate of 50 fps or more. It is said.
 例えば、まばたき回数取得モジュール1035は、画像データを解析して視聴者の目を検知し、所定時間(例えば、1分)における視聴者のまばたきの回数を取得する。また、まばたき回数取得モジュール1035は、視聴者の目が閉じた状態になり、その状態から所定の時間(例えば、1秒)内に視聴者の目が開いた状態になったとき、視聴者がまばたきをしたと認識し、まばたきの回数としてカウントする。このような構成は、視聴者の目が閉じた状態になり、一定時間以上目が開かない場合、当該視聴者はその間映像コンテンツを視聴しているとはいえず、例えば睡眠状態であると考えられるため、睡眠状態と区別するために設定している。 For example, the blink count acquisition module 1035 analyzes the image data to detect the viewer's eyes and acquires the number of blinks of the viewer in a predetermined time (for example, 1 minute). Further, in the blink count acquisition module 1035, when the viewer's eyes are closed and the viewer's eyes are opened within a predetermined time (for example, 1 second) from that state, the viewer Recognizes that you have blinked and counts it as the number of blinks. In such a configuration, if the viewer's eyes are closed and the eyes are not opened for a certain period of time, it cannot be said that the viewer is watching the video content during that period, and it is considered that the viewer is in a sleeping state, for example. Therefore, it is set to distinguish it from the sleeping state.
 顔動き取得モジュール1036は、視聴者が映像コンテンツを視聴している際に装着している音声出力装置30から、センサ部33が検知して出力した検知信号を取得し、検知信号を解析して視聴者の顔の動きを取得する処理を制御する。顔動き取得モジュール1036は、センサ部33を構成するジャイロセンサが検知する角速度とその動きの時間から、視聴者の顔の動きを取得する。 The face movement acquisition module 1036 acquires a detection signal detected and output by the sensor unit 33 from the audio output device 30 worn when the viewer is viewing the video content, and analyzes the detection signal. Controls the process of acquiring the movement of the viewer's face. The face movement acquisition module 1036 acquires the movement of the viewer's face from the angular velocity detected by the gyro sensor constituting the sensor unit 33 and the time of the movement.
 例えば、顔動き取得モジュール1036は、センサ部33を構成するジャイロセンサが検知する角速度とその動きの時間から、視聴者が顔を前後左右に傾けた動き、または視聴者が顔の向きを移動させた動き(左右方向に回転させた動き)を取得する。ここで、視聴者が顔を傾けた動きとは、首を動かすことによる動きでもよく、体幹部分を動かすことによる上半身全体の動きであってもよい。また、取得する視聴者の顔の動きの具体例は、視聴者が所定時間(例えば、10分)内に所定の移動範囲以上に顔を動かした回数でもよく、顔を動かした角度や移動距離でもよい。 For example, in the face movement acquisition module 1036, the viewer tilts the face back and forth and left and right, or the viewer moves the direction of the face based on the angular velocity detected by the gyro sensor constituting the sensor unit 33 and the time of the movement. Acquires the movement (movement rotated in the left-right direction). Here, the movement in which the viewer tilts his / her face may be a movement by moving the neck or a movement of the entire upper body by moving the trunk portion. Further, a specific example of the movement of the viewer's face to be acquired may be the number of times the viewer moves the face beyond the predetermined movement range within a predetermined time (for example, 10 minutes), and the angle or distance of movement of the face may be obtained. But it may be.
 生体情報取得モジュール1037は、視聴者が映像コンテンツを視聴している際に装着している音声出力装置30から、センサ部33が検知して出力した検知信号を取得し、検知信号を解析して視聴者の生体情報を取得する処理を制御する。生体情報取得モジュール1037は、センサ部33を構成する心電センサが検知する、視聴者の脈拍、血圧や体温を取得する。 The biological information acquisition module 1037 acquires a detection signal detected and output by the sensor unit 33 from the audio output device 30 worn when the viewer is viewing the video content, and analyzes the detection signal. Controls the process of acquiring the viewer's biometric information. The biological information acquisition module 1037 acquires the viewer's pulse, blood pressure, and body temperature detected by the electrocardiographic sensor constituting the sensor unit 33.
 満足度算出モジュール1038は、開き度合い取得モジュール1034が取得した目の開き度合い、まばたき回数取得モジュール1035が取得したまばたきの回数、顔動き取得モジュール1036が取得した顔の動き、及び生体情報取得モジュール1037が取得した生体情報をパラメータとして、視聴者の映像コンテンツに対する満足度を算出する処理を制御する。満足度算出モジュール1038は、パラメータとして目の開き度合い及びまばたきの回数のみから満足度を算出してもよく、またはパラメータとして顔の動きを加えてもよく、またはパラメータとして生体情報を加えてもよい。満足度とは、上記のように映像コンテンツを視聴した視聴者の生体反応等から推測される、映像コンテンツに対する満足の度合いを所定時間(例えば、1分)ごとに時系列で数値化したものであり、視聴者ごとに当該視聴者の満足度として算出され、または複数の視聴者の満足度が合計値や平均値として算出されて当該映像コンテンツの評価指標とするものである。また、この満足度は、所定時間ごとの時系列の推移として取得してもよい。 The satisfaction calculation module 1038 includes the degree of eye opening acquired by the degree of opening acquisition module 1034, the number of blinks acquired by the number of blinks acquisition module 1035, the movement of the face acquired by the face movement acquisition module 1036, and the biometric information acquisition module 1037. Controls the process of calculating the viewer's satisfaction with the video content using the biometric information acquired by The satisfaction calculation module 1038 may calculate the satisfaction only from the degree of eye opening and the number of blinks as parameters, may add facial movement as a parameter, or may add biometric information as a parameter. .. Satisfaction is a time-series quantification of the degree of satisfaction with the video content, which is estimated from the biological reaction of the viewer who viewed the video content as described above, at predetermined time intervals (for example, 1 minute). Yes, it is calculated as the satisfaction level of the viewer for each viewer, or the satisfaction level of a plurality of viewers is calculated as a total value or an average value and used as an evaluation index of the video content. Further, this satisfaction level may be acquired as a time-series transition at predetermined time intervals.
 例えば、視聴者の個々の満足度の推移を把握することで、ある映像コンテンツに対する満足度の推移が近い視聴者をグルーピングしたり、グループごとの視聴者の属性等を分析したりすることで、映像コンテンツのマーケティングの指標値として使用してもよい。また、満足度の推移が近い視聴者をレコメンドしたり、満足度の推移が近い他の視聴者が視聴した映像コンテンツを当該視聴者にとっても満足度が高いと推測することで、その映像コンテンツをレコメンドしたりしてもよい。 For example, by grasping the transition of individual satisfaction of viewers, it is possible to group viewers who have similar transitions of satisfaction with a certain video content, or to analyze the attributes of viewers for each group. It may be used as an index value for marketing video content. In addition, by recommending viewers whose satisfaction levels are close to each other, or by assuming that the video content viewed by other viewers whose satisfaction levels are close to each other is also highly satisfying to the viewers, the video content can be selected. You may recommend it.
 ここで、目の大きさやまばたきの回数、顔の動き、生体情報は視聴者の生体活動に依存する値であるため、個人差が大きいものと考えられる。そのため、満足度算出モジュール1038は、例えば最初に取得した視聴者の目や顔、身体状態を初期値(基準値)とし、その後のそれぞれの値の変化、または変化率から満足度を算出してもよく、所定の時間内におけるそれぞれの値の変化、または変化率から満足度を算出してもよく、一般的な視聴者の目や顔、身体状態(例えば、平均値)との差で満足度を算出してもよい。 Here, the size of the eyes, the number of blinks, the movement of the face, and the biological information are values that depend on the biological activity of the viewer, so it is considered that there are large individual differences. Therefore, the satisfaction calculation module 1038 sets, for example, the eyes, face, and physical condition of the viewer first acquired as initial values (reference values), and calculates the satisfaction from the subsequent changes in the respective values or the rate of change. Satisfaction may be calculated from the change in each value within a predetermined time or the rate of change, and the difference from the eyes, face, and physical condition (for example, average value) of a general viewer is satisfied. The degree may be calculated.
 例えば、満足度算出モジュール1038は、上記の例により数値化された目の開き度合い、まばたきの回数、顔の動き、及び生体情報の値をパラメータとして、それぞれ所定の演算を行い、演算後のそれぞれの値を加算することで満足度を算出する。具体的には、目の開き度合いを算出する関数Jを、取得した開き度合いの値jの関数とする。関数Jは、例として、
 J=αj+β
 のような1次関数でもよく、
 J=αj^n+βj^(n-1)+・・・+γ
 のような高次関数でもよく、
 J=α^(j+n)+β
 のような指数関数であってもよい。同様に、まばたきの回数を算出する関数Kを、取得したまばたきの回数の値kの関数とする。顔の動きを算出する関数Lを、取得した顔の動きの値lの関数とする。生体情報を算出する関数Mを、取得した生体情報の値mとする。ここで、それぞれの関数は、それぞれの値をそのまま使用するものであってもよい。満足度Pは、以下の式のようにそれぞれの関数を加算し、
 P=J+K+L+M
 のように算出される。
For example, the satisfaction calculation module 1038 performs predetermined calculations with the parameters of the degree of eye opening, the number of blinks, the movement of the face, and the value of biometric information quantified by the above example, and after the calculation, each of them performs a predetermined calculation. Satisfaction is calculated by adding the values of. Specifically, the function J for calculating the degree of opening of the eyes is a function of the acquired degree of opening value j. The function J is, for example,
J = αj + β
It may be a linear function such as
J = αj ^ n + βj ^ (n-1) + ... + γ
Higher-order functions such as
J = α ^ (j + n) + β
It may be an exponential function such as. Similarly, let the function K for calculating the number of blinks be a function of the value k of the acquired number of blinks. Let the function L for calculating the movement of the face be a function of the acquired value l of the movement of the face. Let the function M for calculating the biological information be the value m of the acquired biological information. Here, each function may use each value as it is. Satisfaction P is calculated by adding each function as shown in the following formula.
P = J + K + L + M
It is calculated as.
 また、例えば、満足度算出モジュール1038は、上記の例により目の開き度合い、まばたきの回数、顔の動き、及び生体情報の関数に対して、それぞれ異なる所定の係数を重み付けとして乗算し、乗算後のそれぞれの値を加算することで満足度を算出する。上記の例の場合、満足度Pは、以下の式
 P=aJ+bK+cL+dM
 により算出される。ここで、a,b,c,dは、それぞれ所定の係数である。
Further, for example, the satisfaction calculation module 1038 multiplies the degree of eye opening, the number of blinks, the movement of the face, and the function of biometric information by different predetermined coefficients as weights according to the above example, and after multiplication. Satisfaction is calculated by adding each value of. In the case of the above example, the satisfaction level P is the following formula P = aJ + bK + cL + dM.
Is calculated by. Here, a, b, c, and d are predetermined coefficients, respectively.
 また、上記のように、映像コンテンツを視聴している者が集中して視聴している場合、目の開きが大きくなり、まばたきの回数が減少し、顔の動きが少なくなり、脈拍数及び体温が増加することが経験的に知られている。そのため、係数a,dを正の値の係数とし、係数b,cを負の値の係数として、満足度を算出してもよい。また、上記それぞれの値は、視聴している映像コンテンツのジャンルによっては、変化の値や割合が異なると考えられる。そのため、映像コンテンツのジャンルごとに係数a,b,c,dの重み付けを変更してもよい。さらに、上記それぞれの値は、視聴している映像コンテンツのジャンルやシーンごとの視聴者の感情によっては、視聴者が反応するまでタイムラグが発生することがあると考えられる。そのため、タイムラグを補正してそれぞれの値を取得してもよい。 In addition, as described above, when the viewers of the video content are intensively watching, the eye opening is widened, the number of blinks is reduced, the movement of the face is reduced, the pulse rate and the body temperature are reduced. Is empirically known to increase. Therefore, the degree of satisfaction may be calculated by using the coefficients a and d as positive coefficient and the coefficients b and c as negative coefficients. Further, it is considered that the value and the rate of change of each of the above values differ depending on the genre of the video content being viewed. Therefore, the weighting of the coefficients a, b, c, and d may be changed for each genre of the video content. Further, it is considered that a time lag may occur until the viewer reacts to each of the above values depending on the genre of the video content being viewed and the emotion of the viewer for each scene. Therefore, each value may be acquired by correcting the time lag.
 さらに、例えば、満足度算出モジュール1038は、上記の例による目の開き度合い、及びまばたきの回数の関数を加算して、集中度とし、集中度に対して所定の演算を行うことで満足度を算出してもよい。上記の例の場合、集中度Qは、以下の式
 Q=aJ+bK
 により算出され、満足度Pは、以下の式
 P=αQ+βR
 により算出される。ここで、Rは、顔の動き、及び生体情報の関数に対して所定の演算を行った関数または値でもよく、その他の値をパラメータとして算出した関数または値でもよい。
Further, for example, the satisfaction calculation module 1038 adds the functions of the degree of opening of the eyes and the number of blinks according to the above example to obtain the degree of concentration, and obtains the degree of satisfaction by performing a predetermined calculation on the degree of concentration. It may be calculated. In the case of the above example, the concentration Q is calculated by the following formula Q = aJ + bK.
Satisfaction P is calculated by the following formula P = αQ + βR
Is calculated by. Here, R may be a function or value obtained by performing a predetermined operation on the function of facial movement and biometric information, or may be a function or value calculated using other values as parameters.
 <1.3 音声出力装置30の機能的な構成>
 図6は、実施の形態1の満足度算出システム1を構成する音声出力装置30の機能的な構成を示す図である。図6に示すように、音声出力装置30は、通信部301と、検知部302と、記憶部303と、制御部304としての機能を発揮する。
<1.3 Functional configuration of audio output device 30>
FIG. 6 is a diagram showing a functional configuration of the audio output device 30 constituting the satisfaction calculation system 1 of the first embodiment. As shown in FIG. 6, the voice output device 30 exerts functions as a communication unit 301, a detection unit 302, a storage unit 303, and a control unit 304.
 通信部301は、音声出力装置30が外部の装置と通信するための処理を行う。 The communication unit 301 performs processing for the voice output device 30 to communicate with an external device.
 検知部302は、音声出力装置30が視聴者の顔の動きと生体情報とを取得するための処理を行う。上記のジャイロセンサと心電センサとにより構成されている。 The detection unit 302 performs processing for the voice output device 30 to acquire the movement of the viewer's face and biometric information. It is composed of the above-mentioned gyro sensor and electrocardiographic sensor.
 記憶部303は、音声出力装置30が使用するデータ及びプログラムを記憶する。記憶部303は、ジャイロセンサ検知信号データベース381と、生体センサ検知信号データベース382等を記憶する。 The storage unit 303 stores the data and the program used by the voice output device 30. The storage unit 303 stores the gyro sensor detection signal database 381, the biosensor detection signal database 382, and the like.
 ジャイロセンサ検知信号データベース381は、検知部302を構成するジャイロセンサの検知信号を時系列で保持するためのデータベースである。 The gyro sensor detection signal database 381 is a database for holding the detection signals of the gyro sensors constituting the detection unit 302 in chronological order.
 生体センサ検知信号データベース382は、検知部302を構成する心電センサの検知信号を時系列で保持するためのデータベースである。 The biosensor detection signal database 382 is a database for holding the detection signals of the electrocardiographic sensors constituting the detection unit 302 in chronological order.
 制御部304は、音声出力装置30のプロセッサがプログラムに従って処理を行うことにより、各種モジュールとして受信制御モジュール3031、送信制御モジュール3032、音声出力モジュール3033、ジャイロセンサ制御モジュール3034、生体センサ制御モジュール3035に示す機能を発揮する。 The control unit 304 performs processing according to a program by the processor of the voice output device 30, and causes the reception control module 3031, the transmission control module 3032, the voice output module 3033, the gyro sensor control module 3034, and the biosensor control module 3035 as various modules. Demonstrate the function shown.
 受信制御モジュール3031は、音声出力装置30が外部の装置から通信プロトコルに従って信号を受信する処理を制御する。 The reception control module 3031 controls a process in which the voice output device 30 receives a signal from an external device according to a communication protocol.
 送信制御モジュール3032は、音声出力装置30が外部の装置に対し通信プロトコルに従って信号を送信する処理を制御する。 The transmission control module 3032 controls a process in which the voice output device 30 transmits a signal to an external device according to a communication protocol.
 音声出力モジュール3033は、サーバ10から送信された音声データを、スピーカ34から映像コンテンツの音声として出力する処理を制御する。 The audio output module 3033 controls a process of outputting audio data transmitted from the server 10 as audio of video content from the speaker 34.
 ジャイロセンサ制御モジュール3034は、検知部302を構成するジャイロセンサの動作処理を制御する。また、ジャイロセンサ制御モジュール3034は、検知信号をサーバ10へ送信する。 The gyro sensor control module 3034 controls the operation processing of the gyro sensor constituting the detection unit 302. Further, the gyro sensor control module 3034 transmits a detection signal to the server 10.
 生体センサ制御モジュール3035は、検知部302を構成する心電センサの動作処理を制御する。また、生体センサ制御モジュール3035は、検知信号をサーバ10へ送信する。 The biosensor control module 3035 controls the operation processing of the electrocardiographic sensor constituting the detection unit 302. Further, the biosensor control module 3035 transmits a detection signal to the server 10.
 <2 データ構造>
 図7は、サーバ10が記憶する映像コンテンツデータベース181、視聴者データベース184のデータ構造を示す図である。
<2 data structure>
FIG. 7 is a diagram showing the data structures of the video content database 181 and the viewer database 184 stored in the server 10.
 図7に示すように、映像コンテンツデータベース181のレコードのそれぞれは、項目「施設ID」と、項目「施設名」と、項目「コンテンツ上映情報」等を含む。 As shown in FIG. 7, each record in the video content database 181 includes an item "facility ID", an item "facility name", an item "content screening information", and the like.
 項目「施設ID」は、満足度算出システム1が設置された施設それぞれを識別する情報である。1つの施設に映像出力装置20が複数設置され、それぞれ異なる映像コンテンツを上映する場合、異なる値の項目「施設ID」が付与され、別々に管理される。 The item "facility ID" is information that identifies each facility in which the satisfaction calculation system 1 is installed. When a plurality of video output devices 20 are installed in one facility and different video contents are screened, the item "facility ID" having a different value is assigned and managed separately.
 項目「施設名」は、満足度算出システム1が設置された施設の名称を示す情報である。 The item "facility name" is information indicating the name of the facility in which the satisfaction calculation system 1 is installed.
 項目「コンテンツ上映情報」は、満足度算出システム1を利用する施設で上映される映画等の映像コンテンツの上映スケジュールの情報であり、具体的には、項目「開始日時」と、項目「終了時刻」と、項目「コンテンツ名」等を含む。図7に示すように、1の項目「施設ID」に対して、1または複数の項目「コンテンツ上映情報」が時系列で格納されている。 The item "content screening information" is information on the screening schedule of video content such as a movie to be screened at a facility using the satisfaction calculation system 1. Specifically, the item "start date and time" and the item "end time". , And the item "content name" and the like. As shown in FIG. 7, one or a plurality of items "content screening information" are stored in chronological order for one item "facility ID".
 項目「開始日時」は、当該施設で上映される映像コンテンツの開始日及び開始時刻を示す情報である。 The item "start date and time" is information indicating the start date and start time of the video content to be screened at the facility.
 項目「終了時刻」は、当該施設で上映される映像コンテンツの終了時刻を示す情報である。 The item "end time" is information indicating the end time of the video content to be screened at the facility.
 項目「コンテンツ名」は、当該施設で上映される映像コンテンツの名称を示す情報である。 The item "content name" is information indicating the name of the video content to be screened at the facility.
 サーバ10は、施設の管理者から映像コンテンツの上映情報の入力を受け付けることに伴って、映像コンテンツデータベース181を更新する。 The server 10 updates the video content database 181 in response to receiving input of video content screening information from the facility manager.
 視聴者データベース184のレコードのそれぞれは、「視聴者ID」と、項目「視聴者名」と、項目「視聴コンテンツ」と、項目「満足度情報」等を含む。 Each of the records in the viewer database 184 includes a "viewer ID", an item "viewer name", an item "viewing content", an item "satisfaction information", and the like.
 項目「視聴者ID」は、満足度算出システム1を利用する施設で上映された映像コンテンツを視聴した視聴者それぞれを識別する情報である。 The item "viewer ID" is information that identifies each viewer who has viewed the video content screened at the facility that uses the satisfaction calculation system 1.
 項目「視聴者名」は、満足度算出システム1を利用する施設で上映された映像コンテンツを視聴した視聴者の氏名を示す情報である。なお、当該施設で視聴者の氏名情報が取得できなかった場合は、図7に示す項目「視聴者ID」が「#0202」で示すレコードのように、格納されなくてもよい。 The item "viewer name" is information indicating the name of the viewer who has viewed the video content screened at the facility using the satisfaction calculation system 1. If the viewer's name information cannot be obtained at the facility, the item "viewer ID" shown in FIG. 7 does not have to be stored as in the record shown by "# 0202".
 項目「視聴コンテンツ」は、当該視聴者が視聴した映像コンテンツの名称を示す情報であり、映像コンテンツデータベース281の項目「コンテンツ名」に対応している。 The item "viewing content" is information indicating the name of the video content viewed by the viewer, and corresponds to the item "content name" of the video content database 281.
 項目「満足度情報」は、満足度算出システム1を利用する施設で上映された映像コンテンツを視聴した視聴者の満足度の情報であり、具体的には、項目「経過時刻」と、項目「満足度」等を含む。図7に示すように、1の項目「視聴者ID」に対して、1または複数の項目「満足度情報」が所定時間(例えば、1分)ごとの時系列で格納されている。 The item "satisfaction information" is information on the satisfaction of the viewer who has viewed the video content screened at the facility using the satisfaction calculation system 1, and specifically, the item "elapsed time" and the item "elapsed time". Includes "satisfaction" and the like. As shown in FIG. 7, one or a plurality of items "satisfaction information" are stored in a time series for each predetermined time (for example, one minute) for one item "viewer ID".
 項目「経過時刻」は、当該映像コンテンツの開始からの所定時間ごとの経過時刻を示す情報である。 The item "elapsed time" is information indicating the elapsed time for each predetermined time from the start of the video content.
 項目「満足度」は、項目「経過時刻」の時刻における、当該映像コンテンツを視聴した視聴者の満足度を示す情報であり、満足度算出モジュール1038で算出された満足度の値である。 The item "satisfaction" is information indicating the satisfaction of the viewer who has viewed the video content at the time of the item "elapsed time", and is the value of the satisfaction calculated by the satisfaction calculation module 1038.
 サーバ10の満足度算出モジュール1038は、映像コンテンツを視聴した視聴者の満足度を算出するごとに、視聴者データベース184を更新する。 The satisfaction calculation module 1038 of the server 10 updates the viewer database 184 every time the satisfaction of the viewer who has viewed the video content is calculated.
 <3 動作>
 以下、図8及び図9を参照しながら、第1の実施の形態における満足度算出システム1による満足度算出の各処理について説明する。
<3 operation>
Hereinafter, each process of satisfaction calculation by the satisfaction calculation system 1 in the first embodiment will be described with reference to FIGS. 8 and 9.
 図8は、実施の形態1の満足度算出システム1による映像コンテンツ出力処理を行う流れの一例を示すフローチャートである。 FIG. 8 is a flowchart showing an example of a flow of performing video content output processing by the satisfaction calculation system 1 of the first embodiment.
 ステップS111において、サーバ10の映像コンテンツ出力制御モジュール1033は、映像コンテンツデータベース181を参照し、映像コンテンツを上映させる日時を管理する。映像コンテンツ出力制御モジュール1033は、映像コンテンツを上映するために送信すべき時間になると、ステップS112の処理へ進む。 In step S111, the video content output control module 1033 of the server 10 refers to the video content database 181 and manages the date and time when the video content is screened. The video content output control module 1033 proceeds to the process of step S112 when it is time to transmit the video content in order to screen it.
 ステップS112において、サーバ10の映像コンテンツ出力制御モジュール1033は、映像コンテンツを視聴者に視聴させるため、映像コンテンツの映像データを映像出力装置20へ送信し、音声データを音声出力装置30へ送信する。 In step S112, the video content output control module 1033 of the server 10 transmits the video data of the video content to the video output device 20 and the audio data to the audio output device 30 in order to allow the viewer to view the video content.
 ステップS122において、映像出力装置20は、サーバ10が送信する映像コンテンツの映像データを受け付ける。音声出力装置30は、サーバ10が送信する映像コンテンツの音声データを受け付ける。 In step S122, the video output device 20 receives the video data of the video content transmitted by the server 10. The audio output device 30 receives audio data of video content transmitted by the server 10.
 ステップS123において、映像出力装置20は、受け付けた映像コンテンツの映像データを映像として上映する。音声出力装置30は、受け付けた映像コンテンツの音声データを音声として出力する。 In step S123, the video output device 20 screens the video data of the received video content as a video. The audio output device 30 outputs the audio data of the received video content as audio.
 以上のように、満足度算出システム1は、所定の時間になると映像コンテンツの映像データを映像出力装置20へ送信し、音声データを音声出力装置30へ送信する。映像出力装置20は、映像データを映像として上映し、音声出力装置30は、音声データを音声として出力する。これにより、映像コンテンツの上映が行われ、視聴者は音声出力装置30を装着して視聴する。 As described above, the satisfaction calculation system 1 transmits the video data of the video content to the video output device 20 and the audio data to the audio output device 30 at a predetermined time. The video output device 20 screens video data as video, and the audio output device 30 outputs audio data as audio. As a result, the video content is screened, and the viewer wears the audio output device 30 for viewing.
 図9は、実施の形態1の満足度算出システム1による満足度算出処理を行う流れの一例を示すフローチャートである。 FIG. 9 is a flowchart showing an example of a flow of performing satisfaction calculation processing by the satisfaction calculation system 1 of the first embodiment.
 ステップS211において、音声出力装置30のジャイロセンサ制御モジュール3034は、検知部302を構成するジャイロセンサの動作制御を行い、ジャイロセンサの検知信号を、通信部301を介してサーバ10へ送信する。また、生体センサ制御モジュール3035は、検知部302を構成する生体センサの動作制御を行い、生体センサの検知信号を、通信部301を介してサーバ10へ送信する。 In step S211 the gyro sensor control module 3034 of the voice output device 30 controls the operation of the gyro sensors constituting the detection unit 302, and transmits the detection signal of the gyro sensor to the server 10 via the communication unit 301. Further, the biosensor control module 3035 controls the operation of the biosensors constituting the detection unit 302, and transmits the detection signal of the biosensor to the server 10 via the communication unit 301.
 ステップS231において、カメラ40は、撮影した画像データをサーバ10へ送信する。 In step S231, the camera 40 transmits the captured image data to the server 10.
 ステップS221において、サーバ10は、音声出力装置30から送信された検知信号を、通信部101を介して受け付ける。また、サーバ10は、カメラ40から送信された画像データを、通信部101を介して受け付ける。サーバ10は、受け付けた検知信号をセンサ検知信号データベース182に、受け付けた画像データをカメラ撮影画像データベース183に、それぞれ格納する。 In step S221, the server 10 receives the detection signal transmitted from the voice output device 30 via the communication unit 101. Further, the server 10 receives the image data transmitted from the camera 40 via the communication unit 101. The server 10 stores the received detection signal in the sensor detection signal database 182 and the received image data in the camera-captured image database 183.
 ステップS222において、サーバ10の開き度合い取得モジュール1034は、カメラ撮影画像データベース183に格納されている画像データを取得する。開き度合い取得モジュール1034は、画像データを解析して視聴者の目を認識して検知し、視聴者の目における目の開き度合いを取得する。 In step S222, the open degree acquisition module 1034 of the server 10 acquires the image data stored in the camera-captured image database 183. The opening degree acquisition module 1034 analyzes the image data, recognizes and detects the eyes of the viewer, and acquires the degree of opening of the eyes in the eyes of the viewer.
 ステップS223において、サーバ10のまばたき回数取得モジュール1035は、カメラ撮影画像データベース183に格納されている画像データを取得する。まばたき回数取得モジュール1035は、画像データを解析して視聴者の目を認識して検知し、視聴者のまばたきの回数を取得する。 In step S223, the blink count acquisition module 1035 of the server 10 acquires the image data stored in the camera-captured image database 183. The blink count acquisition module 1035 analyzes the image data, recognizes and detects the viewer's eyes, and acquires the number of blinks of the viewer.
 ステップS224において、サーバ10の顔動き取得モジュール1036は、センサ検知信号データベース182に格納されている検知信号を取得する。顔動き取得モジュール1036は、ジャイロセンサの検知信号を解析し、検知信号の角速度とその動きの時間から視聴者の顔の動きを取得する。 In step S224, the face movement acquisition module 1036 of the server 10 acquires the detection signal stored in the sensor detection signal database 182. The face movement acquisition module 1036 analyzes the detection signal of the gyro sensor and acquires the movement of the viewer's face from the angular velocity of the detection signal and the time of the movement.
 ステップS225において、サーバ10の生体情報取得モジュール1037は、センサ検知信号データベース182に格納されている検知信号を取得する。生体情報取得モジュール1037は、心電センサの検知信号を解析し、視聴者の生体情報(視聴者の脈拍や体温)を取得する。 In step S225, the biometric information acquisition module 1037 of the server 10 acquires the detection signal stored in the sensor detection signal database 182. The biological information acquisition module 1037 analyzes the detection signal of the electrocardiographic sensor and acquires the biological information of the viewer (the pulse and body temperature of the viewer).
 ステップS226において、サーバ10の満足度算出モジュール1038は、ステップS222で取得した目の開き度合い、ステップS223で取得したまばたきの回数、ステップS224で取得した顔の動き、ステップS225で取得した生体情報に基づき、視聴者の映像コンテンツに対する満足度を算出する。 In step S226, the satisfaction calculation module 1038 of the server 10 is based on the degree of eye opening acquired in step S222, the number of blinks acquired in step S223, the facial movement acquired in step S224, and the biometric information acquired in step S225. Based on this, the degree of satisfaction of the viewer with the video content is calculated.
 ステップS227において、サーバ10の満足度算出モジュール1038は、算出した満足度を視聴者データベース184に格納し、データベースを更新する。 In step S227, the satisfaction calculation module 1038 of the server 10 stores the calculated satisfaction in the viewer database 184 and updates the database.
 以上のように、満足度算出システム1のサーバ10は、視聴者を撮影した画像データから目の開き度合い、及びまばたきの回数を取得し、センサが検知した検知信号から顔の動き、及び生体情報を取得する。これらの情報から、映像コンテンツに対する満足度を算出する。これにより、映像コンテンツに対する評価を分析することが可能になる。 As described above, the server 10 of the satisfaction calculation system 1 acquires the degree of eye opening and the number of blinks from the image data of the viewer, and the face movement and biological information from the detection signal detected by the sensor. To get. From this information, the degree of satisfaction with the video content is calculated. This makes it possible to analyze the evaluation of the video content.
 <4 画面例>
 以下、図10を参照しながら、満足度算出システム1による各パラメータの推移を表示した画面例について説明する。
<4 screen example>
Hereinafter, a screen example in which the transition of each parameter by the satisfaction calculation system 1 is displayed will be described with reference to FIG.
 図10は、満足度算出処理における各パラメータの推移を表示したサーバ10の画面例を示す図である。図10の画面例は、サーバ10の入出力IF13(ディスプレイ)に、目の開き度合い、まばたきの回数、顔の動き、及び生体情報(脈拍)の時系列の推移を表示した画面例を示す。図9のステップS226による算出結果に相当する。 FIG. 10 is a diagram showing a screen example of the server 10 displaying the transition of each parameter in the satisfaction calculation process. The screen example of FIG. 10 shows a screen example in which the degree of eye opening, the number of blinks, the movement of the face, and the time-series transition of biological information (pulse) are displayed on the input / output IF13 (display) of the server 10. It corresponds to the calculation result in step S226 of FIG.
 図10に示すように、サーバ10の入出力IF13(ディスプレイ)には、視聴者の目の開き度合い、まばたきの回数、顔の動き、及び生体情報(脈拍)の時系列の推移を示すグラフ1031aが例として表示されている。このグラフ1031aは、映像コンテンツの開始時点の目の開き度合い、まばたきの回数、顔の動き、及び脈拍を100とし、その後の経過による各パラメータの変化率を算出し、グラフ化したものである。また、映像コンテンツの上映時間は2時間である。 As shown in FIG. 10, on the input / output IF13 (display) of the server 10, a graph 1031a showing the degree of opening of the eyes of the viewer, the number of blinks, the movement of the face, and the time-series transition of biological information (pulse). Is displayed as an example. In this graph 1031a, the degree of eye opening, the number of blinks, the movement of the face, and the pulse at the start of the video content are set to 100, and the rate of change of each parameter with the subsequent passage is calculated and graphed. The screening time of the video content is 2 hours.
 図10のグラフ1031aを参照すると、目の開き度合い、及び脈拍は、映像コンテンツの開始以降減少を続け、開始後30分の時点で上昇に反転し、開始後90分の時点をピークに減少を続け、映像コンテンツの終了時点で100付近の値になっている。また、まばたきの回数、及び顔の動きは、映像コンテンツの開始以降上昇を続け、開始後30分の時点をピークに減少に反転し、開始後90分の時点で上昇に反転し、映像コンテンツの終了時点で100付近の値になっている。 Referring to the graph 1031a of FIG. 10, the degree of eye opening and the pulse continue to decrease after the start of the video content, reverse to increase at 30 minutes after the start, and decrease after peaking at 90 minutes after the start. Subsequently, the value is around 100 at the end of the video content. In addition, the number of blinks and the movement of the face continue to increase after the start of the video content, peak at 30 minutes after the start and reverse to decrease, and then reverse to increase at 90 minutes after the start of the video content. At the end, the value is around 100.
 上記のように、映像コンテンツを視聴している者が集中して視聴している場合、目の開きが大きくなり、まばたきの回数が減少し、顔の動きが少なくなり、脈拍数が増加する傾向があることから、図10に示す例の場合、開始後30分の時点が最も集中していない状態であり、開始後90分の時点が最も集中している状態であることが分かる。そのため、満足度を算出する際の係数は、目の開き度合い、及び脈拍が大きい場合に満足度が高くなるように(正の係数)設定し、まばたきの回数、及び顔の動きが小さい場合に満足度が高くなるように(負の係数)設定する。 As mentioned above, when the people who are watching the video content are watching intensively, the opening of the eyes is widened, the number of blinks is reduced, the movement of the face is reduced, and the pulse rate tends to increase. Therefore, in the case of the example shown in FIG. 10, it can be seen that the time point 30 minutes after the start is the least concentrated state and the time point 90 minutes after the start is the most concentrated state. Therefore, the coefficient for calculating the degree of satisfaction is set so that the degree of eye opening and the degree of satisfaction are high when the pulse is large (positive coefficient), and when the number of blinks and the movement of the face are small. Set so that the satisfaction level is high (negative coefficient).
 <小括>
 以上のように、本実施形態によると、視聴者を撮影した画像データを解析し、目の開き度合い、及びまばたきの回数を取得する。センサが検知した検知信号から、顔の動き、及び生体情報を取得する。これらの情報に基づき、映像コンテンツに対する満足度を算出する。そのため、映像コンテンツに対する視聴者の反応を示す情報を取得することが可能である。これにより、映像コンテンツに対する評価を分析することが可能になる。
<Summary>
As described above, according to the present embodiment, the image data obtained by capturing the viewer is analyzed, and the degree of eye opening and the number of blinks are acquired. Face movement and biometric information are acquired from the detection signal detected by the sensor. Based on this information, the degree of satisfaction with the video content is calculated. Therefore, it is possible to acquire information indicating the reaction of the viewer to the video content. This makes it possible to analyze the evaluation of the video content.
 また、目の開き度合い、まばたきの回数、顔の動き、及び生体情報を所定時間ごとに取得し、所定時間ごとに満足度を算出する。これにより、映像コンテンツの時間経過ごとに細かく視聴者の反応を示す情報を取得することが可能である。これにより、映像コンテンツのシーンごとの評価を分析することが可能になり、例えば、プロモーション等に使用される重要なシーンの抽出を容易に行うことが可能になる。 In addition, the degree of eye opening, the number of blinks, the movement of the face, and the biological information are acquired at predetermined time intervals, and the satisfaction level is calculated at predetermined time intervals. As a result, it is possible to acquire detailed information indicating the reaction of the viewer with each passage of time of the video content. This makes it possible to analyze the evaluation of video content for each scene, and for example, it becomes possible to easily extract important scenes used for promotion and the like.
 以上、開示に係る実施形態について説明したが、これらはその他の様々な形態で実施することが可能であり、種々の省略、置換及び変更を行なって実施することができる。これらの実施形態及び変形例ならびに省略、置換及び変更を行なったものは、特許請求の範囲の技術的範囲とその均等の範囲に含まれる。 Although the embodiments related to the disclosure have been described above, these can be implemented in various other embodiments, and can be implemented by making various omissions, substitutions, and changes. These embodiments and modifications, as well as those omitted, replaced or modified, are included in the technical scope of the claims and the equivalent scope thereof.
 <付記>
 以上の各実施形態で説明した事項を、以下に付記する。
<Additional notes>
The matters described in each of the above embodiments will be added below.
 (付記1)制御部を備える情報処理装置であって、制御部は、コンテンツを視聴する視聴者を撮影した画像データから、視聴者の目を検知し、視聴者の目の開き度合いを取得する開き度合い取得部と、画像データから、視聴者の目を検知し、視聴者のまばたきの回数を取得するまばたき回数取得部と、開き度合い、及びまばたきの回数をパラメータとして、視聴者の映像コンテンツに対する満足度を算出する算出部と、を備える情報処理装置。 (Appendix 1) An information processing device including a control unit, in which the control unit detects the viewer's eyes from image data obtained by capturing a viewer viewing the content and acquires the degree of opening of the viewer's eyes. The opening degree acquisition unit, the blinking count acquisition unit that detects the viewer's eyes from the image data and acquires the number of blinks of the viewer, and the opening degree and the number of blinks as parameters for the viewer's video content. An information processing device including a calculation unit for calculating satisfaction.
 (付記2)制御部は、さらに、画像データから、または視聴者が装着する音声出力装置に設けられたセンサから出力された検知信号から、視聴者の顔の動きを取得する動き取得部を備え、算出部は、顔の動きをパラメータとして、視聴者の映像コンテンツに対する満足度を算出する、(付記1)に記載の情報処理装置。 (Appendix 2) The control unit further includes a motion acquisition unit that acquires the movement of the viewer's face from image data or from a detection signal output from a sensor provided in the audio output device worn by the viewer. The information processing device according to (Appendix 1), wherein the calculation unit calculates the degree of satisfaction of the viewer with the video content using the movement of the face as a parameter.
 (付記3)動き取得部は、視聴者が自己の耳に装着する音声出力装置に設けられたセンサから出力された検知信号から、視聴者の顔の動きを取得する、(付記2)に記載の情報処理装置。 (Appendix 3) The motion acquisition unit acquires the motion of the viewer's face from the detection signal output from the sensor provided in the audio output device worn by the viewer on his / her ear, according to (Appendix 2). Information processing device.
 (付記4)動き取得部は、音声出力装置に設けられたジャイロセンサから出力された検知信号から、視聴者の顔を傾けた動き、または視聴者の顔の向きを移動させた動きを取得する、(付記2)または(付記3)に記載の情報処理装置。 (Appendix 4) The motion acquisition unit acquires a motion in which the viewer's face is tilted or a motion in which the viewer's face is moved from the detection signal output from the gyro sensor provided in the voice output device. , (Appendix 2) or (Appendix 3).
 (付記5)制御部は、さらに、視聴者が装着する音声出力装置に設けられたセンサから出力された検知信号から、視聴者の生体情報を取得する生体情報取得部を備え、算出部は、生体情報をパラメータとして、視聴者の映像コンテンツに対する満足度を算出する、(付記1)から(付記4)のいずれかに記載の情報処理装置。 (Appendix 5) The control unit further includes a biometric information acquisition unit that acquires the biometric information of the viewer from the detection signal output from the sensor provided in the audio output device worn by the viewer, and the calculation unit includes a biometric information acquisition unit. The information processing device according to any one of (Appendix 1) to (Appendix 4), which calculates the degree of satisfaction of the viewer with the video content using biological information as a parameter.
 (付記6)生体情報取得部は、視聴者が装着する音声出力装置に設けられた脈拍センサから出力された検知信号から、視聴者の脈拍または血圧の情報を生体情報として取得する、(付記5)に記載の情報処理装置。 (Appendix 6) The biometric information acquisition unit acquires the viewer's pulse or blood pressure information as biometric information from the detection signal output from the pulse sensor provided in the audio output device worn by the viewer (Appendix 5). ).
 (付記7)生体情報取得部は、視聴者が装着する音声出力装置に設けられた温度センサから出力された検知信号から、視聴者の体温の情報を生体情報として取得する、(付記5)または(付記6)に記載の情報処理装置。 (Appendix 7) The biometric information acquisition unit acquires the viewer's body temperature information as biometric information from the detection signal output from the temperature sensor provided in the audio output device worn by the viewer (Appendix 5) or. The information processing apparatus according to (Appendix 6).
 (付記8)算出部は、パラメータのそれぞれの値の所定の時刻における基準値と比較した差分、基準値との変化率、所定の時間内における変化の値、所定の時間内における変化率のいずれか1つまたは複数に基づき、満足度を算出する、(付記1)から(付記7)のいずれかに記載の情報処理装置。 (Appendix 8) The calculation unit determines whether the difference between each value of the parameter and the reference value at a predetermined time, the rate of change from the reference value, the value of change within a predetermined time, or the rate of change within a predetermined time. The information processing apparatus according to any one of (Appendix 1) to (Appendix 7), which calculates satisfaction based on one or more of them.
 (付記9)算出部は、パラメータのそれぞれの値に対して所定の関数による演算を行い、演算後のそれぞれの値を加算して満足度を算出する、(付記1)から(付記8)のいずれかに記載の情報処理装置。 (Appendix 9) The calculation unit performs an operation by a predetermined function for each value of the parameter, and adds each value after the operation to calculate the satisfaction level, from (Appendix 1) to (Appendix 8). The information processing device according to any one.
 (付記10)算出部は、パラメータまたは演算後のそれぞれの値に対して、それぞれ所定の係数を乗算し、乗算後のそれぞれの値を加算して満足度を算出する、(付記9)に記載の情報処理装置。 (Appendix 10) The calculation unit calculates satisfaction by multiplying each value after the parameter or calculation by a predetermined coefficient and adding each value after the multiplication. Information processing equipment.
 (付記11)算出部は、パラメータまたは演算後のそれぞれの所定の値に対して、正の値の係数を乗算し、パラメータまたは演算後のそれぞれの所定の値に対して、負の値の係数を乗算し、乗算後のそれぞれの値を加算する、(付記10)に記載の情報処理装置。 (Appendix 11) The calculation unit multiplies each predetermined value after the parameter or calculation by a coefficient of a positive value, and the coefficient of a negative value with respect to each predetermined value after the parameter or calculation. The information processing apparatus according to (Appendix 10), wherein is multiplied by and the respective values after multiplication are added.
 (付記12)算出部は、1または複数の視聴者をそれぞれ識別し、それぞれの視聴者ごとに満足度を算出する、(付記1)から(付記11)のいずれかに記載の情報処理装置。 (Appendix 12) The information processing device according to any one of (Appendix 1) to (Appendix 11), wherein the calculation unit identifies one or a plurality of viewers, and calculates satisfaction for each viewer.
 (付記13)開き度合い取得部は、視聴者の目を開いた部分の面積を算出し、視聴者の目の開き度合いとして取得する、(付記1)から(付記12)のいずれかに記載の情報処理装置。 (Appendix 13) The opening degree acquisition unit calculates the area of the portion where the viewer's eyes are opened and acquires it as the viewer's eye opening degree, according to any one of (Appendix 1) to (Appendix 12). Information processing device.
 (付記14)開き度合い取得部は、視聴者の目を開いた部分における、上瞼と下瞼との接点と、上瞼の上方の端点、または虹彩と上瞼との接点と、下瞼の下方の端点、または虹彩と下瞼との接点と、の位置の情報を算出し、位置の情報に基づき、視聴者の目の開き度合いとして取得する、(付記1)から(付記12)のいずれかに記載の情報処理装置。 (Appendix 14) The degree of opening acquisition unit is the contact point between the upper eyelid and the lower eyelid and the upper end point of the upper eyelid, or the contact point between the iris and the upper eyelid, and the lower eyelid in the portion where the viewer's eyes are opened. Any of (Appendix 1) to (Appendix 12) that calculates the position information of the lower end point or the contact point between the iris and the lower eyelid and acquires it as the degree of opening of the viewer's eyes based on the position information. Information processing device described in the eyelids.
 (付記15)まばたき回数取得部は、視聴者を撮影した画像データの所定時間における視聴者のまばたきの回数を取得する、(付記1)から(付記14)のいずれかに記載の情報処理装置。 (Appendix 15) The information processing device according to any one of (Appendix 1) to (Appendix 14), wherein the blink count acquisition unit acquires the number of blinks of the viewer in a predetermined time of the image data obtained by photographing the viewer.
 (付記16)まばたき回数取得部は、視聴者の目が閉じた状態から、所定の時間内に視聴者の目が開いた状態になったとき、視聴者のまばたきと認識して回数をカウントする、(付記1)から(付記15)のいずれかに記載の情報処理装置。 (Appendix 16) When the viewer's eyes are opened within a predetermined time from the state in which the viewer's eyes are closed, the blink count acquisition unit recognizes the viewer's blink and counts the number of times. , (Appendix 1) to (Appendix 15).
 (付記17)カメラと、情報処理装置と、を備えるシステムであって、カメラは、映像コンテンツを視聴する視聴者を撮影し、画像データを生成して情報処理装置に取得させ、情報処理装置は、制御部を備え、制御部は、映像コンテンツを視聴する視聴者を撮影した画像データから、視聴者の目を検知し、視聴者の目の開き度合いを取得する開き度合い取得部と、画像データから、視聴者の目を検知し、視聴者のまばたきの回数を取得するまばたき回数取得部と、開き度合い、及びまばたきの回数をパラメータとして、視聴者の映像コンテンツに対する満足度を算出する算出部と、を備える、システム。 (Appendix 17) A system including a camera and an information processing device. The camera photographs a viewer who views video content, generates image data, and causes the information processing device to acquire the image data. , The control unit includes an opening degree acquisition unit that detects the eyes of the viewer from the image data of the viewer viewing the video content and acquires the opening degree of the viewer's eyes, and image data. From, a blinking count acquisition unit that detects the viewer's eyes and acquires the number of blinks of the viewer, and a calculation unit that calculates the satisfaction level of the viewer with the video content using the degree of opening and the number of blinks as parameters. , A system.
 (付記18)プロセッサを備えるコンピュータにより実行されるための方法であって、方法は、プロセッサが、映像コンテンツを視聴する視聴者を撮影した画像データから、視聴者の目を検知し、視聴者の目の開き度合いを取得するステップと、画像データから、視聴者の目を検知し、視聴者のまばたきの回数を取得するステップと、開き度合い、及びまばたきの回数をパラメータとして、視聴者の映像コンテンツに対する満足度を算出するステップと、を実行する、方法。 (Appendix 18) A method for being executed by a computer equipped with a processor, wherein the processor detects the eyes of the viewer from the image data of the viewer who views the video content, and the viewer's eyes are detected. The step of acquiring the degree of eye opening, the step of detecting the viewer's eyes from the image data and acquiring the number of blinks of the viewer, and the degree of opening and the number of blinks as parameters of the viewer's video content. Steps to calculate satisfaction with, and how to perform.
 (付記19)プロセッサを備えるコンピュータに実行させるためのプログラムであって、プログラムは、プロセッサに、映像コンテンツを視聴する視聴者を撮影した画像データから、視聴者の目を検知し、視聴者の目の開き度合いを取得するステップと、画像データから、視聴者の目を検知し、視聴者のまばたきの回数を取得するステップと開き度合い、及びまばたきの回数をパラメータとして、視聴者の映像コンテンツに対する満足度を算出するステップと、を実行させる、プログラム。 (Appendix 19) A program for causing a computer equipped with a processor to execute the program, in which the program detects the eyes of the viewer from the image data obtained by capturing the viewer viewing the video content, and the eyes of the viewer are detected. Satisfaction with the viewer's video content with the steps of acquiring the degree of opening and the step of detecting the viewer's eyes from the image data and acquiring the number of blinks of the viewer, the degree of opening, and the number of blinks as parameters. A program that executes the steps to calculate the degree.
10 サーバ、20 映像出力装置、30 音声出力装置、40 カメラ、80 ネットワーク、12 通信IF、13 入出力IF、15 メモリ、16 ストレージ、19 プロセッサ、33 センサ部、101 通信部、102 記憶部、181 映像コンテンツデータベース、182 センサ検知信号データベース、183 カメラ撮影画像データベース、184 視聴者データベース、103 制御部、1033 映像コンテンツ出力制御モジュール、1034 開き度合い取得モジュール(開き度合い取得部)、1035 まばたき回数取得モジュール(まばたき回数取得部)、1036 顔動き取得モジュール(動き取得部)、1037 生体情報取得モジュール(生体情報取得部)、1038 満足度算出モジュール(算出部)

 
10 server, 20 video output device, 30 audio output device, 40 camera, 80 network, 12 communication IF, 13 input / output IF, 15 memory, 16 storage, 19 processor, 33 sensor unit, 101 communication unit, 102 storage unit, 181 Video content database, 182 sensor detection signal database, 183 camera shot image database, 184 viewer database, 103 control unit, 1033 video content output control module, 1034 opening degree acquisition module (opening degree acquisition unit), 1035 blink count acquisition module ( Blink count acquisition unit), 1036 Face movement acquisition module (movement acquisition unit), 1037 Biometric information acquisition module (biological information acquisition unit), 1038 Satisfaction calculation module (calculation unit)

Claims (19)

  1.  制御部を備える情報処理装置であって、
     前記制御部は、
     映像コンテンツを視聴する視聴者を撮影した画像データから、前記視聴者の目を検知し、前記視聴者の目の開き度合いを取得する開き度合い取得部と、
     前記画像データから、前記視聴者の目を検知し、前記視聴者のまばたきの回数を取得するまばたき回数取得部と、
     前記開き度合い、及び前記まばたきの回数をパラメータとして、前記視聴者の前記映像コンテンツに対する満足度を算出する算出部と、を備える情報処理装置。
    An information processing device equipped with a control unit
    The control unit
    An opening degree acquisition unit that detects the eyes of the viewer and acquires the degree of opening of the viewer's eyes from image data obtained by capturing a viewer viewing the video content.
    From the image data, a blink count acquisition unit that detects the eyes of the viewer and acquires the number of blinks of the viewer.
    An information processing device including a calculation unit that calculates the degree of satisfaction of the viewer with the video content by using the degree of opening and the number of blinks as parameters.
  2.  前記制御部は、さらに、
     前記画像データから、または前記視聴者が装着する音声出力装置に設けられたセンサから出力された検知信号から、前記視聴者の顔の動きを取得する動き取得部を備え、
     前記算出部は、前記顔の動きをパラメータとして、前記視聴者の前記映像コンテンツに対する満足度を算出する、請求項1に記載の情報処理装置。
    The control unit further
    It is provided with a motion acquisition unit that acquires the motion of the viewer's face from the image data or from a detection signal output from a sensor provided in the audio output device worn by the viewer.
    The information processing device according to claim 1, wherein the calculation unit calculates the satisfaction level of the viewer with respect to the video content using the movement of the face as a parameter.
  3.  前記動き取得部は、前記視聴者が自己の耳に装着する音声出力装置に設けられたセンサから出力された検知信号から、前記視聴者の顔の動きを取得する、請求項2に記載の情報処理装置。 The information according to claim 2, wherein the motion acquisition unit acquires the motion of the viewer's face from a detection signal output from a sensor provided in an audio output device worn by the viewer on his / her ear. Processing equipment.
  4.  前記動き取得部は、前記音声出力装置に設けられたジャイロセンサから出力された検知信号から、前記視聴者の顔を傾けた動き、または前記視聴者の顔の向きを移動させた動きを取得する、請求項2または請求項3に記載の情報処理装置。 The motion acquisition unit acquires a motion in which the viewer's face is tilted or a motion in which the viewer's face is moved from a detection signal output from a gyro sensor provided in the voice output device. , The information processing apparatus according to claim 2 or 3.
  5.  前記制御部は、さらに、
     前記視聴者が装着する音声出力装置に設けられたセンサから出力された検知信号から、前記視聴者の生体情報を取得する生体情報取得部を備え、
     前記算出部は、前記生体情報をパラメータとして、前記視聴者の前記映像コンテンツに対する満足度を算出する、請求項1から請求項4のいずれか1項に記載の情報処理装置。
    The control unit further
    It is provided with a biometric information acquisition unit that acquires biometric information of the viewer from a detection signal output from a sensor provided in the audio output device worn by the viewer.
    The information processing device according to any one of claims 1 to 4, wherein the calculation unit calculates the satisfaction level of the viewer with respect to the video content using the biometric information as a parameter.
  6.  前記生体情報取得部は、前記視聴者が装着する音声出力装置に設けられた脈拍センサから出力された検知信号から、前記視聴者の脈拍または血圧の情報を前記生体情報として取得する、請求項5に記載の情報処理装置。 5. The biometric information acquisition unit acquires information on the viewer's pulse or blood pressure as the biometric information from a detection signal output from a pulse sensor provided in the audio output device worn by the viewer. The information processing device described in.
  7.  前記生体情報取得部は、前記視聴者が装着する音声出力装置に設けられた温度センサから出力された検知信号から、前記視聴者の体温の情報を前記生体情報として取得する、請求項5または請求項6に記載の情報処理装置。 5. The biometric information acquisition unit acquires information on the body temperature of the viewer as the biometric information from a detection signal output from a temperature sensor provided in the audio output device worn by the viewer, claim 5 or the above. Item 6. The information processing apparatus according to item 6.
  8.  前記算出部は、前記パラメータのそれぞれの値の所定の時刻における基準値と比較した差分、前記基準値との変化率、所定の時間内における変化の値、所定の時間内における変化率のいずれか1つまたは複数に基づき、前記満足度を算出する、請求項1から請求項7のいずれか1項に記載の情報処理装置。 The calculation unit is any one of a difference between each value of the parameter and a reference value at a predetermined time, a rate of change from the reference value, a value of change within a predetermined time, and a rate of change within a predetermined time. The information processing apparatus according to any one of claims 1 to 7, which calculates the satisfaction level based on one or more.
  9.  前記算出部は、前記パラメータのそれぞれの値に対して所定の関数による演算を行い、演算後のそれぞれの値を加算して前記満足度を算出する、請求項1から請求項8のいずれか1項に記載の情報処理装置。 The calculation unit performs an operation by a predetermined function on each value of the parameter, and adds each value after the operation to calculate the satisfaction level. Any one of claims 1 to 8. The information processing device described in the section.
  10.  前記算出部は、前記パラメータまたは演算後のそれぞれの値に対して、それぞれ所定の係数を乗算し、乗算後のそれぞれの値を加算して前記満足度を算出する、請求項9に記載の情報処理装置。 The information according to claim 9, wherein the calculation unit multiplies each of the parameters or each value after calculation by a predetermined coefficient, and adds each value after multiplication to calculate the satisfaction. Processing equipment.
  11.  前記算出部は、
     前記パラメータまたは演算後のそれぞれの所定の値に対して、正の値の係数を乗算し、
     前記パラメータまたは演算後のそれぞれの所定の値に対して、負の値の係数を乗算し、乗算後のそれぞれの値を加算する、請求項10に記載の情報処理装置。
    The calculation unit
    Multiply each predetermined value of the parameter or after calculation by a coefficient of a positive value.
    The information processing apparatus according to claim 10, wherein a coefficient of a negative value is multiplied by each predetermined value after the parameter or calculation, and each value after multiplication is added.
  12.  前記算出部は、1または複数の視聴者をそれぞれ識別し、それぞれの視聴者ごとに前記満足度を算出する、請求項1から請求項11のいずれか1項に記載の情報処理装置。 The information processing device according to any one of claims 1 to 11, wherein the calculation unit identifies one or a plurality of viewers, and calculates the satisfaction level for each viewer.
  13.  前記開き度合い取得部は、前記視聴者の目を開いた部分の面積を算出し、前記視聴者の目の開き度合いとして取得する、請求項1から請求項12のいずれか1項に記載の情報処理装置。 The information according to any one of claims 1 to 12, wherein the opening degree acquisition unit calculates the area of the portion where the viewer's eyes are opened and acquires the area as the viewer's eye opening degree. Processing equipment.
  14.  前記開き度合い取得部は、前記視聴者の目を開いた部分における、上瞼と下瞼との接点と、上瞼の上方の端点、または虹彩と上瞼との接点と、下瞼の下方の端点、または虹彩と下瞼との接点と、の位置の情報を算出し、前記位置の情報に基づき、前記視聴者の目の開き度合いとして取得する、請求項1から請求項12のいずれか1項に記載の情報処理装置。 The degree of opening acquisition unit is the contact point between the upper eyelid and the lower eyelid and the upper end point of the upper eyelid, or the contact point between the iris and the upper eyelid, and the lower part of the lower eyelid in the portion where the viewer's eyes are opened. Any one of claims 1 to 12, which calculates the position information of the end point or the contact point between the iris and the lower eyelid and acquires it as the degree of eye opening of the viewer based on the position information. The information processing device described in the section.
  15.  前記まばたき回数取得部は、前記視聴者を撮影した画像データの所定時間における前記視聴者のまばたきの回数を取得する、請求項1から請求項14のいずれか1項に記載の情報処理装置。 The information processing device according to any one of claims 1 to 14, wherein the blink number acquisition unit acquires the number of blinks of the viewer in a predetermined time of image data obtained by photographing the viewer.
  16.  前記まばたき回数取得部は、前記視聴者の目が閉じた状態から、所定の時間内に前記視聴者の目が開いた状態になったとき、前記視聴者のまばたきと認識して回数をカウントする、請求項1から請求項15のいずれか1項に記載の情報処理装置。 When the viewer's eyes are opened within a predetermined time from the state in which the viewer's eyes are closed, the blink count acquisition unit recognizes the viewer's blinks and counts the number of times. , The information processing apparatus according to any one of claims 1 to 15.
  17.  カメラと、情報処理装置と、を備えるシステムであって、
     前記カメラは、映像コンテンツを視聴する視聴者を撮影し、画像データを生成して前記情報処理装置に取得させ、
     前記情報処理装置は、制御部を備え、
      前記制御部は、
      映像コンテンツを視聴する視聴者を撮影した画像データから、前記視聴者の目を検知し、前記視聴者の目の開き度合いを取得する開き度合い取得部と、
      前記画像データから、前記視聴者の目を検知し、前記視聴者のまばたきの回数を取得するまばたき回数取得部と、
      前記開き度合い、及び前記まばたきの回数をパラメータとして、前記視聴者の前記映像コンテンツに対する満足度を算出する算出部と、を備える、システム。
    A system equipped with a camera and an information processing device.
    The camera captures a viewer viewing the video content, generates image data, and causes the information processing device to acquire the image data.
    The information processing device includes a control unit.
    The control unit
    An opening degree acquisition unit that detects the eyes of the viewer and acquires the degree of opening of the viewer's eyes from image data obtained by capturing a viewer viewing the video content.
    From the image data, a blink count acquisition unit that detects the eyes of the viewer and acquires the number of blinks of the viewer.
    A system including a calculation unit that calculates the degree of satisfaction of the viewer with the video content by using the degree of opening and the number of blinks as parameters.
  18.  プロセッサを備えるコンピュータにより実行されるための方法であって、
     前記方法は、前記プロセッサが、
     映像コンテンツを視聴する視聴者を撮影した画像データから、前記視聴者の目を検知し、前記視聴者の目の開き度合いを取得するステップと、
     前記画像データから、前記視聴者の目を検知し、前記視聴者のまばたきの回数を取得するステップと、
     前記開き度合い、及び前記まばたきの回数をパラメータとして、前記視聴者の前記映像コンテンツに対する満足度を算出するステップと、を実行する、方法。
    A method for being executed by a computer with a processor,
    In the method, the processor
    A step of detecting the eyes of the viewer from image data obtained by capturing a viewer viewing the video content and acquiring the degree of opening of the eyes of the viewer.
    A step of detecting the eyes of the viewer from the image data and acquiring the number of blinks of the viewer, and
    A method of executing a step of calculating the degree of satisfaction of the viewer with the video content using the degree of opening and the number of blinks as parameters.
  19.  プロセッサを備えるコンピュータに実行させるためのプログラムであって、
     前記プログラムは、前記プロセッサに、
     映像コンテンツを視聴する視聴者を撮影した画像データから、前記視聴者の目を検知し、前記視聴者の目の開き度合いを取得するステップと、
     前記画像データから、前記視聴者の目を検知し、前記視聴者のまばたきの回数を取得するステップと、
     前記開き度合い、及び前記まばたきの回数をパラメータとして、前記視聴者の前記映像コンテンツに対する満足度を算出するステップと、を実行させる、プログラム。
    A program to be executed by a computer equipped with a processor.
    The program is delivered to the processor.
    A step of detecting the eyes of the viewer from image data obtained by capturing a viewer viewing the video content and acquiring the degree of opening of the eyes of the viewer.
    A step of detecting the eyes of the viewer from the image data and acquiring the number of blinks of the viewer, and
    A program for executing a step of calculating the degree of satisfaction of the viewer with the video content, using the degree of opening and the number of blinks as parameters.
PCT/JP2021/015615 2020-04-16 2021-04-15 Information processing device, system, method, and program WO2021210652A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2020073382A JP6856959B1 (en) 2020-04-16 2020-04-16 Information processing equipment, systems, methods and programs
JP2020-073382 2020-04-16

Publications (1)

Publication Number Publication Date
WO2021210652A1 true WO2021210652A1 (en) 2021-10-21

Family

ID=75377961

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2021/015615 WO2021210652A1 (en) 2020-04-16 2021-04-15 Information processing device, system, method, and program

Country Status (2)

Country Link
JP (2) JP6856959B1 (en)
WO (1) WO2021210652A1 (en)

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002057765A (en) * 2000-08-07 2002-02-22 Matsushita Electric Ind Co Ltd Headset for wireless telephone
JP2002344904A (en) * 2001-02-06 2002-11-29 Sony Corp Device for reproducing content, device for receiving content, method for controlling content display, method for collecting and analyzing content evaluation, device for collecting and analyzing content evaluation, method for managing collection of content evaluation, and device for managing collection of content evaluation
JP2007003618A (en) * 2005-06-21 2007-01-11 Sharp Corp Display device and mobile terminal device
JP2009159073A (en) * 2007-12-25 2009-07-16 Panasonic Corp Acoustic playback apparatus and acoustic playback method
JP2010520554A (en) * 2007-03-06 2010-06-10 エムセンス コーポレイション Method and system for creating an aggregated view of user responses in time-varying media using physiological data
JP2011524656A (en) * 2008-04-30 2011-09-01 ディーピー テクノロジーズ インコーポレイテッド Improved headset
JP2013114595A (en) * 2011-11-30 2013-06-10 Canon Inc Information processor, information processing method, and program
JP2014127896A (en) * 2012-12-27 2014-07-07 Samsung R&D Institute Japan Co Ltd Signal processing apparatus and signal processing method
JP2017041673A (en) * 2015-08-17 2017-02-23 パナソニックIpマネジメント株式会社 Viewing state detection device, viewing state detection system and viewing state detection method

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006293979A (en) * 2005-03-18 2006-10-26 Advanced Telecommunication Research Institute International Content providing system
FR3083346B1 (en) * 2018-07-02 2020-07-10 Airbus Operations METHOD AND DEVICE FOR MONITORING THE CAPACITY OF A CREW MEMBER OF AN AIRCRAFT

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002057765A (en) * 2000-08-07 2002-02-22 Matsushita Electric Ind Co Ltd Headset for wireless telephone
JP2002344904A (en) * 2001-02-06 2002-11-29 Sony Corp Device for reproducing content, device for receiving content, method for controlling content display, method for collecting and analyzing content evaluation, device for collecting and analyzing content evaluation, method for managing collection of content evaluation, and device for managing collection of content evaluation
JP2007003618A (en) * 2005-06-21 2007-01-11 Sharp Corp Display device and mobile terminal device
JP2010520554A (en) * 2007-03-06 2010-06-10 エムセンス コーポレイション Method and system for creating an aggregated view of user responses in time-varying media using physiological data
JP2009159073A (en) * 2007-12-25 2009-07-16 Panasonic Corp Acoustic playback apparatus and acoustic playback method
JP2011524656A (en) * 2008-04-30 2011-09-01 ディーピー テクノロジーズ インコーポレイテッド Improved headset
JP2013114595A (en) * 2011-11-30 2013-06-10 Canon Inc Information processor, information processing method, and program
JP2014127896A (en) * 2012-12-27 2014-07-07 Samsung R&D Institute Japan Co Ltd Signal processing apparatus and signal processing method
JP2017041673A (en) * 2015-08-17 2017-02-23 パナソニックIpマネジメント株式会社 Viewing state detection device, viewing state detection system and viewing state detection method

Also Published As

Publication number Publication date
JP2021170248A (en) 2021-10-28
JP2021170315A (en) 2021-10-28
JP6856959B1 (en) 2021-04-14

Similar Documents

Publication Publication Date Title
US20220337693A1 (en) Audio/Video Wearable Computer System with Integrated Projector
US11418893B2 (en) Selective modification of background noises
KR101788485B1 (en) Intelligent device mode shifting based on activity
RU2601287C1 (en) Device for creating interest on the part of a viewer viewing the content
KR20240011874A (en) Directing live entertainment using biometric sensor data for detection of neurological state
JP2019522300A (en) Mobile and wearable video capture and feedback platform for the treatment of mental disorders
US20130245396A1 (en) Mental state analysis using wearable-camera devices
US11626127B2 (en) Systems and methods for processing audio based on changes in active speaker
CN106104650A (en) Remote Device Control is carried out via gaze detection
TW201526667A (en) Hearing aid system and voice capturing method for hearing aid system
US20220232321A1 (en) Systems and methods for retroactive processing and transmission of words
WO2017064891A1 (en) Information processing system, information processing method, and storage medium
US20200250289A1 (en) Automatic low radiation mode for a wearable device
KR20180017821A (en) Broadcasting service apparatus for delivering live audience reaction
US11929087B2 (en) Systems and methods for selectively attenuating a voice
EP2402839A2 (en) System and method for indexing content viewed on an electronic device
WO2021038295A1 (en) Hearing aid system with differential gain
US11580727B2 (en) Systems and methods for matching audio and image information
CN109255314B (en) Information prompting method and device, intelligent glasses and storage medium
US20220284915A1 (en) Separation of signals based on direction of arrival
WO2018075523A1 (en) Audio/video wearable computer system with integrated projector
US20200301398A1 (en) Information processing device, information processing method, and program
KR20140115153A (en) Apparatus and Method for Analysing Audience Reaction And System for Analysing Audience Reaction Using the Same
WO2021210652A1 (en) Information processing device, system, method, and program
WO2022091493A1 (en) Program, information processing device, and method

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21789166

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21789166

Country of ref document: EP

Kind code of ref document: A1