WO2020059047A1 - Evaluation system, server device, terminal device, information processing method, and information processing program - Google Patents

Evaluation system, server device, terminal device, information processing method, and information processing program Download PDF

Info

Publication number
WO2020059047A1
WO2020059047A1 PCT/JP2018/034671 JP2018034671W WO2020059047A1 WO 2020059047 A1 WO2020059047 A1 WO 2020059047A1 JP 2018034671 W JP2018034671 W JP 2018034671W WO 2020059047 A1 WO2020059047 A1 WO 2020059047A1
Authority
WO
WIPO (PCT)
Prior art keywords
information
sound
content
evaluation
video
Prior art date
Application number
PCT/JP2018/034671
Other languages
French (fr)
Japanese (ja)
Inventor
大久 谷川
英士 福田
久美 余川
和俊 木下
吉田 真人
渉 鈴木
Original Assignee
楽天株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 楽天株式会社 filed Critical 楽天株式会社
Priority to PCT/JP2018/034671 priority Critical patent/WO2020059047A1/en
Priority to JP2019503504A priority patent/JP6543429B1/en
Publication of WO2020059047A1 publication Critical patent/WO2020059047A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications

Definitions

  • the present invention relates to a technical field of an evaluation system for allowing a user to view content composed of at least one of video and sound and input evaluation information for the content.
  • Patent Literature 1 discloses a technique for ascertaining the taste of a viewer of a television broadcast.
  • the remote control of the set-top box is provided with a notification button for notifying the viewer's preference for people, objects, music, and the like on the air.
  • the notification button When the notification button is pressed, the set-top box stores the viewing log as a viewing event and transmits the viewing log to the broadcasting device at predetermined time intervals.
  • the broadcasting device counts viewing events that match the conditions set by the program creator or the like.
  • a dedicated remote control is used for a television receiver or a set-top box or the like for receiving a television broadcast signal and displaying a video of a television program, and is evaluated via the television receiver or the set-top box or the like. Since the information is transmitted to the broadcast station or the program creator, it is assured that the user has actually viewed the program and entered the evaluation information. Therefore, the reliability of the input evaluation information is ensured.
  • a general-purpose terminal device such as a mobile phone or a tablet computer is used for inputting the evaluation information by the user, and the evaluation information of each user for the content is provided in a highly versatile system that can be used for other purposes.
  • the issue is how to ensure the reliability of the evaluation information. The reason is that, when such a terminal device is used, the terminal device cannot be dedicated for evaluation input to the television receiver, so that the input operation of the evaluation information by the user is caused by an actual program that watches the television receiver. This is because it is easy to be performed irrespective of the viewing act of the user.
  • Patent Document 2 discloses that a portable communication device acquires sound output from a speaker of a television receiver by a built-in microphone and stores sound data, and broadcasts the sound data by a tuner built in the portable communication device. It is disclosed that data indicating the viewing status of a broadcast program is stored based on acquiring program data, extracting audio data, and comparing the audio data with each other. However, the technique disclosed in Patent Literature 2 attempts to evaluate the viewing situation of a broadcast program by a viewer, but does not aim to evaluate the broadcast program itself.
  • the present invention has been made in view of the above points, and is an evaluation system, a server device, and a terminal device that can ensure the reliability of evaluation information on content even when evaluating content using a general-purpose terminal device. , An information processing method, and an information processing program.
  • one embodiment of the present invention is an evaluation system including a terminal device and a server device connected to the terminal device via a network, wherein the terminal device includes Input means for inputting, by a user, evaluation information for at least one of the contents, and a video or sound output by an output device for outputting at least one of the video and the sound, which constitutes the content Transmitting means for transmitting the input evaluation information and the detected information indicating the detected video or sound to the server device, and the server device transmits the content.
  • An acquisition unit configured to acquire content information indicating the video or the sound, and receiving the evaluation information and the detection information from the terminal device.
  • the method further comprises: selecting means for selecting the received evaluation information to be used for evaluating the content;
  • the content to be evaluated is actually determined based on the comparison between the detection information indicating the video or sound detected by the terminal device and the content information indicating the video or sound of the content acquired by the server device. It is possible to estimate whether or not the user is watching or listening. Therefore, it is possible to preferentially use the evaluation information that is presumed that the user has watched or listened to the content to input the content. Therefore, it is possible to ensure the reliability of the evaluation information.
  • the content is a content to be broadcast, and during the broadcast of the content, the evaluation information and the detection information are transmitted from each of the plurality of terminal devices to the server device as the terminal device.
  • the server device is configured to transmit at least one of the evaluation information and the detection information by at least one terminal device of the plurality of terminal devices, and to transmit at least one of the other terminal devices of the plurality of terminal devices.
  • Timing information transmitting unit that performs Further comprising a timing information receiving means for receiving the timing information from the device, the transmission unit, in accordance with the received timing information, and transmits at least one of the evaluation information and the detection information.
  • the transmission timing of at least one of the evaluation information and the detection information is plural. Are distributed. Therefore, the processing load of the server device can be distributed in the time axis direction.
  • One embodiment of the present invention is further characterized in that the timing information transmitting means determines a transmission timing by the at least one other terminal device during an interval between transmission timings by the at least one terminal device. I do.
  • the transmission timing of at least one terminal device and the transmission timing of another at least one terminal device are different from each other, the number of pieces of information received by the server device per unit time is equalized.
  • the processing load on the device can be further distributed.
  • the detection unit and the transmission unit repeatedly detect the video or the sound at predetermined time intervals and transmit the detection information
  • the selection unit includes the reception unit.
  • the detection information repeatedly received at predetermined time intervals, the detection information indicating the video or sound detected by the terminal device at the time closest to the input time before or after the input time of the evaluation information, The selection is performed based on at least a comparison result with the content information.
  • the server device determines whether to use the evaluation information for evaluating the content by using the detection information indicating the video or the sound detected at a time close to the time when the user inputs the evaluation information. . Therefore, even in a mode in which the terminal device periodically performs detection of video or sound and transmission of the detection information, the estimation accuracy of whether the user is watching or listening to the content at the time of inputting the evaluation information. Can be increased.
  • the selection unit further includes, among the detection information periodically received by the reception unit, a video or a sound detected at a time relatively close to an input time of the evaluation information. The selection is performed on the basis of a comparison result between two or more pieces of detection information each indicating the following and the content information.
  • two or more pieces of detection information indicating a video or a sound detected at a time close to the time when the user inputs the evaluation information are used. Therefore, it is possible to further increase the estimation accuracy of whether or not the user is watching or listening to the content when the evaluation information is input.
  • the detecting unit and the transmitting unit detect the video or the sound when the evaluation information is input, and transmit the detection information to the server device together with the evaluation information.
  • the selection means selects the evaluation information received together with the detection information to be used for evaluating the content.
  • the server device determines whether or not to use the evaluation information for evaluating the content, using the detection information indicating the video or the sound detected when the user inputs the evaluation information. Therefore, it is possible to increase the estimation accuracy of whether or not the user is watching or listening to the content at the time when the evaluation information is input.
  • the selection unit further includes, among the content information, an output time of the video or sound indicated by the portion determined to match the detection information by the output device, And the input time of the evaluation information received together with the evaluation information.
  • the server device specifies, as the input time of the evaluation information, the output time of the video or sound in the content that matches the video or the sound detected when the user inputs the evaluation information. Therefore, it is possible to appropriately specify which scene of the content is the evaluation information input, and thus it is possible to appropriately evaluate the content.
  • the acquisition unit further includes a step of acquiring specific information converted from characteristic information time-series data composed of a time series of characteristic information indicating characteristics of the video or the sound constituting the content.
  • Specific information time-series data configured in time series, wherein the corresponding characteristic information can be specified based on each of the specific information, and the information amount of each of the specific information is the information of the corresponding characteristic information.
  • the terminal device acquires specific information time-series data less than the amount as the content information, and the terminal device obtains the characteristic information time-series data in advance, the characteristic information time-series data obtaining unit, and the detected video or sound.
  • Extracting means for extracting characteristic information indicating the characteristic of the characteristic information, and a degree of coincidence between the extracted characteristic information among the characteristic information included in the acquired characteristic information time-series data is determined.
  • Generating means for generating specific information that specifies characteristic information exceeding the value, and generating specific information having an information amount smaller than the information amount of the characteristic information, wherein the transmitting unit includes: Is transmitted as the detection information, and the selecting unit outputs the output time corresponding to the specific information that matches the specific information received as the detection information among the specific information included in the characteristic information time-series data. Is specified as the input time.
  • the terminal device compares the characteristic information time-series data of the content acquired before the output of the content with the characteristic information extracted from the detected video or sound.
  • the terminal device generates specific information having a smaller information amount from the characteristic information when there is characteristic information in the characteristic information time-series data in which the degree of coincidence with the extracted characteristic information exceeds a predetermined value.
  • the identification information is transmitted to the server device as detection information.
  • the server device compares the specific information time-series data converted from the characteristic information time-series data with the specific information received from the terminal device.
  • the server device acquires, as the input time of the evaluation information, the output time of the video or sound corresponding to the specific information that matches the received specific information, from the specific information time-series data. Therefore, since the amount of detection information is reduced, the communication load on the terminal device and the server device can be reduced.
  • the specific information is a hash value of the characteristic information.
  • One embodiment of the present invention is an acquisition unit configured to acquire content information indicating the video or the sound, which constitutes a content configured of at least one of a video and a sound, and evaluation information for the content is input by a user. And, constituting the content, from a terminal device that detects a video or sound output by an output device that outputs at least one of the video and the sound, the evaluation information, the detected video or Receiving means for receiving detection information indicating a sound; a comparing means for comparing the acquired content information with the received detection information; and a comparison result by the comparing means, wherein the content information and the detection information are obtained. If there is a predetermined match between the received evaluation information and the selected Characterized in that it comprises a and.
  • One embodiment of the present invention provides an input unit in which a user inputs evaluation information for a content composed of at least one of a video and a sound, and at least one of the video and the sound that constitutes the content Detection means for detecting the video or sound output by the output device that outputs the input evaluation information, and the detection information indicating the detected video or sound, constituting the content, the video and According to a comparison result of the content information indicating at least one of the sounds and the detection information, if there is a predetermined match between the content information and the detection information, the evaluation information is compared with the content. Transmitting means for transmitting to a server device selected to be used for evaluation.
  • One embodiment of the present invention relates to an information processing method in an evaluation system including a terminal device and a server device connected to the terminal device via a network, wherein the server device has at least one of video and sound.
  • the apparatus stores the acquired evaluation information and the detection information indicating the detected video or sound in the support.
  • One embodiment of the present invention provides, in an information processing method executed by a computer of a server device, acquiring content information indicating the video or the sound, which constitutes a content configured of at least one of a video and a sound.
  • An acquisition step, and a terminal device that receives the evaluation information for the content by a user and detects a video or a sound output by an output device that outputs at least one of the video and the sound that constitutes the content
  • One aspect of the present invention provides a computer of a server device, an acquisition unit configured to acquire content information indicating the video or the sound, which constitutes a content configured of at least one of a video and a sound, Evaluation information is input by a user, and, from the terminal device that detects the video or sound output by the output device that outputs at least one of the video and the sound that constitutes the content, The detection information indicating the detected video or sound, receiving means for receiving, the comparing means for comparing the acquired content information and the received detection information, and a comparison result by the comparing means, If there is a predetermined match between the content information and the detection information, the received evaluation information is used to evaluate the content. Characterized in that to function as a selecting means for selecting to use.
  • One embodiment of the present invention provides a computer of a terminal device, which acquires evaluation information for a content that is input by a user to input means provided in the terminal device and that is configured by at least one of video and sound.
  • Means from the detection means for detecting the video or sound output by the output device that outputs at least one of the video and the sound that constitutes the content, from the detection information indicating the detected video or sound Detecting information obtaining means for obtaining, the obtained evaluation information, the obtained detection information, the content information constituting at least one of the video and the sound constituting the content, and the detection information If there is a predetermined match between the content information and the detection information according to the comparison result of Wherein transmitting means for transmitting to the server device selected to use in the evaluation of Ntsu, that function as.
  • the reliability of the evaluation information for a content can be ensured even when the content is evaluated using a general-purpose terminal device.
  • FIG. 3 is a diagram illustrating an example of functional blocks of a system control unit 11 of the server 1 according to an embodiment. It is a block diagram showing an example of the outline composition of user terminal 2 concerning one embodiment.
  • FIG. 3 is a diagram illustrating an example of functional blocks of a system control unit 21 of the user terminal 2 according to one embodiment.
  • FIG. 3 is a diagram illustrating an example of functional blocks of a system control unit 11 of the server 1 according to an embodiment.
  • FIG. 3 is a diagram illustrating an example of the outline composition of user terminal 2 concerning one embodiment.
  • FIG. 3 is a diagram illustrating an example of functional blocks of a system control unit 21 of the user terminal 2 according to one embodiment.
  • FIG. 9 is a diagram illustrating an example of information transmission timings by a plurality of user terminals 2.
  • FIG. 9 is a diagram illustrating an example of a generated report.
  • 9 is a flowchart illustrating an example of a terminal process executed by a system control unit 21 of the user terminal 2.
  • 5 is a flowchart illustrating an example of a server process executed by the system control unit 11 of the server 1. It is a figure showing an example of the processing outline in program evaluation system S.
  • 9 is a flowchart illustrating an example of a terminal process executed by a system control unit 21 of the user terminal 2.
  • 5 is a flowchart illustrating an example of a server process executed by the system control unit 11 of the server 1.
  • the content that can be evaluated in the present invention includes at least one of a video (especially a moving image) and a sound.
  • Broadcasting, on-demand distribution, and the like are examples of content distribution modes. Examples of the broadcasting form include terrestrial digital television broadcasting, satellite broadcasting, cable television, radio broadcasting, Internet broadcasting, and the like. Examples of the form of on-demand distribution include satellite broadcasting, cable television, and the Internet.
  • the embodiment described below is an embodiment in which the present invention is applied to a system for evaluating a program in digital terrestrial television broadcasting.
  • FIG. 1 is a diagram illustrating an example of a schematic configuration of a program evaluation system S according to the present embodiment.
  • the program evaluation system S includes a server 1 and one or a plurality of user terminals 2.
  • the server 1 and each user terminal 2 can mutually transmit and receive data via the network NW using, for example, TCP / IP as a communication protocol.
  • the network NW is constructed by, for example, the Internet, a dedicated communication line (for example, a CATV (Community @ Antenna @ Television) line), a mobile communication network (including a base station or the like), a gateway, and the like.
  • the server 1 is a server device for counting or analyzing user evaluations on television programs broadcasted by the broadcasting station 3. In order to determine the reliability of the evaluation by each user, the server 1 acquires information indicating the sound of the program to be evaluated as program sound information during or before the broadcast of the program. During the broadcast of the program to be evaluated, the server 1 receives, from each user terminal 2, evaluation information indicating an evaluation by the user. Further, the server 1 receives, from each user terminal 2, audio data indicating a sound detected by the user terminal 2 and detected sound information. Then, the server 1 determines whether or not to use the evaluation information from the user terminal 2 for evaluating the program based on the detected sound information.
  • Each user terminal 2 is used by a user who has registered as a member in the program evaluation system S.
  • the user terminal 2 used by that user detects the sound output from the television receiver 4 and outputs the detected sound information. Send to server 1.
  • each user terminal 2 transmits evaluation information indicating the input evaluation to the server 1.
  • evaluation information There is no particular upper limit on the number of times that a user can input evaluation information per unit time while viewing a program. For example, a user may input evaluation information at intervals of several seconds or hundreds of milliseconds.
  • the user terminal 2 is preferably a portable computer, but may be a stationary computer. Examples of the user terminal 2 include a mobile information terminal such as a smartphone and a tablet computer, a mobile phone, a PDA (Personal Digital Assistant), and a personal computer.
  • FIG. 2A is a block diagram illustrating an example of a schematic configuration of the server 1 according to the present embodiment.
  • the server 1 includes a system control unit 11, a system bus 12, an input / output interface 13, a storage unit 14, and a communication unit 15.
  • the system control unit 11 and the input / output interface 13 are connected via a system bus 12.
  • the system control unit 11 includes a CPU (Central Processing Unit) 11a, a ROM (Read Only Memory) 11b, a RAM (Random Access Memory) 11c, and the like.
  • the CPU 11a is an example of a processor.
  • the present invention can be applied to various processors different from the CPU.
  • the storage unit 14, the ROM 11b, and the RAM 11c are each an example of a memory.
  • the present invention is applicable to various memories different from a hard disk, a ROM, and a RAM.
  • the input / output interface 13 performs an interface process between the storage unit 14 and the communication unit 15 and the system control unit 11.
  • the storage unit 14 is configured by, for example, a hard disk drive or the like.
  • the storage unit 14 stores program sound information acquired from the broadcast station 3.
  • the storage unit 14 stores the evaluation information received from each user terminal 2.
  • the storage unit 14 stores a user database.
  • the user database stores information on the users registered as members in the program evaluation system S. For example, in the user database, user attributes such as user ID, name, date of birth, gender, and occupation are stored in association with each user.
  • the user ID is information for identifying a user.
  • the storage unit 14 stores various programs such as an operating system and a server program.
  • the server program is a program for acquiring program information, receiving evaluation information and detected sound information, determining whether to use evaluation information, and the like.
  • the server program may be obtained from another server device or the like via the network NW, or may be recorded on a recording medium such as a magnetic tape, an optical disk, or a memory card and read via a drive device. It may be.
  • the communication unit 15 connects to the network NW and controls a communication state with each user terminal 2.
  • FIG. 3A is a block diagram illustrating an example of a schematic configuration of the user terminal 2 according to the present embodiment.
  • the user terminal 2 includes a system control unit 21, a system bus 22, an input / output interface 23, a storage unit 24, a communication unit 25, an operation input unit 26, a display unit 27, A microphone 28 and a camera 29 are provided.
  • the system control unit 21 and the input / output interface 23 are connected via a system bus 22.
  • the system control unit 21 includes a CPU 21a, a ROM 21b, a RAM 21c, and the like.
  • the input / output interface 23 performs an interface process between the storage unit 24 to the camera 29 and the system control unit 21.
  • the storage unit 24 includes, for example, a flash memory, a hard disk drive, and the like.
  • the storage unit 24 stores various programs such as an operating system, a web browser, and a program evaluation application.
  • the program evaluation application is a program for performing processing for using the program evaluation system S.
  • the program evaluation application may be obtained from a server device such as the server 1 via the network NW, or may be recorded on a recording medium such as an optical disk or a memory card and read via a drive device. You may. Note that the program evaluation application may be a web application.
  • the communication unit 25 connects to the network NW and controls a communication state with the server 1.
  • the operation input unit 26 receives an operation performed by the user, and outputs a signal corresponding to the operation content to the system control unit 21.
  • Examples of the operation input unit 26 include a touch panel, a button, a switch, a key, a keyboard, a mouse, and the like.
  • the operation input unit 26 functions as a unit for inputting evaluation information for a broadcast program by a user.
  • the display unit 27 displays information such as images and characters under the control of the system control unit 21.
  • Examples of the display unit 27 include a liquid crystal display and an organic EL (Electro Luminescence) display.
  • the microphone 28 is a device that converts a sound wave into an audio signal and outputs the audio signal to the system control unit 21.
  • the microphone 28 functions as a unit that detects a sound of a broadcast program output from the television receiver 4.
  • the camera 29 captures a still image or a moving image.
  • the camera 29 is configured by, for example, a CCD (Colony Collapse Disorder) image sensor or the like.
  • FIG. 3B is a diagram illustrating an example of functional blocks of the system control unit 21 of the user terminal 2 according to the present embodiment.
  • the system control unit 21 causes the CPU 21a to read and execute various codes included in the program evaluation application, thereby obtaining an evaluation information acquisition unit 211, a detection information acquisition unit 212, an evaluation information / detection information transmission unit. 213 etc.
  • FIG. 4A is a diagram showing an example of a processing outline in the program evaluation system S.
  • the evaluation information acquisition unit 211 acquires, from the operation input unit 26, the evaluation information input by the user for the broadcast program. As shown in FIG. 4A, the user normally inputs evaluation information to the user terminal 2 while viewing a broadcast program to be evaluated on the television receiver 4.
  • FIG. 5 is a diagram showing an example of an input screen for inputting evaluation information for a program.
  • the evaluation information acquisition unit 211 causes the display unit 27 of the user terminal 2 to display the input screen shown in FIG.
  • the evaluation information acquisition unit 211 may perform display control so that the input screen can be displayed only in the broadcast time slot of the program to be evaluated.
  • the input screen may include, for example, evaluation buttons 101 and 102, a comment input area 103, a comment transmission button 104, and the like.
  • the evaluation button 101 is a button that is pressed when the user feels that the program is boring.
  • the evaluation information obtaining unit 211 obtains evaluation information indicating “boring”.
  • the evaluation button 102 is a button that is pressed when the user feels that the program is interesting.
  • the evaluation information obtaining unit 211 obtains evaluation information indicating “interesting”.
  • the comment input area 103 is a button for inputting a comment on the program.
  • the comment transmission button 104 is a button for transmitting a comment input to the comment input area 103.
  • the evaluation information acquisition unit 211 acquires evaluation information including a character string of the input comment.
  • the evaluation of the program is not limited to “interesting”, “boring” and comments. For example, it may be possible to input “funny XX%” by touching the indicator bar displayed on the screen. That is, the degree and tendency of the evaluation may be visually input. In this case, if the indicator bar is selected near the center, the evaluation input is made as "funny 50%".
  • Various other forms of evaluation can be input as evaluation information.
  • the detection information acquisition unit 212 acquires detection sound information indicating a sound detected by the microphone 28. As shown in FIG. 4A, when the user inputs the evaluation information while watching the broadcast program to be evaluated, the sound of the program is detected by the microphone 28.
  • the detection information acquisition unit 212 may generate the detection sound information by, for example, converting an audio signal output from the microphone 28.
  • the detection information acquisition unit 212 may extract the waveform information indicating the characteristics of the sound waveform from the audio signal by analyzing the audio signal as detected sound information.
  • the detection information acquisition unit 212 may define a plurality of amplitude bands and specify an amplitude band in which a waveform of an audio signal is sampled at predetermined sampling intervals.
  • the detection information acquisition unit 212 may count the number of waveform samples for each amplitude band, and generate an array of these sample numbers as waveform information. Alternatively, the detection information acquisition unit 212 may extract the feature amount of the detected sound from the audio signal by analyzing the audio signal. For example, a feature amount is extracted using a discrete Fourier transform or the like. The detection information acquisition unit 212 acquires this feature amount as detection sound information. Alternatively, an audio signal may be converted into audio data in a format such as MP3 (MPEG-1 Audio Layer-3) and acquired as detected sound information.
  • MP3 MPEG-1 Audio Layer-3
  • the detection information acquisition unit 212 controls on / off of sound detection by the microphone 28.
  • the detection information acquiring unit 212 may acquire the detected sound information by turning on the sound detection at all times during the broadcast of the evaluation target program.
  • the detection information acquisition unit 212 may cause the microphone 28 to repeatedly detect sound at predetermined time intervals.
  • the sound detection interval may be, for example, 1 second, 5 seconds, 10 seconds, 30 seconds, 1 minute, or the like.
  • content evaluation which will be described later, can be performed in a timely manner in accordance with changes in the content, and feedback to content creation and editing can be accurately performed.
  • the detection information acquisition unit 212 causes the microphone 28 to continue detecting the sound for a shorter time than the detection interval every time the sound is detected. Alternatively, the detection information acquisition unit 212 may cause the microphone 28 to detect a sound at the timing when the evaluation information is input by the user.
  • the evaluation information / detection information transmitting unit 213 transmits to the server 1 the evaluation information obtained by the evaluation information obtaining unit 211 and the detected sound information obtained by the detection information obtaining unit 212.
  • the evaluation information / detection information transmitting unit 213 may transmit at least one of the evaluation information and the detected sound information collectively after the end of the broadcast of the program to be evaluated.
  • the evaluation information / detection information transmitting unit 213 may transmit the evaluation information and the detected sound information while the program to be evaluated is being broadcast.
  • the evaluation information / detection information transmitting unit 213 may transmit the evaluation information at the timing when the evaluation information is input, and may repeatedly transmit the detection sound information at predetermined time intervals.
  • the evaluation information / detection information transmitting unit 213 may transmit the detected sound information every time the detection is performed. By transmitting the detected sound information to the server 1 periodically, the server 1 can constantly grasp the user's viewing status of the program during the broadcast of the evaluation target program.
  • the evaluation information / detection information transmitting unit 213 may repeatedly transmit the evaluation information together with the detected sound information at predetermined time intervals. In this case, the transmission of the evaluation information and the detected sound information is reserved until the periodic transmission timing comes after the input of the evaluation information.
  • the evaluation information / detection information transmitting unit 213 may transmit the evaluation information together with the detection information at the timing when the evaluation information is input. Although details will be described later, the evaluation information / detection information transmitting unit 213 may transmit the detection sound information at a timing according to the timing information transmitted from the server 1.
  • FIG. 2B is a diagram illustrating an example of functional blocks of the system control unit 11 of the server 1 according to the present embodiment.
  • the system control unit 11 causes the CPU 11a to read and execute various codes included in the server program, thereby obtaining a program information obtaining unit 111, an evaluation information / detection information receiving unit 112, a comparing unit 113, It functions as the information use determining unit 114, the evaluation unit 115, and the like.
  • the program information acquisition unit 111 acquires content information indicating at least one of a video and a sound, which constitutes the content of the broadcast program to be evaluated.
  • the program information acquisition unit 111 acquires program sound information indicating the sound of a broadcast program as content information.
  • the server 1 may receive the audio data of the program from the broadcast station 3 via the network NW before the broadcast of the program starts.
  • audio data of a program may be recorded on a recording medium and loaded into the server 1 via a drive device.
  • a tuner (not shown) may receive a broadcast signal transmitted from the broadcast station 3 while a program is being broadcast, and the server 1 may acquire audio data extracted from the broadcast signal by the tuner in real time.
  • the program information acquisition unit 111 may cause the storage unit 14 to store the audio data of the program as program sound information. Alternatively, the program information acquisition unit 111 may extract the waveform information or feature amount of the sound of the program from the audio data at predetermined time intervals, similarly to the detection information acquisition unit 212 of the user terminal 2. Then, the program information acquisition unit 111 may convert the time series data composed of the time series of the waveform information or the feature amount into a database as the program sound information. In the program sound information, each waveform information or feature amount may be associated with a time at which a sound corresponding to the waveform information or the feature amount is broadcasted in the program to be evaluated. This broadcast time may be an absolute time or a relative time from the broadcast start time.
  • the evaluation information / detection information receiving unit 112 receives the evaluation information and the detection sound information from each user terminal 2. As described above, the evaluation information and the detected sound information may be transmitted together, or may be transmitted at different timings.
  • the evaluation information / detection information receiving unit 112 determines that the transmission timing of at least one of the evaluation information and the detection sound information by at least one user terminal 2 among the plurality of user terminals 2 The transmission timing of at least one of the evaluation information and the detected sound information is determined for each user terminal 2 so as to be different from the transmission timing of at least one of the evaluation information and the detected sound information by at least one user terminal 2. May be. Then, the server 1 may transmit timing information indicating the determined timing to each of the plurality of user terminals 2.
  • the evaluation information / detection information receiving unit 112 may determine the transmission timing of at least one other user terminal 2 during the interval of the transmission timing of at least one user terminal 2 among the plurality of user terminals 2. . Thereby, the number of pieces of detection information received by the server device per unit time is made uniform. In this case, the server 1 causes each user terminal 2 to periodically transmit the detected sound information.
  • the transmission cycle of the information (at least one of the evaluation information and the detected sound information) by each user terminal 2 is P seconds, and the transmission timing of the information is distributed to N transmission timings.
  • the start time of the i-th transmission cycle from the start of program broadcasting is set to Ti.
  • the information transmission timing in the i-th cycle is, for example, Ti + 0 seconds, Ti + 1 ⁇ P / N seconds, Ti + 2 ⁇ P / N seconds,... Ti + (N ⁇ 1) ⁇ P / N seconds. 0, 1 ⁇ P / N, 2 ⁇ P / N,... (N ⁇ 1) ⁇ P / N are offsets from the start time.
  • each user terminal 2 may notify the server 1 when an operation for starting evaluation of a program is performed by a user.
  • the evaluation information / detection information receiving unit 112 may determine one transmission timing from among a plurality of transmission timings cyclically or randomly according to the order. Since the transmission timings of the plurality of user terminals 2 need only be dispersed as a whole, there is no problem that the transmission timings of some of the user terminals 2 overlap.
  • the evaluation information / detection information receiving unit 112 transmits timing information indicating the determined transmission timing to the user terminal 2 that has transmitted the notification.
  • the evaluation information / detection information transmitting unit 213 of the user terminal 2 transmits information according to the timing indicated in the timing information.
  • the timing information may include, for example, a start time of a transmission cycle (for example, 0 seconds per minute), an offset (for example, 0 seconds, 20 seconds, 40 seconds, and the like).
  • FIG. 6 is a diagram illustrating an example of information transmission timings by a plurality of user terminals 2.
  • P is 60 seconds and N is 3.
  • the user terminal 2-1 receives timing information indicating “0 second per minute” from the server 1.
  • the user terminal 2-2 receives timing information indicating “after 20 seconds from 0 seconds every minute” from the server 1.
  • the user terminal 2-3 receives timing information indicating “after a lapse of 40 seconds from every minute” from the server 1. It is assumed that a certain transmission cycle starts at time T. In this case, the user terminal 2-1 transmits information during a period from time T to T + 20 seconds.
  • the user terminal 2-2 transmits information between times T + 20 and T + 40 seconds.
  • the user terminal 2-3 transmits information between times T + 40 and T + 60 seconds. Further, the user terminal 2-1 transmits information between times T + 60 and T + 80 seconds.
  • the user terminal 2-2 transmits information between times T + 80 and T + 100 seconds.
  • the user terminal 2-3 transmits information between times T + 100 and T
  • the comparing unit 113 compares the program sound information acquired by the program information acquiring unit 111 with the detected sound information received by the evaluation information / detected information receiving unit 112. For example, when the program sound information and the detected sound information are waveform information, the comparing unit 113 compares the waveform information or feature amount of the sound detected by the user terminal 2 indicated by the detected sound information with the program indicated by the program sound information.
  • the degree of coincidence between the detected waveform information of the sound and the waveform information of the sound of the program may be calculated by comparing the time-series waveform information of the waveform information of the sound. For example, the degree of coincidence may be calculated based on the coincidence and non-coincidence of the number of samples for each amplitude band.
  • the comparing unit 113 compares only the waveform information broadcasted within a predetermined time before and after the time when the sound is detected in the user terminal 2 with the detected sound information, out of the time series of the waveform information indicated in the program sound information. Is also good. The same applies to the case where the program sound information and the detected sound information are feature amounts.
  • the coincidence in the case of the feature amount may be, for example, a cosine similarity.
  • the program evaluation application may be programmed so that the user can select a program to be evaluated using the user terminal 2.
  • the comparison unit 113 compares the program sound information of the selected program with the detected sound information, assuming that the user is viewing the selected program.
  • the comparing unit 113 may compare the program sound information of each of the plurality of programs with the detected sound information, and determine that the program with the highest degree of coincidence is the program being watched by the user.
  • the evaluation information use determining unit 114 converts the evaluation information received by the evaluation information / detected information receiving unit 112 into Select to use for program evaluation. For example, as shown in FIG. 4A, when it is determined that the sound detected by the user terminal 2 and a part of the sound of the program match, the evaluation information use determining unit 114 uses the evaluation information for evaluation. If it is determined and it is determined that they do not match, it may be determined that the evaluation information is not used for the evaluation.
  • the evaluation information use determining unit 114 determines whether or not the program sound information has a portion that is determined to match the detected sound information. For example, when the program sound information and the detected sound information are waveform information, the evaluation information use determining unit 114 determines that the degree of coincidence with the waveform information of the sound detected by the user terminal 2 among the waveform information included in the program sound information.
  • the evaluation information use determining unit 114 may determine that the detected sound matches the sound of the program. In the present embodiment, accurate voice recognition is not required, and if it is possible to determine the consistency of voice waveforms, it is possible to determine whether the user is actually watching a program.
  • the processing of the evaluation information use determining unit 114 is basically the same. There is a possibility that the microphone 28 of the user terminal 2 may detect a voice of the user or the like, an environmental sound, or the like, together with the sound of the program. Therefore, the threshold may be set lower.
  • the evaluation information use determination unit 114 may determine that the detected sound matches the sound of the program when all the degrees of coincidence exceed the threshold, or that the average value of the degree of coincidence exceeds the threshold and If the standard deviation of the degree of coincidence is smaller than a predetermined value, it may be determined that the detected sound matches the sound of the program. Note that in a situation where the microphone 28 of the user terminal 2 detects a voice or environmental sound of the user or the like, preprocessing may be performed to remove the noise as noise.
  • the evaluation information use determining unit 114 determines whether the sound of the program and the detected sound are different.
  • the match determination can be performed at predetermined time intervals. By shortening the interval, it is possible to accurately determine in real time whether or not the user is watching the program when the user inputs the evaluation information.
  • a user's interest in a program may change in seconds as the program progresses, and if the program is not interesting, the user may stop watching the program or watch another program in the middle of the program. Therefore, the evaluation information in seconds for the program is important. The reliability of the real-time evaluation information can be secured.
  • the evaluation information use determining unit 114 evaluates the evaluation information of the program based on a comparison between the sound detected by the microphone 28 at the same time as or near the time when the evaluation information is input by the user and the sound of the program. May be determined. While the program to be evaluated is being broadcast, whether or not the user is watching the program may change. The evaluation information use determining unit 114 makes a determination according to the change in the viewing situation.
  • FIG. 4B is a diagram illustrating an example of a user action that may occur when a method of determining whether to use evaluation information for evaluating a program is performed. As shown in FIG. 4B, a certain user inputs the evaluation information, for example, five minutes after the start of the broadcast of the program.
  • the evaluation information is used for evaluation. Thereafter, the user leaves the room where the television receiver 4 is located, and inputs evaluation information 30 minutes after the broadcast of the program has started. The sound detected at this time did not match the sound of the program. In this case, the evaluation information is not used for evaluation. Thereafter, the user returned to the room, and input evaluation information 50 minutes after the broadcast of the program started. The sound detected around this time coincided with the sound of the program. In this case, the evaluation information is used for evaluation.
  • the evaluation information use determining unit 114 uses the evaluation information for program evaluation based on a comparison between the detected sound information and the program sound information. It may be determined whether or not there is.
  • the evaluation information use determination unit 114 receives the evaluation information periodically.
  • the comparison result between the detected sound information indicating the sound detected by the microphone 28 of the user terminal 2 at the time before or after the input time of the evaluation information and the time closest to the input time and the program sound information.
  • the user terminal 2 may transmit the input time of the evaluation information to the server 1 together with the evaluation information, or may transmit the time at which the evaluation information is transmitted to the server 1 as the input time of the evaluation information. Alternatively, the time at which the server 1 receives the evaluation information from the user terminal 2 may be used as the input time of the evaluation information. Further, the user terminal 2 may transmit the time at which the sound is detected by the microphone 28 to the server 1 together with the detected sound information, or the time at which the detected sound information is transmitted to the server 1 as the time at which the sound is detected. May be sent. Alternatively, the time at which the server 1 receives the detected sound information from the user terminal 2 may be used as the time at which the sound is detected.
  • the evaluation information use determination unit 114 detects two or more detection sounds indicating the sounds detected by the microphone 28 at a time relatively close to the input time of the evaluation information among the detection sound information repeatedly received at predetermined time intervals. Based on the comparison result between the sound information detected sound information and the program sound information, it may be determined whether or not the evaluation information is used for evaluating the program. That is, two or more pieces of detected sound information are used to determine whether to use the evaluation information for evaluation. This makes it possible to increase the accuracy of determining whether or not the user is watching the program to be evaluated.
  • the evaluation information use determination unit 114 may use a predetermined number of pieces of detected sound information in order of the detected time being closer to the input time before or after the input time of the evaluation information, or the detected time may be the input time. Alternatively, detection sound information within a predetermined time from may be used. Alternatively, the evaluation information use determination unit 114 may use a predetermined number of pieces of detected sound information in order of the detected time being closer to the input time before and after the input time of the evaluation information, or the detected time may be changed from the input time. Detection sound information within a predetermined time may be used.
  • the evaluation information use determining unit 114 may, for example, determine that the evaluation information is used for evaluation when the degree of coincidence for all of the two or more pieces of detected sound information exceeds the threshold, or the average value of the degree of coincidence may exceed the threshold. When the standard deviation of the degree of coincidence is smaller than a predetermined value, it may be determined that the evaluation information is used for the evaluation.
  • whether or not to use the evaluation information for evaluating the program is determined based on only the sound of the program.
  • the evaluation information use determining unit 114 determines whether or not to use the evaluation information based on only the video of the program. Whether to use the evaluation information for evaluating the program may be determined based on both the sound and the sound.
  • the user points the lens of the camera 29 of the user terminal 2 to the television receiver 4.
  • the detection information acquisition unit 212 of the user terminal 2 causes the camera 29 to detect an image, for example, periodically or when evaluation information is input.
  • the detection information obtaining unit 212 may generate the detected video information by extracting the feature amount of the video from the video data output from the camera 29, for example.
  • the feature amount may be extracted using an algorithm such as SIFT (Scale Invariant Feature Transform) or SURF (Speeded Up Robust Feature).
  • the evaluation information / detection information transmitting unit 213 transmits the detected video information to the server 1.
  • the program information acquisition unit 111 of the server 1 causes a tuner to receive a broadcast signal from a broadcast station or acquires video data via a network NW, extracts a feature amount of a video of a program, and generates program video information. May be.
  • the comparing unit 113 compares the program video information with the detected video information, and the evaluation information use determining unit 114 determines whether to use the evaluation information for evaluating the program based on the comparison.
  • the details and modifications when using video may be the same as when using sound.
  • the evaluation information use determination unit 114 may determine that the evaluation information is used for evaluating the program, for example, when it is determined that the video matches and the sound matches.
  • the evaluation unit 115 executes a process of evaluating a program based on the evaluation information determined to be used by the evaluation information use determining unit 114 among the evaluation information received by the evaluation information / detection information receiving unit 112. For example, the evaluation information / detection information receiving unit 112 counts and analyzes the evaluation information. The evaluation unit 115 may calculate the total evaluation of each item such as “uninteresting” and “interesting”, and the number of evaluations input at each time from the start to the end of the broadcast of the program. In addition, the evaluation unit 115 may count the total number of evaluations as a whole or the number of evaluations input at each time.
  • the evaluation unit 115 may generate information indicating a transition of the number of evaluations, information indicating a ranking at a time when the number of evaluations is large, and the like. In addition, the evaluation unit 115 determines the number of users who have participated in the evaluation of the program, and the number of users who have been determined by the evaluation information use determination unit 114 to use at least one piece of evaluation information (the number of effective evaluations is at least one). (The number of users who performed once) may be counted. In addition, the evaluation unit 115 may generate a distribution of the attributes of the user who has performed the effective evaluation. Further, the evaluation unit 115 may generate a list of comments. The evaluation unit 115 may generate a report as a program evaluation result. FIG.
  • the server 1 transmits the generated report via the network NW, for example, in response to a request from a terminal device (not shown) in the broadcasting station 3.
  • the process of evaluating a program may be executed by, for example, a terminal device used by a manager of the program evaluation system S.
  • the evaluation unit 115 may give a privilege such as a point to the ID of the user who has performed the evaluation. As a result, a large amount of evaluation information can be collected by using the motivation of the user, and the reliability of the evaluation result can be increased by increasing the population.
  • the server 1 acquires program sound information in advance and stores it in the storage unit 14. It is assumed that the user terminal 2 periodically detects a sound and transmits detected sound information to the server 1. The server 1 uses the detected sound information received before the evaluation information reception time and at the time closest to the reception time, and determines whether or not to use the evaluation information for program evaluation. It is assumed that waveform information is used as the program sound information and the detected sound information.
  • FIG. 8 is a flowchart illustrating an example of a terminal process executed by the system control unit 21 of the user terminal 2.
  • the user activates the program evaluation application and performs a program survey start operation.
  • the system control unit 21 executes terminal processing according to the program evaluation application.
  • the evaluation information / detection information transmitting unit 213 transmits a survey start notification to the server 1 together with the user ID of the user who uses the user terminal 2 (step S1).
  • the evaluation information / detection information transmission unit 213 receives the timing information from the server 1 and stores it in the RAM 21c (Step S2).
  • the detection information acquisition unit 212 determines whether or not the sound detection timing has arrived based on the current time (step S3). For example, the detection information acquisition unit 212 determines the detection timing in time for the transmission timing based on the transmission timing of the detected sound information indicated by the timing information, the time for continuing the sound detection, and the like. If the detection information acquisition unit 212 determines that the detection timing has come (step S2: YES), the process proceeds to step S4. In step S4, the detection information acquisition unit 212 causes the microphone 28 to detect a sound. Next, the detection information acquisition unit 212 extracts waveform information from the audio signal output from the microphone 28 as detection sound information (step S5).
  • step S3 when the detection information acquisition unit 212 determines that the detection timing has not arrived (step S2: NO), the detection information acquisition unit 212 proceeds with the process to step S6.
  • step S6 the evaluation information / detection information transmitting unit 213 determines whether the transmission timing indicated by the timing information has arrived based on the current time. When it is determined that the transmission timing has arrived (step S6: YES), the evaluation information / detection information transmitting unit 213 advances the process to step S7.
  • step S7 the evaluation information / detection information transmission unit 213 transmits the detection sound information stored in the RAM 21c to the server 1 together with the user ID.
  • step S6 when the evaluation information / detection information transmitting unit 213 determines that the transmission timing has not arrived (step S6: YES), the process proceeds to step S8.
  • step S8 the evaluation information acquisition unit 211 determines whether evaluation information has been input based on a signal from the operation input unit 26. When determining that the evaluation information has been input (step S8: YES), the evaluation information acquisition unit 211 advances the processing to step S9. In step S9, the evaluation information acquisition unit 211 transmits the input evaluation information to the server 1 together with the user ID.
  • step S5 When step S5, S7 or S9 is completed, or when it is determined in step S8 that the evaluation information has not been input (step S8: NO), the system control unit 21 arrives at the broadcast end time of the program to be evaluated. It is determined whether or not (Step S10). If it is determined that the end time has not arrived (step S10: NO), the system control unit 21 advances the processing to step S3. On the other hand, when it is determined that the end time has arrived (step S10: YES), the system control unit 21 ends the terminal processing.
  • FIG. 9 is a flowchart illustrating an example of a server process executed by the system control unit 11 of the server 1.
  • the system control unit 11 starts the server processing, for example, a predetermined time before the start of the broadcast of the program to be evaluated, according to the server program.
  • the evaluation information / detection information receiving unit 112 determines whether a survey start notification has been received from any of the user terminals 2 (step S21). When the evaluation information / detection information receiving unit 112 determines that the investigation start notification has been received (step S21: YES), the process proceeds to step S22. In step S22, the evaluation information / detection information receiving unit 112 stores the viewing flag set to FALSE in the RAM 11c in association with the user ID received together with the survey start notification. The viewing flag is information indicating whether or not the user is viewing the program to be evaluated. Next, the evaluation information / detection information receiving unit 112 determines any one of a plurality of predetermined transmission timings, for example, at random (Step S23). The evaluation information / detection information receiving unit 112 transmits timing information indicating the determined transmission timing to the user terminal 2 that has transmitted the investigation start notification (step S24).
  • step S21 when the evaluation information / detection information receiving unit 112 determines that the investigation start notification has not been received (step S21: NO), the process proceeds to step S25.
  • step S25 the evaluation information / detection information receiving unit 112 determines whether the detection sound information has been received from any of the user terminals 2.
  • step S25 YES
  • the process proceeds to step S26.
  • step S26 the comparing unit 113 calculates the degree of coincidence between the waveform information indicated in the received detected sound information and each piece of waveform information included in the program information.
  • the comparing unit 113 determines whether or not there is any of the pieces of waveform information included in the program information whose calculated coincidence exceeds a threshold.
  • the comparing unit 113 determines that there is waveform information whose degree of coincidence exceeds the threshold (step S26: YES)
  • the comparing unit 113 determines that there is no waveform information whose matching degree exceeds the threshold value (step S26: NO)
  • the viewing flag associated with the user ID received together with the detected sound information is set to FALSE. (Step S28).
  • step S25 when the evaluation information / detection information receiving unit 112 determines that the detection sound information has not been received (step S25: NO), the process proceeds to step S29.
  • step S29 the evaluation information / detection information receiving unit 112 determines whether evaluation information has been received from any of the user terminals 2.
  • step S30 the evaluation information use determining unit 114 determines whether the viewing flag associated with the user ID received along with the evaluation information is TRUE.
  • step S30 determines that the viewing flag is TRUE (step S30: YES)
  • the process proceeds to step S31.
  • step S31 the evaluation information use determining unit 114 sets the input time of the received evaluation information to the reception time of the evaluation information.
  • the evaluation information use determining unit 114 stores the evaluation information, the input time, and the user ID in the storage unit 14 in association with each other (Step S32).
  • the evaluation information use determining unit 114 discards the received evaluation information (step S33).
  • step S24, S27, S28, S32 or S33 is completed, or when it is determined in step S29 that the evaluation information has not been received (step S29: NO)
  • the evaluation unit 115 determines the broadcast end time of the program to be evaluated. Is determined (step S34).
  • step S34: NO the evaluation unit 115 advances the processing to step S21.
  • step S35 the evaluation unit 115 executes a program evaluation process using the evaluation information (step S35).
  • the evaluation unit 115 counts and analyzes the evaluation information stored in the storage unit 14.
  • the evaluation unit 115 may use the input time to calculate the transition of the number of evaluations for each evaluation item, or determine the ranking of the time with the highest number of evaluations.
  • the evaluation unit 115 generates a report indicating the evaluation result, and causes the storage unit 14 to store the report.
  • the evaluation unit 115 ends the server processing.
  • the user terminal 2 detects the sound output by the television receiver 4 that outputs the video and the sound constituting the program. Also, the user terminal 2 transmits the evaluation information input by the user and the detected sound information indicating the detected sound to the server 1.
  • the server 1 acquires program sound information indicating the sound that constitutes the program. Further, the server 1 receives the evaluation information and the detected sound information from the user terminal 2. Further, the server 1 compares the acquired program sound information with the received detected sound information. If the server 1 determines from the comparison result that there is a predetermined match between the program sound information and the received detected sound information, the server 1 selects the received evaluation information to be used for evaluating the program. Therefore, it is possible to estimate whether the user is actually watching the program to be evaluated. Therefore, only the evaluation information that is estimated that the user is watching the program can be used for evaluating the program. Therefore, it is possible to ensure the reliability of the evaluation information.
  • the content may be a content of a broadcast program.
  • the server 1 determines that at least one of the evaluation information and the detected sound information transmitted by at least one of the user terminals 2 among the plurality of user terminals 2 during the broadcast of the program is transmitted to another of the plurality of user terminals 2
  • the transmission timing may be determined for each of the plurality of user terminals 2 so as to be different from the transmission timing by at least one user terminal 2.
  • the server 1 may transmit timing information indicating the determined transmission timing to each of the plurality of user terminals 2.
  • Each user terminal 2 may receive timing information from the server 1. Further, each user terminal 2 may transmit at least one of the evaluation information and the detected sound information according to the transmission timing indicated in the received timing information.
  • the transmission timing of at least one of the evaluation information and the detected sound information is dispersed to a plurality. You. Therefore, it is possible to suppress the processing load of the server 1 from being concentrated at the same point and to distribute the load.
  • the server 1 may determine the transmission timing of at least one other user terminal 2 during the interval of the transmission timing of at least one user terminal 2 among the plurality of user terminals 2. In this case, the number of pieces of detection information received by the server 1 per unit time is made uniform, so that the processing load on the server device can be further dispersed.
  • the user terminal 2 may repeatedly detect the sound at predetermined time intervals and transmit the detected sound information.
  • the server 1 includes, among the detected sound information repeatedly received at predetermined time intervals, detected sound information indicating a sound detected by the user terminal 2 at a time before or after the input time of the evaluation information and closest to the input time. Whether to use the evaluation information for evaluating the program may be determined based at least on the result of comparison with the program sound information. In this case, even in a mode in which the user terminal 2 periodically performs sound detection and transmission of the detected sound information, it is possible to increase the estimation accuracy of whether or not the user is watching the program at the time when the evaluation information is input. it can.
  • the server 1 compares the program sound information with two or more pieces of detected sound information indicating sounds detected at a time relatively close to the input time of the evaluation information among the detected sound information received periodically. Based on the result, it may be determined whether to use the evaluation information for evaluating the program. In this case, the accuracy of estimating whether or not the user is watching the program at the time when the evaluation information is input can be further increased.
  • the user terminal 2 may detect a sound when the evaluation information is input.
  • the server 1 may determine that the evaluation information received together with the detected sound information is used for evaluating the program. In this case, the accuracy of estimating whether the user is watching the program at the time when the evaluation information is input can be improved.
  • the user terminal 2 detects sound at the timing when the evaluation information is input, and acquires the detected sound information.
  • the transmission timing of the evaluation information and the detected sound information may be a point in time when the detected sound information is detected, or may be a regular period.
  • the server 1 specifies the time at which the evaluation information was input to the user terminal 2 based on the detected sound information.
  • the evaluation information use determining unit 114 determines that the evaluation information received together with the detected sound information from the user terminal 2 is used for evaluating the program. .
  • the evaluation information use determining unit 114 receives the output time (broadcast time) of the sound indicated by the portion determined to match the detection information in the program to be evaluated by the television receiver 4 together with the detected sound information. May be specified as the input time of the evaluation information. This time may be an absolute time or a relative time from the broadcast start time. As a result, it is possible to evaluate the program using the input time with high accuracy.
  • This input time specifying method is also effective for on-demand distribution in which the distribution time zone is not predetermined.
  • a feature value of a sound may be used, but feature value specifying information capable of specifying the feature value may be used.
  • This feature amount specifying information is information having an information amount smaller than the feature amount.
  • This characteristic amount specifying information is basically different for each characteristic amount in at least one program. When there are a plurality of programs to be evaluated, it is preferable that the feature amount specifying information differs between the programs.
  • the characteristic amount specifying information may be, for example, a hash value indicating a summary of the characteristic amount, or may be identification information given to the characteristic amount based on a predetermined criterion.
  • the characteristic amount specifying information may include a broadcast time of a sound corresponding to the characteristic amount.
  • waveform information and information capable of specifying the waveform information and information having a smaller information amount than the waveform information may be used.
  • FIG. 10 is a diagram illustrating an example of a processing outline in the program evaluation system S.
  • the program information acquisition unit 111 extracts a feature time from the audio data of the program acquired via the network NW in the time series of the feature amount of the sound of the program. Generate series data. In other words, the audio data of the program is divided into a plurality of pieces and those arranged in chronological order are stored in the storage unit 24 in a state where they are each converted into a feature amount. Further, the program information acquisition unit 111 converts the generated feature amount time series data into specific information time series data.
  • the program information acquisition unit 111 generates a hash value of each feature amount of the feature amount time-series data using a predetermined hash function. Then, the program information acquisition unit 111 generates specific information time-series data composed of a time series of hash values as program sound information and causes the storage unit 14 to store the information.
  • each hash value may be associated with a time at which a sound corresponding to the hash value is broadcast in the program to be evaluated.
  • the user terminal 2 acquires the feature amount time-series data as program sound feature amount information in advance. For example, the user operates the user terminal 2 before the broadcast of the program to be evaluated starts, and preliminarily enters a program survey. Upon receiving the entry notification from the user terminal 2, the server 1 transmits the program sound feature information. The user terminal 2 causes the storage unit 24 to store the program sound feature information.
  • the detection information acquisition unit 212 causes the microphone 28 to detect a sound, and receives an audio signal from the microphone 28.
  • the detection information acquisition unit 212 extracts a feature amount from the audio signal, compares the feature amount with each feature amount in the program sound feature amount information, and calculates a degree of coincidence.
  • the detection information acquisition unit 212 converts the hash value of the feature amount into the same hash function as the hash function used in the server 1. Generated by a function (for example, included in the program evaluation application).
  • the detection information acquisition unit 212 acquires this hash value as detection sound information.
  • the evaluation information / detection information transmitting unit 213 transmits a hash value as detected sound information to the server 1 together with the evaluation information.
  • the comparison unit 113 of the server 1 compares the detected sound information received from the user terminal 2 with each time-series hash value of the hash value as the program sound information. As a result of the comparison, when there is a hash value that matches the detected sound information in the time series of the hash values, the evaluation information / detection information transmitting unit 213 determines that the evaluation information is used for evaluating the program. The evaluation information / detection information transmitting unit 213 specifies a time at which a sound corresponding to a hash value matching the detected sound information is broadcasted in the program to be evaluated as an input time of the evaluation information.
  • the accuracy of the input time used for the evaluation is increased, and the communication amount between the server 1 and the user terminal 2 during the broadcast of the program can be reduced.
  • This reduction in the communication amount is realized by reducing the information amount of the detected sound information by replacing the detected sound information from the characteristic amount to the characteristic amount specifying information. Further, when the sound detected by the microphone 28 does not match the sound of the program, the user terminal 2 does not need to transmit any of the detected sound information and the evaluation information to the server 1, so that the number of times of communication is reduced. You.
  • the user terminal 2 has determined that the feature amount 102 in the program sound feature amount information matches the detected sound information. Therefore, the user terminal 2 transmits the hash value of the feature amount 102 to the server 1. On the server 1 side, the hash value 102 originally generated from the feature amount 102 in the program sound information matches the hash value received from the user terminal 2. Therefore, the time T102 associated with the hash value 102 is the input time of the evaluation information.
  • the operation of the program evaluation system S will be described with reference to FIGS.
  • the user terminal 2 transmits the detected sound information to the server 1 together with the evaluation information at the timing when the evaluation information is input.
  • the user terminal 2 previously acquires the program sound feature amount information and stores it in the storage unit 24.
  • FIG. 11 is a flowchart illustrating an example of a terminal process executed by the system control unit 21 of the user terminal 2. 11, the same processes as those in FIG. 8 are denoted by the same reference numerals.
  • the evaluation information acquisition unit 211 determines whether the evaluation information has been input (step S8). When it is determined that the evaluation information has been input (step S8: YES), the evaluation information acquisition unit 211 proceeds to step S4. In step S4, the detection information acquisition unit 212 causes the microphone 28 to detect a sound. Next, the detection information acquisition unit 212 extracts a feature amount from the audio signal output from the microphone 28 (Step S41). Next, the detection information acquisition unit 212 compares the generated feature amount with each feature amount in the program sound feature amount information. The evaluation information acquisition unit 211 determines whether or not a feature amount whose degree of coincidence with the generated feature amount exceeds a threshold exists in the program sound feature amount information (step S42).
  • step S42 determines that there is a feature amount whose coincidence exceeds the threshold
  • step S43 the detection information acquisition unit 212 generates a hash value of the feature amount having the highest matching degree among the feature amounts whose matching degree of the generated feature amount exceeds the threshold value in the program sound feature amount information.
  • the evaluation information / detection information transmitting unit 213 transmits the evaluation information and the generated hash value to the server 1 together with the user ID (Step S44).
  • step S44 when it is determined in step S8 that the evaluation information has not been input (step S8: NO), or when it is determined in step S42 that there is no feature amount whose matching degree exceeds the threshold value ( (Step S42: NO), the system control unit 21 determines whether or not the broadcast end time of the evaluation target program has arrived (step S10). If it is determined that the end time has not arrived (step S10: NO), the system control unit 21 advances the processing to step S8. On the other hand, when it is determined that the end time has arrived (step S10: YES), the system control unit 21 ends the terminal processing.
  • FIG. 12 is a flowchart illustrating an example of a server process executed by the system control unit 11 of the server 1.
  • the same processes as those in FIG. 9 are denoted by the same reference numerals.
  • the evaluation information / detection information receiving unit 112 determines whether evaluation information and a hash value have been received from any of the user terminals 2 (step S51). When the evaluation information / detection information receiving unit 112 determines that the evaluation information and the hash value have been received (step S51: YES), the process proceeds to step S52. In step S52, the comparison unit 52 determines whether a hash value that matches the received hash value exists in the program sound information. If the evaluation information / detection information receiving unit 112 determines that there is a hash value that matches the received hash value (step S52: YES), the process proceeds to step S53.
  • step S53 the evaluation information use determining unit 114 sets the input time of the received evaluation information to the broadcast time associated with the hash value that matches the received hash value.
  • the evaluation information use determining unit 114 stores the evaluation information, the input time, and the user ID in the storage unit 14 in association with each other (Step S32).
  • the evaluation information use determining unit 114 discards the received evaluation information (step S33).
  • step S32 or S32 When step S32 or S32 is completed, or when it is determined in step S51 that the evaluation information and the hash value have not been received (step S51: NO), the evaluation unit 115 arrives at the broadcast end time of the program to be evaluated. It is determined whether or not (Step S34). If the evaluation unit 115 determines that the end time has not arrived (step S34: NO), the processing proceeds to step S51. On the other hand, when the evaluation unit 115 determines that the end time has arrived (step S34: YES), the evaluation unit 115 executes a program evaluation process using the evaluation information (step S35), and ends the server process.
  • the server 1 determines the output time of the sound indicated by the portion of the program sound information determined to match the detected sound information by the television receiver 4 by using the detected sound information. It is specified as the input time of the evaluation information received together with. Therefore, it is possible to appropriately specify which scene of the program the evaluation information is input to, so that the program can be appropriately evaluated.
  • the server 1 may acquire, as the program sound information, the time series of the feature amount specifying information for specifying each of the feature amounts in the time series of the feature amounts of the sounds constituting the program content.
  • the user terminal 2 may acquire in advance program sound feature amount information indicating a time series of sound feature amounts. Further, when the user terminal 2 includes a feature amount whose degree of coincidence with the detected sound feature amount exceeds a threshold value in the program sound feature amount information, the feature amount identification information for identifying the feature amount May be transmitted as detected sound information.
  • the server 1 specifies, as the input time of the evaluation information, the output time of the sound corresponding to the characteristic amount specifying information that matches the characteristic amount specifying information received from the user terminal 2 in the program sound information. You may. In this case, since the amount of detected sound information is reduced, the communication load on the user terminal 2 and the server 1 can be reduced.
  • server 2 user terminal 11 system control unit 12 system bus 13 input / output interface 14 storage unit 15 communication unit 111 program information acquisition unit 112 evaluation information / detection information reception unit 113 comparison unit 114 evaluation information use determination unit 115 evaluation unit 21 system control Unit 22 system bus 23 input / output interface 24 storage unit 25 communication unit 26 operation input unit 27 display unit microphone 29 camera 211 evaluation information acquisition unit 212 detection information acquisition unit 213 information evaluation information / detection information transmission unit NW network

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Databases & Information Systems (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Abstract

The purpose of the present invention is to ensure the reliability of content evaluation information. This terminal device detects a video or a sound outputted from an output device which outputs at a video and/or a sound constituting content. The terminal device transmits evaluation information inputted from a user regarding the content and detection information indicating the detected video or sound to the server device. The server device acquires content information indicating the video or the sound which constitutes the content. The server device receives the evaluation information and the detection information from the terminal device. The server device compares the acquired content information with the received detection information. The server device selects the received evaluation information so as to be used for evaluating the content when a prescribed degree of matching is recognized between the content information and the detection information from the comparison result.

Description

評価システム、サーバ装置、端末装置、情報処理方法、及び情報処理プログラムEvaluation system, server device, terminal device, information processing method, and information processing program
 本発明は、映像及び音の少なくとも何れか一方で構成されるコンテンツをユーザが視聴してそのコンテンツに対する評価情報を入力するための評価システムの技術分野に関する。 {Circle around (1)} The present invention relates to a technical field of an evaluation system for allowing a user to view content composed of at least one of video and sound and input evaluation information for the content.
 従来、例えばテレビ放送番組等のコンテンツの視聴者がそのコンテンツに対する評価情報を入力することを可能とし、各視聴者からの評価情報を集計、分析等するためのシステムが知られている。例えば、テレビ受像機で番組視聴中に、テレビ受像機のリモコンにおいて所定のボタンが押下されると、テレビ受像機の画面に番組に関する選択肢が表示される。リモコンのカラーボタンの何れかが押下されると、押下されたボタンに対応する選択肢が選択されて、選択結果がテレビ受像機から放送局側へ送信される。また、特許文献1には、テレビ放送の視聴者の嗜好を把握するための技術が開示されている。具体的に、セットトップボックスのリモコンに、放送中の人、物、音楽等に対する視聴者の嗜好を通知するための通知ボタンが備えられる。通知ボタンが押下されると、セットトップボックスは、これを視聴イベントとして視聴ログを記憶して所定時間間隔ごとに視聴ログを放送装置へ送信する。放送装置は、番組制作者等が設定した条件に合致する視聴イベントをカウントする。 Conventionally, there has been known a system which enables a viewer of a content such as a television broadcast program to input evaluation information for the content, and tallies and analyzes evaluation information from each viewer. For example, when a predetermined button is pressed on the remote control of the television receiver while viewing the program on the television receiver, options relating to the program are displayed on the screen of the television receiver. When one of the color buttons on the remote controller is pressed, an option corresponding to the pressed button is selected, and the selection result is transmitted from the television receiver to the broadcast station. Further, Patent Literature 1 discloses a technique for ascertaining the taste of a viewer of a television broadcast. Specifically, the remote control of the set-top box is provided with a notification button for notifying the viewer's preference for people, objects, music, and the like on the air. When the notification button is pressed, the set-top box stores the viewing log as a viewing event and transmits the viewing log to the broadcasting device at predetermined time intervals. The broadcasting device counts viewing events that match the conditions set by the program creator or the like.
 上述した技術においては、テレビ放送信号を受信してテレビ番組の映像を表示するためのテレビ受像機又はセットトップボックス等に専用のリモコンが用いられ、テレビ受像機又はセットトップボックス等を介して評価情報が放送局又は番組制作者側へ送信されるので、ユーザが実際に番組を視聴して評価情報を入力したことが保証される。そのため、入力された評価情報の信頼性が確保される。 In the above-described technology, a dedicated remote control is used for a television receiver or a set-top box or the like for receiving a television broadcast signal and displaying a video of a television program, and is evaluated via the television receiver or the set-top box or the like. Since the information is transmitted to the broadcast station or the program creator, it is assured that the user has actually viewed the program and entered the evaluation information. Therefore, the reliability of the input evaluation information is ensured.
 しかしながら、例えば携帯電話機、タブレット式コンピュータ等の汎用の端末装置をユーザによる評価情報の入力に用い、他の用途にも用いることができるような汎用性の高いシステムにてコンテンツに対する各ユーザの評価情報の集計、分析等を行う場合には、評価情報の信頼性を如何にして確保するかが課題となる。その理由は、そのような端末装置を用いた場合、端末装置をテレビ受像機に対する評価入力用として専用化できていないので、ユーザによる評価情報の入力行為が、テレビ受像機を見た実際の番組の視聴行為とは無関係に行われ易くなるからである。 However, for example, a general-purpose terminal device such as a mobile phone or a tablet computer is used for inputting the evaluation information by the user, and the evaluation information of each user for the content is provided in a highly versatile system that can be used for other purposes. When performing aggregation, analysis, and the like, the issue is how to ensure the reliability of the evaluation information. The reason is that, when such a terminal device is used, the terminal device cannot be dedicated for evaluation input to the television receiver, so that the input operation of the evaluation information by the user is caused by an actual program that watches the television receiver. This is because it is easy to be performed irrespective of the viewing act of the user.
 なお、特許文献2には、携帯通信装置が、テレビ受像機のスピーカから出力された音声を、内蔵するマイクにて取得して音声データを記憶するとともに、携帯通信装置が内蔵するチューナにて放送番組データを取得して音声データを抽出し、音声データ同士を比較することに基づいて、放送番組の視聴状況を示すデータを記憶することが開示されている。しかし、特許文献2に開示された技術では、視聴者による放送番組の視聴状況を評価しようとしているが、放送番組それ自体の評価を目的としていない。 Patent Document 2 discloses that a portable communication device acquires sound output from a speaker of a television receiver by a built-in microphone and stores sound data, and broadcasts the sound data by a tuner built in the portable communication device. It is disclosed that data indicating the viewing status of a broadcast program is stored based on acquiring program data, extracting audio data, and comparing the audio data with each other. However, the technique disclosed in Patent Literature 2 attempts to evaluate the viewing situation of a broadcast program by a viewer, but does not aim to evaluate the broadcast program itself.
特開2000-333154号公報JP 2000-333154 A 特開2017-060060号公報JP-A-2017-060060
 本発明は、以上の点に鑑みてなされたものであり、汎用の端末装置を用いてコンテンツを評価する場合においても、コンテンツに対する評価情報の信頼性を確保可能な評価システム、サーバ装置、端末装置、情報処理方法、及び情報処理プログラムを提供することを目的とする。 The present invention has been made in view of the above points, and is an evaluation system, a server device, and a terminal device that can ensure the reliability of evaluation information on content even when evaluating content using a general-purpose terminal device. , An information processing method, and an information processing program.
 上記課題を解決するために、本発明の一の形態は、端末装置と、前記端末装置とネットワークを介して接続されるサーバ装置と、を含む評価システムにおいて、前記端末装置は、映像及び音の少なくとも何れか一方で構成されるコンテンツに対する評価情報がユーザにより入力される入力手段と、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を出力する出力装置により出力された映像又は音を検出する検出手段と、前記入力された評価情報と、前記検出された映像又は音を示す検出情報と、を前記サーバ装置へ送信する送信手段と、を備え、前記サーバ装置は、前記コンテンツを構成する、前記映像又は前記音を示すコンテンツ情報を取得する取得手段と、前記端末装置から前記評価情報及び前記検出情報を受信する受信手段と、前記取得されたコンテンツ情報と前記受信された検出情報とを比較する比較手段と、前記比較手段による比較結果により、前記コンテンツ情報と前記検出情報との間に所定の一致がある場合 、前記受信された評価情報を前記コンテンツの評価に用いるよう選択する選択手段と、 を備えることを特徴とする。 In order to solve the above-described problem, one embodiment of the present invention is an evaluation system including a terminal device and a server device connected to the terminal device via a network, wherein the terminal device includes Input means for inputting, by a user, evaluation information for at least one of the contents, and a video or sound output by an output device for outputting at least one of the video and the sound, which constitutes the content Transmitting means for transmitting the input evaluation information and the detected information indicating the detected video or sound to the server device, and the server device transmits the content. An acquisition unit configured to acquire content information indicating the video or the sound, and receiving the evaluation information and the detection information from the terminal device. Receiving means, a comparing means for comparing the obtained content information with the received detection information, and a result of the comparison by the comparing means shows that there is a predetermined match between the content information and the detection information. In this case, the method further comprises: selecting means for selecting the received evaluation information to be used for evaluating the content;
 この発明によれば、端末装置により検出された映像又は音を示す検出情報と、サーバ装置が取得しているコンテンツの映像又は音を示すコンテンツ情報との比較に基づいて、評価対象のコンテンツを実際にユーザが観ているか否か又は聴いているか否かを推定することが可能である。従って、ユーザがコンテンツを観ているか又は聴いて入力したと推定される評価情報を優先的に、コンテンツの評価に用いることができる。そのため、評価情報の信頼性を確保することが可能となる。 According to the present invention, the content to be evaluated is actually determined based on the comparison between the detection information indicating the video or sound detected by the terminal device and the content information indicating the video or sound of the content acquired by the server device. It is possible to estimate whether or not the user is watching or listening. Therefore, it is possible to preferentially use the evaluation information that is presumed that the user has watched or listened to the content to input the content. Therefore, it is possible to ensure the reliability of the evaluation information.
 本発明の一の形態は、さらに、前記コンテンツは、放送されるコンテンツであり、前記コンテンツの放送中に、前記端末装置として複数の端末装置それぞれから前記サーバ装置へ前記評価情報及び前記検出情報が送信され、前記サーバ装置は、前記複数の端末装置のうち少なくとも一の端末装置による前記評価情報及び前記検出情報の少なくとも何れか一方の送信タイミングが、前記複数の端末装置のうち他の少なくとも一の端末装置による送信タイミングと相違するように、前記複数の端末装置それぞれについて、前記送信タイミングを決定するタイミング決定手段と、前記決定された送信タイミングを示すタイミング情報を、前記複数の端末装置それぞれに送信するタイミング情報送信手段と、を更に備え、前記端末装置は、前記サーバ装置から前記タイミング情報を受信するタイミング情報受信手段を更に備え、前記送信手段は、前記受信されたタイミング情報に従って、前記評価情報及び前記検出情報の少なくとも何れか一方を送信することを特徴とする。 In one embodiment of the present invention, the content is a content to be broadcast, and during the broadcast of the content, the evaluation information and the detection information are transmitted from each of the plurality of terminal devices to the server device as the terminal device. The server device is configured to transmit at least one of the evaluation information and the detection information by at least one terminal device of the plurality of terminal devices, and to transmit at least one of the other terminal devices of the plurality of terminal devices. Transmitting timing information indicating the transmission timing to each of the plurality of terminal devices to each of the plurality of terminal devices so as to differ from the transmission timing by the terminal device; And a timing information transmitting unit that performs Further comprising a timing information receiving means for receiving the timing information from the device, the transmission unit, in accordance with the received timing information, and transmits at least one of the evaluation information and the detection information.
 この発明によれば、コンテンツの放送中に複数の端末装置それぞれからサーバ装置へ評価情報及び検出情報が送信される態様であっても、評価情報及び検出情報の少なくとも何れか一方の送信タイミングが複数に分散される。従って、サーバ装置の処理負荷を時間軸方向に分散させることができる。 According to the present invention, even when the evaluation information and the detection information are transmitted from each of the plurality of terminal devices to the server device during the broadcast of the content, the transmission timing of at least one of the evaluation information and the detection information is plural. Are distributed. Therefore, the processing load of the server device can be distributed in the time axis direction.
 本発明の一の形態は、さらに、前記タイミング情報送信手段は、前記少なくとも一の端末装置による送信タイミングの間隔の間に、前記他の少なくとも一の端末装置による送信タイミングを決定することを特徴とする。 One embodiment of the present invention is further characterized in that the timing information transmitting means determines a transmission timing by the at least one other terminal device during an interval between transmission timings by the at least one terminal device. I do.
 この発明によれば、少なくとも一の端末装置による送信タイミングと他の少なくとも一の端末装置による送信タイミングとがずれるので、単位時間当たりにサーバ装置が受信する情報の数が均一化されるので、サーバ装置の処理負荷を更に分散させることができる。 According to the present invention, since the transmission timing of at least one terminal device and the transmission timing of another at least one terminal device are different from each other, the number of pieces of information received by the server device per unit time is equalized. The processing load on the device can be further distributed.
 本発明の一の形態は、さらに、前記検出手段及び前記送信手段は、所定時間間隔を置いて繰り返し前記映像又は前記音を検出して前記検出情報を送信し、前記選択手段は、前記受信手段により所定時間間隔を置いて繰り返し受信される前記検出情報のうち、前記評価情報の入力時刻以前又は以後で前記入力時刻に最も近い時刻に前記端末装置により検出される映像又は音を示す検出情報と前記コンテンツ情報との比較結果に少なくとも基づいて、前記選択を行うことを特徴とする。 In one embodiment of the present invention, the detection unit and the transmission unit repeatedly detect the video or the sound at predetermined time intervals and transmit the detection information, and the selection unit includes the reception unit. Among the detection information repeatedly received at predetermined time intervals, the detection information indicating the video or sound detected by the terminal device at the time closest to the input time before or after the input time of the evaluation information, The selection is performed based on at least a comparison result with the content information.
 この発明によれば、サーバ装置は、ユーザが評価情報を入力した時刻に近い時刻に検出された映像又は音を示す検出情報を用いて、評価情報をコンテンツの評価に用いるか否かを判定する。そのため、端末装置が映像又は音の検出及び検出情報の送信を定期的に実行する態様においても、評価情報を入力した時点においてユーザがコンテンツを観ているか否か又は聴いているか否かの推定精度を高めることができる。 According to the present invention, the server device determines whether to use the evaluation information for evaluating the content by using the detection information indicating the video or the sound detected at a time close to the time when the user inputs the evaluation information. . Therefore, even in a mode in which the terminal device periodically performs detection of video or sound and transmission of the detection information, the estimation accuracy of whether the user is watching or listening to the content at the time of inputting the evaluation information. Can be increased.
 本発明の一の形態は、さらに、前記選択手段は、前記受信手段により定期的に受信される前記検出情報のうち、前記評価情報の入力時刻から相対的に近い時刻に検出された映像又は音をそれぞれ示す2以上の検出情報と前記コンテンツ情報との比較結果に基づいて、前記選択を行うことを特徴とする。 In one embodiment of the present invention, the selection unit further includes, among the detection information periodically received by the reception unit, a video or a sound detected at a time relatively close to an input time of the evaluation information. The selection is performed on the basis of a comparison result between two or more pieces of detection information each indicating the following and the content information.
 この発明によれば、ユーザが評価情報を入力した時刻に近い時刻に検出された映像又は音を示す2以上の検出情報が用いられる。そのため、評価情報を入力した時点においてユーザがコンテンツを観ているか否か又は聴いているか否かの推定精度を更に高めることができる。 According to the present invention, two or more pieces of detection information indicating a video or a sound detected at a time close to the time when the user inputs the evaluation information are used. Therefore, it is possible to further increase the estimation accuracy of whether or not the user is watching or listening to the content when the evaluation information is input.
 本発明の一の形態は、さらに、前記検出手段及び前記送信手段は、前記評価情報が入力されたときに前記映像又は前記音を検出して、前記評価情報とともに前記検出情報をサーバ装置へ送信し、前記選択手段は、前記コンテンツ情報が、前記検出情報と一致すると判定される部分を有する場合、前記検出情報とともに受信された前記評価情報を、前記コンテンツの評価に用いるよう選択することを特徴とする。 In one embodiment of the present invention, the detecting unit and the transmitting unit detect the video or the sound when the evaluation information is input, and transmit the detection information to the server device together with the evaluation information. When the content information has a portion determined to match the detection information, the selection means selects the evaluation information received together with the detection information to be used for evaluating the content. And
 この発明によれば、サーバ装置は、ユーザが評価情報を入力した時点で検出された映像又は音を示す検出情報を用いて、その評価情報をコンテンツの評価に用いるか否かを判定する。そのため、評価情報を入力した時点においてユーザがコンテンツを観ているか否か又は聴いているか否かの推定精度を高めることができる。 According to the present invention, the server device determines whether or not to use the evaluation information for evaluating the content, using the detection information indicating the video or the sound detected when the user inputs the evaluation information. Therefore, it is possible to increase the estimation accuracy of whether or not the user is watching or listening to the content at the time when the evaluation information is input.
 本発明の一の形態は、さらに、前記選択手段は、前記コンテンツ情報のうち、前記検出情報と一致すると判定された前記部分により示される映像又は音の前記出力装置による出力時刻を、前記検出情報とともに受信された前記評価情報の入力時刻として特定することを特徴とする。 In one embodiment of the present invention, the selection unit further includes, among the content information, an output time of the video or sound indicated by the portion determined to match the detection information by the output device, And the input time of the evaluation information received together with the evaluation information.
 この発明によれば、サーバ装置は、ユーザが評価情報を入力した時点で検出された映像又は音と一致する、コンテンツにおける映像又は音の出力時刻を、評価情報の入力時刻として特定する。そのためコンテンツの何れの場面に対して入力された評価情報であるかを適切に特定することが可能であるので、コンテンツに対して適切な評価を行うことができる。 According to the present invention, the server device specifies, as the input time of the evaluation information, the output time of the video or sound in the content that matches the video or the sound detected when the user inputs the evaluation information. Therefore, it is possible to appropriately specify which scene of the content is the evaluation information input, and thus it is possible to appropriately evaluate the content.
 本発明の一の形態は、さらに、前記取得手段は、前記コンテンツを構成する前記映像又は前記音の特徴を示す特徴情報の時系列で構成される特徴情報時系列データから変換された特定情報の時系列で構成される特定情報時系列データであって、各前記特定情報に基づいて対応する前記特徴情報が特定可能であり、且つ各前記特定情報の情報量は、対応する前記特徴情報の情報量よりも少ない特定情報時系列データを、前記コンテンツ情報として取得し、前記端末装置は、前記特徴情報時系列データを事前に取得する特徴情報時系列データ取得手段と、前記検出された映像又は音の特徴を示す特徴情報を抽出する抽出手段と、前記取得された特徴情報時系列データに含まれる前記特徴情報のうち、前記抽出された特徴情報との間の一致度が所定値を超える特徴情報を特定する特定情報であって、該特徴情報の情報量よりも少ない情報量の特定情報を生成する生成手段と、を更に備え、前記送信手段は、前記生成された特定情報を、前記検出情報として送信し、前記選択手段は、前記特徴情報時系列データに含まれる前記特定情報のうち、前記検出情報として受信された前記特定情報と一致する特定情報に対応する前記出力時刻を、前記入力時刻として特定することを特徴とする。 In one embodiment of the present invention, the acquisition unit further includes a step of acquiring specific information converted from characteristic information time-series data composed of a time series of characteristic information indicating characteristics of the video or the sound constituting the content. Specific information time-series data configured in time series, wherein the corresponding characteristic information can be specified based on each of the specific information, and the information amount of each of the specific information is the information of the corresponding characteristic information. The terminal device acquires specific information time-series data less than the amount as the content information, and the terminal device obtains the characteristic information time-series data in advance, the characteristic information time-series data obtaining unit, and the detected video or sound. Extracting means for extracting characteristic information indicating the characteristic of the characteristic information, and a degree of coincidence between the extracted characteristic information among the characteristic information included in the acquired characteristic information time-series data is determined. Generating means for generating specific information that specifies characteristic information exceeding the value, and generating specific information having an information amount smaller than the information amount of the characteristic information, wherein the transmitting unit includes: Is transmitted as the detection information, and the selecting unit outputs the output time corresponding to the specific information that matches the specific information received as the detection information among the specific information included in the characteristic information time-series data. Is specified as the input time.
 この発明によれば、端末装置が、コンテンツの出力前に取得しておいたそのコンテンツの特徴情報時系列データと、検出された映像又は音から抽出された特徴情報とを比較する。端末装置は、特徴情報時系列データの中に、抽出された特徴情報との間の一致度が所定値を超える特徴情報が存在する場合、この特徴情報から情報量がより少ない特定情報を生成して、この特定情報を検出情報としてサーバ装置へ送信する。サーバ装置は、特徴情報時系列データから変換された特定情報時系列データと、端末装置から受信された特定情報を比較する。サーバ装置は、特定情報時系列データのうち、受信された特定情報と一致する特定情報に対応する映像又は音の出力装置による出力時刻を、評価情報の入力時刻として取得する。従って、検出情報の情報量が削減されるので、端末装置及びサーバ装置の通信負荷を削減することができる。 According to the present invention, the terminal device compares the characteristic information time-series data of the content acquired before the output of the content with the characteristic information extracted from the detected video or sound. The terminal device generates specific information having a smaller information amount from the characteristic information when there is characteristic information in the characteristic information time-series data in which the degree of coincidence with the extracted characteristic information exceeds a predetermined value. Then, the identification information is transmitted to the server device as detection information. The server device compares the specific information time-series data converted from the characteristic information time-series data with the specific information received from the terminal device. The server device acquires, as the input time of the evaluation information, the output time of the video or sound corresponding to the specific information that matches the received specific information, from the specific information time-series data. Therefore, since the amount of detection information is reduced, the communication load on the terminal device and the server device can be reduced.
 本発明の一の形態は、さらに、前記特定情報は、前記特徴情報のハッシュ値であることを特徴とする。 In one embodiment of the present invention, the specific information is a hash value of the characteristic information.
 本発明の一の形態は、映像及び音の少なくとも何れか一方で構成されるコンテンツを構成する、前記映像又は前記音を示すコンテンツ情報を取得する取得手段と、前記コンテンツに対する評価情報がユーザにより入力され、且つ、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を出力する出力装置により出力された映像又は音を検出する端末装置から、前記評価情報と、前記検出された映像又は音を示す検出情報と、を受信する受信手段と、前記取得されたコンテンツ情報と前記受信された検出情報とを比較する比較手段と、前記比較手段による比較結果により、前記コンテンツ情報と前記検出情報との間に所定の一致がある場合 、前記受信された評価情報を前記コンテンツの評価に用いるよう選択する選択手段と、を備えることを特徴とする。 One embodiment of the present invention is an acquisition unit configured to acquire content information indicating the video or the sound, which constitutes a content configured of at least one of a video and a sound, and evaluation information for the content is input by a user. And, constituting the content, from a terminal device that detects a video or sound output by an output device that outputs at least one of the video and the sound, the evaluation information, the detected video or Receiving means for receiving detection information indicating a sound; a comparing means for comparing the acquired content information with the received detection information; and a comparison result by the comparing means, wherein the content information and the detection information are obtained. If there is a predetermined match between the received evaluation information and the selected Characterized in that it comprises a and.
 本発明の一の形態は、映像及び音の少なくとも何れか一方で構成されるコンテンツに対する評価情報がユーザにより入力される入力手段と、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を出力する出力装置により出力された映像又は音を検出する検出手段と、前記入力された評価情報と、前記検出された映像又は音を示す検出情報とを、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を示すコンテンツ情報と、前記検出情報と、の比較結果比較結果により、前記コンテンツ情報と前記検出情報との間に所定の一致がある場合、前記評価情報を前記コンテンツの評価に用いるよう選択するサーバ装置へ送信する送信手段と、を備えることを特徴とする。 One embodiment of the present invention provides an input unit in which a user inputs evaluation information for a content composed of at least one of a video and a sound, and at least one of the video and the sound that constitutes the content Detection means for detecting the video or sound output by the output device that outputs the input evaluation information, and the detection information indicating the detected video or sound, constituting the content, the video and According to a comparison result of the content information indicating at least one of the sounds and the detection information, if there is a predetermined match between the content information and the detection information, the evaluation information is compared with the content. Transmitting means for transmitting to a server device selected to be used for evaluation.
 本発明の一の形態は、端末装置と、前記端末装置とネットワークを介して接続されるサーバ装置と、を含む評価システムにおける情報処理方法において、前記サーバ装置が、映像及び音の少なくとも何れか一方で構成されるコンテンツを構成する、前記映像又は前記音を示すコンテンツ情報を取得する取得ステップと、前記端末装置が、前記端末装置が備える入力手段にユーザにより入力された、前記コンテンツに対する評価情報を取得する評価情報取得ステップと、前記端末装置が、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を出力する出力装置により出力された映像又は音を検出する検出ステップと、前記端末装置が、前記取得された評価情報と、前記検出された映像又は音を示す検出情報と、を前記サーバ装置へ送信する送信ステップと、前記サーバ装置が、前記端末装置から前記評価情報及び前記検出情報を受信する受信ステップと、前記サーバ装置が、前記取得されたコンテンツ情報と前記受信された検出情報とを比較する比較ステップと、前記サーバ装置が、前記比較ステップによる比較結果により、前記コンテンツ情報と前記検出情報との間に所定の一致がある場合、前記受信された評価情報を前記コンテンツの評価に用いるよう選択するする選択ステップと、を含むことを特徴とする。 One embodiment of the present invention relates to an information processing method in an evaluation system including a terminal device and a server device connected to the terminal device via a network, wherein the server device has at least one of video and sound. An acquisition step of acquiring the content information indicating the video or the sound, which constitutes a content configured by the terminal device, the terminal device being input by a user to an input unit provided in the terminal device, the evaluation information for the content, An evaluation information acquiring step of acquiring, the terminal device constituting the content, a detection step of detecting a video or a sound output by an output device that outputs at least one of the video and the sound, and the terminal The apparatus stores the acquired evaluation information and the detection information indicating the detected video or sound in the support. A transmitting step of transmitting to the server device, the server device receiving the evaluation information and the detection information from the terminal device, and the server device transmits the acquired content information and the received detection information. A comparison step of comparing the content information with the detection information according to a result of the comparison performed by the comparison step. And a selecting step of selecting to use for.
 本発明の一の形態は、サーバ装置のコンピュータにより実行される情報処理方法において、映像及び音の少なくとも何れか一方で構成されるコンテンツを構成する、前記映像又は前記音を示すコンテンツ情報を取得する取得ステップと、前記コンテンツに対する評価情報がユーザにより入力され、且つ、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を出力する出力装置により出力された映像又は音を検出する端末装置から、前記評価情報と、前記検出された映像又は音を示す検出情報と、を受信する受信ステップと、前記取得されたコンテンツ情報と前記受信された検出情報とを比較する比較ステップと、前記比較ステップによる比較結果により、前記コンテンツ情報と前記検出情報との間に所定の一致がある場合、前記受信された評価情報を前記コンテンツの評価に用いるよう選択する選択ステップと、を含むことを特徴とする。 One embodiment of the present invention provides, in an information processing method executed by a computer of a server device, acquiring content information indicating the video or the sound, which constitutes a content configured of at least one of a video and a sound. An acquisition step, and a terminal device that receives the evaluation information for the content by a user and detects a video or a sound output by an output device that outputs at least one of the video and the sound that constitutes the content A receiving step of receiving the evaluation information and detection information indicating the detected video or sound, a comparing step of comparing the obtained content information with the received detection information, There is a predetermined match between the content information and the detection information according to the comparison result in the step If, characterized in that it comprises a selection step of selecting to use the received evaluation information to the evaluation of the content.
 本発明の一の形態は、端末装置のコンピュータにより実行される情報処理方法において、前記端末装置が備える入力手段にユーザにより入力された、映像及び音の少なくとも何れか一方で構成されるコンテンツに対する評価情報を取得する評価情報取得ステップと、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を出力する出力装置により出力された映像又は音を検出する検出ステップと、前記取得された評価情報と、前記検出された映像又は音を示す検出情報とを、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を示すコンテンツ情報と、前記検出情報と、の比較結果により、前記コンテンツ情報と前記検出情報との間に所定の一致がある場合、前記評価情報を前記コンテンツの評価に用いるよう選択するサーバ装置へ送信する送信ステップと、を含むことを特徴とする。 According to one embodiment of the present invention, in an information processing method executed by a computer of a terminal device, evaluation of a content constituted by at least one of video and sound, which is input by a user to input means provided in the terminal device, is performed. An evaluation information acquisition step of acquiring information; a detection step of detecting a video or a sound output by an output device that outputs at least one of the video and the sound, which constitutes the content; and the acquired evaluation Information, the detected information indicating the detected video or sound, constituting the content, content information indicating at least one of the video and the sound, and a comparison result of the detection information, If there is a predetermined match between the content information and the detection information, the evaluation information is used to evaluate the content. Characterized in that it comprises a transmission step of transmitting to the server device selected to use, a.
 本発明の一の形態は、サーバ装置のコンピュータを、映像及び音の少なくとも何れか一方で構成されるコンテンツを構成する、前記映像又は前記音を示すコンテンツ情報を取得する取得手段と、前記コンテンツに対する評価情報がユーザにより入力され、且つ、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を出力する出力装置により出力された映像又は音を検出する端末装置から、前記評価情報と、前記検出された映像又は音を示す検出情報と、を受信する受信手段と、前記取得されたコンテンツ情報と前記受信された検出情報とを比較する比較手段と、前記比較手段による比較結果により、前記コンテンツ情報と前記検出情報との間に所定の一致がある場合、前記受信された評価情報を前記コンテンツの評価に用いるよう選択する選択手段と、として機能させることを特徴とする。 One aspect of the present invention provides a computer of a server device, an acquisition unit configured to acquire content information indicating the video or the sound, which constitutes a content configured of at least one of a video and a sound, Evaluation information is input by a user, and, from the terminal device that detects the video or sound output by the output device that outputs at least one of the video and the sound that constitutes the content, The detection information indicating the detected video or sound, receiving means for receiving, the comparing means for comparing the acquired content information and the received detection information, and a comparison result by the comparing means, If there is a predetermined match between the content information and the detection information, the received evaluation information is used to evaluate the content. Characterized in that to function as a selecting means for selecting to use.
 本発明の一の形態は、端末装置のコンピュータを、前記端末装置が備える入力手段にユーザにより入力された、映像及び音の少なくとも何れか一方で構成されるコンテンツに対する評価情報を取得する評価情報取得手段と、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を出力する出力装置により出力された映像又は音を検出する検出手段から、前記検出された映像又は音を示す検出情報を取得する検出情報取得手段と、前記取得された評価情報と、前記取得された検出情報とを、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を示すコンテンツ情報と、前記検出情報と、の比較結果により、前記コンテンツ情報と前記検出情報との間に所定の一致がある場合、前記評価情報を前記コンテンツの評価に用いるよう選択するサーバ装置へ送信する送信手段と、として機能させることを特徴とする。 One embodiment of the present invention provides a computer of a terminal device, which acquires evaluation information for a content that is input by a user to input means provided in the terminal device and that is configured by at least one of video and sound. Means, from the detection means for detecting the video or sound output by the output device that outputs at least one of the video and the sound that constitutes the content, from the detection information indicating the detected video or sound Detecting information obtaining means for obtaining, the obtained evaluation information, the obtained detection information, the content information constituting at least one of the video and the sound constituting the content, and the detection information If there is a predetermined match between the content information and the detection information according to the comparison result of Wherein transmitting means for transmitting to the server device selected to use in the evaluation of Ntsu, that function as.
 本発明によれば汎用の端末装置を用いてコンテンツを評価する場合においても、コンテンツに対する評価情報の信頼性を確保することができる。 According to the present invention, the reliability of the evaluation information for a content can be ensured even when the content is evaluated using a general-purpose terminal device.
一実施形態に係る番組評価システムSの概要構成の一例を示す図である。It is a figure showing an example of the outline composition of program evaluation system S concerning one embodiment. 一実施形態に係るサーバ1の概要構成の一例を示すブロック図である。It is a block diagram showing an example of the outline composition of server 1 concerning one embodiment. 一実施形態に係るサーバ1のシステム制御部11の機能ブロックの一例を示す図である。FIG. 3 is a diagram illustrating an example of functional blocks of a system control unit 11 of the server 1 according to an embodiment. 一実施形態に係るユーザ端末2の概要構成の一例を示すブロック図である。It is a block diagram showing an example of the outline composition of user terminal 2 concerning one embodiment. 一実施形態に係るユーザ端末2のシステム制御部21の機能ブロックの一例を示す図である。FIG. 3 is a diagram illustrating an example of functional blocks of a system control unit 21 of the user terminal 2 according to one embodiment. 番組評価システムSにおける処理概要の一例を示す図である。It is a figure showing an example of the processing outline in program evaluation system S. 評価情報を番組の評価に用いるか否かを決定する方法の一例を示す図である。It is a figure showing an example of the method of deciding whether to use evaluation information for program evaluation. 番組に対する評価情報を入力するための入力画面の一例を示す図である。It is a figure showing an example of an input screen for inputting evaluation information on a program. 複数のユーザ端末2による情報の送信タイミングの一例を示す図である。FIG. 3 is a diagram illustrating an example of information transmission timings by a plurality of user terminals 2. 生成されたレポートの一例を示す図である。FIG. 9 is a diagram illustrating an example of a generated report. ユーザ端末2のシステム制御部21により実行される端末処理の一例を示すフローチャートである。9 is a flowchart illustrating an example of a terminal process executed by a system control unit 21 of the user terminal 2. サーバ1のシステム制御部11により実行されるサーバ処理の一例を示すフローチャートである。5 is a flowchart illustrating an example of a server process executed by the system control unit 11 of the server 1. 番組評価システムSにおける処理概要の一例を示す図である。It is a figure showing an example of the processing outline in program evaluation system S. ユーザ端末2のシステム制御部21により実行される端末処理の一例を示すフローチャートである。9 is a flowchart illustrating an example of a terminal process executed by a system control unit 21 of the user terminal 2. サーバ1のシステム制御部11により実行されるサーバ処理の一例を示すフローチャートである。5 is a flowchart illustrating an example of a server process executed by the system control unit 11 of the server 1.
 以下、図面を参照して本発明の実施形態について詳細に説明する。本発明において評価対象となり得るコンテンツは、映像(特に動画)及び音の少なくとも何れか一方で構成される。コンテンツの配信形態の例としては、放送、オンデマンド配信等が挙げられる。放送形態の例として、地上デジタルテレビ放送、衛星放送、ケーブルテレビ、ラジオ放送、インターネット放送等が挙げられる。オンデマンド配信の形態の例として、衛星放送、ケーブルテレビ、インターネット等が挙げられる。以下に説明する実施形態は、地上デジタルテレビ放送における番組を評価するためのシステムに対して本発明を適用した場合の実施形態である。 Hereinafter, embodiments of the present invention will be described in detail with reference to the drawings. The content that can be evaluated in the present invention includes at least one of a video (especially a moving image) and a sound. Broadcasting, on-demand distribution, and the like are examples of content distribution modes. Examples of the broadcasting form include terrestrial digital television broadcasting, satellite broadcasting, cable television, radio broadcasting, Internet broadcasting, and the like. Examples of the form of on-demand distribution include satellite broadcasting, cable television, and the Internet. The embodiment described below is an embodiment in which the present invention is applied to a system for evaluating a program in digital terrestrial television broadcasting.
[1.第1実施形態] [1. First Embodiment]
[1-1.番組評価システムの構成] [1-1. Configuration of Program Evaluation System]
 先ず、本実施形態に係る番組評価システムSの構成について、図1を用いて説明する。図1は、本実施形態に係る番組評価システムSの概要構成の一例を示す図である。 First, the configuration of the program evaluation system S according to the present embodiment will be described with reference to FIG. FIG. 1 is a diagram illustrating an example of a schematic configuration of a program evaluation system S according to the present embodiment.
 図1に示すように、番組評価システムSは、サーバ1と、1又は複数のユーザ端末2と、を含んで構成されている。サーバ1と各ユーザ端末2とは、ネットワークNWを介して、例えば、通信プロトコルにTCP/IP等を用いて相互にデータの送受信が可能になっている。なお、ネットワークNWは、例えば、インターネット、専用通信回線(例えば、CATV(Community Antenna Television)回線)、移動体通信網(基地局等を含む)、及びゲートウェイ等により構築されている。 As shown in FIG. 1, the program evaluation system S includes a server 1 and one or a plurality of user terminals 2. The server 1 and each user terminal 2 can mutually transmit and receive data via the network NW using, for example, TCP / IP as a communication protocol. The network NW is constructed by, for example, the Internet, a dedicated communication line (for example, a CATV (Community @ Antenna @ Television) line), a mobile communication network (including a base station or the like), a gateway, and the like.
 サーバ1は、放送局3により放送されるテレビ番組に対するユーザの評価を集計又は分析するためのサーバ装置である。サーバ1は、各ユーザによる評価の信頼性を判定するために、評価対象の番組の音を示す情報を、番組音情報として、その番組の放送中に又は放送前に取得する。評価対象の番組の放送中に、サーバ1は、各ユーザ端末2から、ユーザによる評価を示す評価情報を受信する。また、サーバ1は、各ユーザ端末2から、ユーザ端末2により検出された音を示す音声データを、検出音情報を受信する。そして、サーバ1は、検出音情報に基づいて、ユーザ端末2からの評価情報を、番組の評価に用いるか否かを判定する。 The server 1 is a server device for counting or analyzing user evaluations on television programs broadcasted by the broadcasting station 3. In order to determine the reliability of the evaluation by each user, the server 1 acquires information indicating the sound of the program to be evaluated as program sound information during or before the broadcast of the program. During the broadcast of the program to be evaluated, the server 1 receives, from each user terminal 2, evaluation information indicating an evaluation by the user. Further, the server 1 receives, from each user terminal 2, audio data indicating a sound detected by the user terminal 2 and detected sound information. Then, the server 1 determines whether or not to use the evaluation information from the user terminal 2 for evaluating the program based on the detected sound information.
 各ユーザ端末2は、番組評価システムSに会員登録しているユーザにより利用される。各ユーザが、それぞれのテレビ受像機4にて評価対象番組を視聴しているとき、そのユーザが利用するユーザ端末2は、テレビ受像機4から出力される音を検出して、検出音情報をサーバ1へ送信する。また、各ユーザ端末2は、ユーザにより番組に対する評価が入力されると、入力された評価を示す評価情報をサーバ1へ送信する。番組の視聴中、単位時間当たりにユーザが評価情報を入力することができる回数に特に上限は設けられていない。例えば、ユーザは数秒間隔又は数百ミリ秒間隔で評価情報を入力してもよい。ユーザ端末2は、携帯可能なコンピュータであることが望ましいが、据え置き型のコンピュータであってもよい。ユーザ端末2の例として、スマートフォン、タブレット式コンピュータ等の携帯情報端末、携帯電話機、PDA(Personal Digital Assistant)、パーソナルコンピュータ等が挙げられる。 Each user terminal 2 is used by a user who has registered as a member in the program evaluation system S. When each user is watching a program to be evaluated on each television receiver 4, the user terminal 2 used by that user detects the sound output from the television receiver 4 and outputs the detected sound information. Send to server 1. Further, when a user inputs an evaluation for a program, each user terminal 2 transmits evaluation information indicating the input evaluation to the server 1. There is no particular upper limit on the number of times that a user can input evaluation information per unit time while viewing a program. For example, a user may input evaluation information at intervals of several seconds or hundreds of milliseconds. The user terminal 2 is preferably a portable computer, but may be a stationary computer. Examples of the user terminal 2 include a mobile information terminal such as a smartphone and a tablet computer, a mobile phone, a PDA (Personal Digital Assistant), and a personal computer.
[1-2.サーバの構成] [1-2. Server configuration]
 次に、サーバ1の構成について、図2Aを用いて説明する。図2Aは、本実施形態に係るサーバ1の概要構成の一例を示すブロック図である。図2Aに示すように、サーバ1は、システム制御部11と、システムバス12と、入出力インターフェース13と、記憶部14と、通信部15と、を備えている。システム制御部11と入出力インターフェース13とは、システムバス12を介して接続されている。 Next, the configuration of the server 1 will be described with reference to FIG. 2A. FIG. 2A is a block diagram illustrating an example of a schematic configuration of the server 1 according to the present embodiment. As shown in FIG. 2A, the server 1 includes a system control unit 11, a system bus 12, an input / output interface 13, a storage unit 14, and a communication unit 15. The system control unit 11 and the input / output interface 13 are connected via a system bus 12.
 システム制御部11は、CPU(Central Processing Unit)11a、ROM(Read Only Memory)11b、RAM(Random Access Memory)11c等により構成されている。CPU11aは、プロセッサの一例である。なお、本発明は、CPUと異なる様々なプロセッサに対しても適用可能である。記憶部14、ROM11b及びRAM11cは、それぞれメモリの一例である。なお、本発明は、ハードディスク、ROM及びRAMと異なる様々なメモリに対しても適用可能である。 The system control unit 11 includes a CPU (Central Processing Unit) 11a, a ROM (Read Only Memory) 11b, a RAM (Random Access Memory) 11c, and the like. The CPU 11a is an example of a processor. The present invention can be applied to various processors different from the CPU. The storage unit 14, the ROM 11b, and the RAM 11c are each an example of a memory. The present invention is applicable to various memories different from a hard disk, a ROM, and a RAM.
 入出力インターフェース13は、記憶部14及び通信部15とシステム制御部11との間のインターフェース処理を行う。 The input / output interface 13 performs an interface process between the storage unit 14 and the communication unit 15 and the system control unit 11.
 記憶部14は、例えば、ハードディスクドライブ等により構成されている。この記憶部14には、放送局3から取得された番組音情報が記憶される。また、記憶部14には、各ユーザ端末2から受信された評価情報が記憶される。また、記憶部14には、ユーザデータベースが記憶されている。ユーザデータベースには、番組評価システムSに会員登録されているユーザに関する情報が記憶される。例えば、ユーザデータベースには、ユーザID、氏名、生年月日、性別、職業等のユーザの属性が、ユーザごとに関連付けて記憶される。ユーザIDは、ユーザを識別する情報である。更に、記憶部14には、オペレーティングシステム、サーバプログラム等の各種プログラムが記憶されている。サーバプログラムは、番組情報の取得、評価情報及び検出音情報の受信、評価情報を利用するか否かの判定等を実行するためのプログラムである。サーバプログラムは、例えば、他のサーバ装置等からネットワークNWを介して取得されるようにしてもよいし、磁気テープ、光ディスク、メモリカード等の記録媒体に記録されてドライブ装置を介して読み込まれるようにしてもよい。 The storage unit 14 is configured by, for example, a hard disk drive or the like. The storage unit 14 stores program sound information acquired from the broadcast station 3. The storage unit 14 stores the evaluation information received from each user terminal 2. The storage unit 14 stores a user database. The user database stores information on the users registered as members in the program evaluation system S. For example, in the user database, user attributes such as user ID, name, date of birth, gender, and occupation are stored in association with each user. The user ID is information for identifying a user. Further, the storage unit 14 stores various programs such as an operating system and a server program. The server program is a program for acquiring program information, receiving evaluation information and detected sound information, determining whether to use evaluation information, and the like. The server program may be obtained from another server device or the like via the network NW, or may be recorded on a recording medium such as a magnetic tape, an optical disk, or a memory card and read via a drive device. It may be.
 通信部15は、ネットワークNWに接続して、各ユーザ端末2との間の通信状態を制御する。 The communication unit 15 connects to the network NW and controls a communication state with each user terminal 2.
[1-3.ユーザ端末の構成] [1-3. Configuration of User Terminal]
 次に、ユーザ端末2の構成について、図3Aを用いて説明する。図3Aは、本実施形態に係るユーザ端末2の概要構成の一例を示すブロック図である。図3Aに示すように、ユーザ端末2は、システム制御部21と、システムバス22と、入出力インターフェース23と、記憶部24と、通信部25と、操作入力部26と、表示部27と、マイク28と、カメラ29とを備えている。システム制御部21と入出力インターフェース23とは、システムバス22を介して接続されている。 Next, the configuration of the user terminal 2 will be described with reference to FIG. 3A. FIG. 3A is a block diagram illustrating an example of a schematic configuration of the user terminal 2 according to the present embodiment. As shown in FIG. 3A, the user terminal 2 includes a system control unit 21, a system bus 22, an input / output interface 23, a storage unit 24, a communication unit 25, an operation input unit 26, a display unit 27, A microphone 28 and a camera 29 are provided. The system control unit 21 and the input / output interface 23 are connected via a system bus 22.
 システム制御部21は、CPU21a、ROM21b、RAM21c等により構成されている。 The system control unit 21 includes a CPU 21a, a ROM 21b, a RAM 21c, and the like.
 入出力インターフェース23は、記憶部24~カメラ29とシステム制御部21との間のインターフェース処理を行う。 The input / output interface 23 performs an interface process between the storage unit 24 to the camera 29 and the system control unit 21.
 記憶部24は、例えば、フラッシュメモリ、ハードディスクドライブ等により構成されている。この記憶部24には、オペレーティングシステム、ウェブブラウザ、番組評価アプリケーション等の各種プログラムが記憶される。番組評価アプリケーションは、番組評価システムSを利用するための処理を行うためのプログラムである。番組評価アプリケーションは、例えば、サーバ1等のサーバ装置からネットワークNWを介して取得されるようにしてもよいし、光ディスク、メモリカード等の記録媒体に記録されてドライブ装置を介して読み込まれるようにしてもよい。なお、番組評価アプリケーションは、ウェブアプリケーションであってもよい。 The storage unit 24 includes, for example, a flash memory, a hard disk drive, and the like. The storage unit 24 stores various programs such as an operating system, a web browser, and a program evaluation application. The program evaluation application is a program for performing processing for using the program evaluation system S. The program evaluation application may be obtained from a server device such as the server 1 via the network NW, or may be recorded on a recording medium such as an optical disk or a memory card and read via a drive device. You may. Note that the program evaluation application may be a web application.
 通信部25は、ネットワークNWに接続して、サーバ1との間の通信状態を制御する。 The communication unit 25 connects to the network NW and controls a communication state with the server 1.
 操作入力部26は、ユーザによる操作を受け付け、操作内容に対応する信号をシステム制御部21に出力する。操作入力部26の例として、タッチパネル、ボタン、スイッチ、キー、キーボード、マウス等が挙げられる。操作入力部26は、放送番組に対する評価情報がユーザにより入力される手段として機能する。 The operation input unit 26 receives an operation performed by the user, and outputs a signal corresponding to the operation content to the system control unit 21. Examples of the operation input unit 26 include a touch panel, a button, a switch, a key, a keyboard, a mouse, and the like. The operation input unit 26 functions as a unit for inputting evaluation information for a broadcast program by a user.
 表示部27は、システム制御部21の制御により、画像、文字等の情報を表示する。表示部27の例として、液晶ディスプレイ、有機EL(Electro Luminescence)ディスプレイ等が挙げられる。 The display unit 27 displays information such as images and characters under the control of the system control unit 21. Examples of the display unit 27 include a liquid crystal display and an organic EL (Electro Luminescence) display.
 マイク28は、音波を音声信号に変換して、音声信号をシステム制御部21へ出力するデバイスである。マイク28は、テレビ受像機4から出力された放送番組の音を検出する手段として機能する。 The microphone 28 is a device that converts a sound wave into an audio signal and outputs the audio signal to the system control unit 21. The microphone 28 functions as a unit that detects a sound of a broadcast program output from the television receiver 4.
 カメラ29は、静止画又は動画を撮影する。カメラ29は、例えばCCD(Colony Collapse Disorder)イメージセンサ等により構成されている。 The camera 29 captures a still image or a moving image. The camera 29 is configured by, for example, a CCD (Colony Collapse Disorder) image sensor or the like.
[1-4.機能概要] [1-4. Functional overview]
 次に、図2B、図3B乃至図6を用いて、サーバ1のシステム制御部11、及びユーザ端末2のシステム制御部21の機能概要を説明する。 Next, an outline of functions of the system control unit 11 of the server 1 and the system control unit 21 of the user terminal 2 will be described with reference to FIGS. 2B and 3B to 6.
 図3Bは、本実施形態に係るユーザ端末2のシステム制御部21の機能ブロックの一例を示す図である。システム制御部21は、CPU21aが、番組評価アプリケーションに含まれる各種コードを読み出し実行することにより、図3Bに示すように、評価情報取得部211、検出情報取得部212、評価情報・検出情報送信部213等として機能する。 FIG. 3B is a diagram illustrating an example of functional blocks of the system control unit 21 of the user terminal 2 according to the present embodiment. As shown in FIG. 3B, the system control unit 21 causes the CPU 21a to read and execute various codes included in the program evaluation application, thereby obtaining an evaluation information acquisition unit 211, a detection information acquisition unit 212, an evaluation information / detection information transmission unit. 213 etc.
 図4Aは、番組評価システムSにおける処理概要の一例を示す図である。評価情報取得部211は、放送番組に対してユーザにより入力された評価情報を操作入力部26から取得する。図4Aに示すように、ユーザは、通常であればテレビ受像機4により評価対象の放送番組を視聴しながら、ユーザ端末2に評価情報を入力する。 FIG. 4A is a diagram showing an example of a processing outline in the program evaluation system S. The evaluation information acquisition unit 211 acquires, from the operation input unit 26, the evaluation information input by the user for the broadcast program. As shown in FIG. 4A, the user normally inputs evaluation information to the user terminal 2 while viewing a broadcast program to be evaluated on the television receiver 4.
 図5は、番組に対する評価情報を入力するための入力画面の一例を示す図である。番組評価アプリケーションを起動してユーザが番組調査の開始操作を行うことにより、評価情報取得部211は、ユーザ端末2の表示部27に、図5に示す入力画面を表示させる。評価情報取得部211は、例えば、評価対象の番組の放送時間帯にのみ入力画面が表示可能となるように、表示制御を行ってもよい。入力画面は、例えば評価ボタン101及び102、コメント入力領域103、コメント送信ボタン104等により構成されてもよい。評価ボタン101は、番組がつまらないとユーザが感じたときに押されるボタンである。評価ボタン101が押されると、評価情報取得部211は、「つまらない」を示す評価情報を取得する。評価ボタン102は、番組がおもしろいとユーザが感じたときに押されるボタンである。評価ボタン102が押されると、評価情報取得部211は、「おもしろい」を示す評価情報を取得する。コメント入力領域103は、番組に対するコメントを入力するためのボタンである。コメント送信ボタン104は、コメント入力領域103に入力されたコメントを送信するためのボタンである。コメント送信ボタン104が押されると、評価情報取得部211は、入力されたコメントの文字列を含む評価情報を取得する。なお、番組に対する評価は、「おもしろい」、「つまらない」及びコメントに限定されるものではない。例えば、画面に表示されたインジケータバーのタッチ操作により「面白さ○○%」を入力できるものであってもよい。つまり、評価の程度や傾向を可視的に入力できるものであってもよい。その場合、インジケータバーを中央付近として選択するならば「面白さ50%」として評価入力されることになる。その他様々な形態の評価が評価情報として入力可能である。 FIG. 5 is a diagram showing an example of an input screen for inputting evaluation information for a program. When the program evaluation application is activated and the user performs a program survey start operation, the evaluation information acquisition unit 211 causes the display unit 27 of the user terminal 2 to display the input screen shown in FIG. For example, the evaluation information acquisition unit 211 may perform display control so that the input screen can be displayed only in the broadcast time slot of the program to be evaluated. The input screen may include, for example, evaluation buttons 101 and 102, a comment input area 103, a comment transmission button 104, and the like. The evaluation button 101 is a button that is pressed when the user feels that the program is boring. When the evaluation button 101 is pressed, the evaluation information obtaining unit 211 obtains evaluation information indicating “boring”. The evaluation button 102 is a button that is pressed when the user feels that the program is interesting. When the evaluation button 102 is pressed, the evaluation information obtaining unit 211 obtains evaluation information indicating “interesting”. The comment input area 103 is a button for inputting a comment on the program. The comment transmission button 104 is a button for transmitting a comment input to the comment input area 103. When the comment transmission button 104 is pressed, the evaluation information acquisition unit 211 acquires evaluation information including a character string of the input comment. The evaluation of the program is not limited to “interesting”, “boring” and comments. For example, it may be possible to input “funny XX%” by touching the indicator bar displayed on the screen. That is, the degree and tendency of the evaluation may be visually input. In this case, if the indicator bar is selected near the center, the evaluation input is made as "funny 50%". Various other forms of evaluation can be input as evaluation information.
 検出情報取得部212は、マイク28により検出された音を示す検出音情報を取得する。図4Aに示すように、ユーザが評価対象の放送番組を視聴しながら評価情報を入力する場合には、マイク28により番組の音が検出されることになる。検出情報取得部212は、例えばマイク28から出力された音声信号を変換することにより、検出音情報を生成してもよい。例えば、検出情報取得部212は、音声信号を解析することにより、音の波形の特徴を示す波形情報を、検出音情報として音声信号から抽出してもよい。例えば、検出情報取得部212は、複数の振幅帯域を定義し、所定のサンプリング間隔ごとに、音声信号の波形がサンプリングされる振幅帯域を特定してもよい。検出情報取得部212は、振幅帯域ごとの波形のサンプル数をカウントして、これらサンプル数の配列を波形情報として生成してもよい。或いは、検出情報取得部212は、音声信号を解析することにより、検出された音の特徴量を音声信号から抽出してもよい。例えば、離散フーリエ変換等を用いて特徴量が抽出される。検出情報取得部212は、この特徴量を検出音情報として取得する。或いは、音声信号がMP3(MPEG-1 Audio Layer-3)等のフォーマットの音声データに変換されて、検出音情報として取得されてもよい。 The detection information acquisition unit 212 acquires detection sound information indicating a sound detected by the microphone 28. As shown in FIG. 4A, when the user inputs the evaluation information while watching the broadcast program to be evaluated, the sound of the program is detected by the microphone 28. The detection information acquisition unit 212 may generate the detection sound information by, for example, converting an audio signal output from the microphone 28. For example, the detection information acquisition unit 212 may extract the waveform information indicating the characteristics of the sound waveform from the audio signal by analyzing the audio signal as detected sound information. For example, the detection information acquisition unit 212 may define a plurality of amplitude bands and specify an amplitude band in which a waveform of an audio signal is sampled at predetermined sampling intervals. The detection information acquisition unit 212 may count the number of waveform samples for each amplitude band, and generate an array of these sample numbers as waveform information. Alternatively, the detection information acquisition unit 212 may extract the feature amount of the detected sound from the audio signal by analyzing the audio signal. For example, a feature amount is extracted using a discrete Fourier transform or the like. The detection information acquisition unit 212 acquires this feature amount as detection sound information. Alternatively, an audio signal may be converted into audio data in a format such as MP3 (MPEG-1 Audio Layer-3) and acquired as detected sound information.
 検出情報取得部212は、マイク28による音検出のオン/オフを制御する。例えば、検出情報取得部212は、評価対象の番組の放送中、常時音検出をオンにさせて検出音情報を取得してもよい。しかしながら、処理負荷、消費電力、ネットワーク負荷等を考慮すると、所定のタイミングにのみ検出を行うことが望ましい。例えば、検出情報取得部212は、所定時間間隔を置いて繰り返しマイク28により音を検出させてもよい。音の検出間隔は、例えば1秒、5秒、10秒、30秒、1分等であってもよい。なお、音の検出間隔を短く設定するに従い、後述するコンテンツ評価をコンテンツの変化に応じてタイムリーに行うことができるようになり、コンテンツの作成や編集へのフィードバックを正確に行うことができるようになる。検出情報取得部212は、1回の音の検出ごとに、検出間隔よりも短い時間の間、マイク28による音の検出を継続させる。或いは、検出情報取得部212は、ユーザにより評価情報が入力されたタイミングで、マイク28により音を検出させてもよい。 The detection information acquisition unit 212 controls on / off of sound detection by the microphone 28. For example, the detection information acquiring unit 212 may acquire the detected sound information by turning on the sound detection at all times during the broadcast of the evaluation target program. However, in consideration of processing load, power consumption, network load, and the like, it is desirable to perform detection only at a predetermined timing. For example, the detection information acquisition unit 212 may cause the microphone 28 to repeatedly detect sound at predetermined time intervals. The sound detection interval may be, for example, 1 second, 5 seconds, 10 seconds, 30 seconds, 1 minute, or the like. In addition, as the sound detection interval is set shorter, content evaluation, which will be described later, can be performed in a timely manner in accordance with changes in the content, and feedback to content creation and editing can be accurately performed. become. The detection information acquisition unit 212 causes the microphone 28 to continue detecting the sound for a shorter time than the detection interval every time the sound is detected. Alternatively, the detection information acquisition unit 212 may cause the microphone 28 to detect a sound at the timing when the evaluation information is input by the user.
 評価情報・検出情報送信部213は、評価情報取得部211により取得された評価情報と、検出情報取得部212により取得された検出音情報を、サーバ1へ送信する。評価情報・検出情報送信部213は、例えば評価対象の番組の放送終了後に、評価情報及び検出音情報の少なくとも何れか一方をまとめて送信してもよい。一方で、評価情報・検出情報送信部213は、評価対象の番組の放送中に、評価情報及び検出音情報を送信してもよい。例えば、評価情報・検出情報送信部213は、評価情報が入力されたタイミングでこの評価情報を送信する一方で、検出音情報は所定時間間隔を置いて繰り返し送信してもよい。例えば、マイク28により所定時間間隔を置いて繰り返し音の検出が行われる場合、評価情報・検出情報送信部213は、検出が行われるたびに検出音情報を送信してもよい。定期的に検出音情報がサーバ1へ送信されることにより、サーバ1において、評価対象の番組の放送中、ユーザによる番組の視聴状況を常時把握することができる。評価情報が入力されたタイミングでマイク28により音の検出が行われる場合、評価情報・検出情報送信部213は、所定時間間隔を置いて繰り返し、検出音情報とともに評価情報を送信してもよい。この場合、評価情報が入力されてから、定期的な送信タイミングが到来するまで、評価情報及び検出音情報の送信が留保される。或いは、評価情報・検出情報送信部213は、評価情報が入力されたタイミングで、検出情報とともに評価情報を送信してもよい。なお、詳細は後述するが、評価情報・検出情報送信部213は、サーバ1から送信されてくるタイミング情報に従ったタイミングで、検出音情報を送信してもよい。 The evaluation information / detection information transmitting unit 213 transmits to the server 1 the evaluation information obtained by the evaluation information obtaining unit 211 and the detected sound information obtained by the detection information obtaining unit 212. The evaluation information / detection information transmitting unit 213 may transmit at least one of the evaluation information and the detected sound information collectively after the end of the broadcast of the program to be evaluated. On the other hand, the evaluation information / detection information transmitting unit 213 may transmit the evaluation information and the detected sound information while the program to be evaluated is being broadcast. For example, the evaluation information / detection information transmitting unit 213 may transmit the evaluation information at the timing when the evaluation information is input, and may repeatedly transmit the detection sound information at predetermined time intervals. For example, when a repeated sound is detected at predetermined time intervals by the microphone 28, the evaluation information / detection information transmitting unit 213 may transmit the detected sound information every time the detection is performed. By transmitting the detected sound information to the server 1 periodically, the server 1 can constantly grasp the user's viewing status of the program during the broadcast of the evaluation target program. When sound is detected by the microphone 28 at the timing when the evaluation information is input, the evaluation information / detection information transmitting unit 213 may repeatedly transmit the evaluation information together with the detected sound information at predetermined time intervals. In this case, the transmission of the evaluation information and the detected sound information is reserved until the periodic transmission timing comes after the input of the evaluation information. Alternatively, the evaluation information / detection information transmitting unit 213 may transmit the evaluation information together with the detection information at the timing when the evaluation information is input. Although details will be described later, the evaluation information / detection information transmitting unit 213 may transmit the detection sound information at a timing according to the timing information transmitted from the server 1.
 図2Bは、本実施形態に係るサーバ1のシステム制御部11の機能ブロックの一例を示す図である。システム制御部11は、CPU11aが、サーバプログラムに含まれる各種コードを読み出し実行することにより、図2Bに示すように、番組情報取得部111、評価情報・検出情報受信部112、比較部113、評価情報利用決定部114、評価部115等として機能する。 FIG. 2B is a diagram illustrating an example of functional blocks of the system control unit 11 of the server 1 according to the present embodiment. As shown in FIG. 2B, the system control unit 11 causes the CPU 11a to read and execute various codes included in the server program, thereby obtaining a program information obtaining unit 111, an evaluation information / detection information receiving unit 112, a comparing unit 113, It functions as the information use determining unit 114, the evaluation unit 115, and the like.
 番組情報取得部111は、評価対象の放送番組のコンテンツを構成する、映像及び音の少なくとも何れか一方を示すコンテンツ情報を取得する。本実施形態において、番組情報取得部111は、放送番組の音を示す番組音情報を、コンテンツ情報として取得する。例えば、評価対象の番組が事前に収録されている場合、サーバ1は、番組の放送開始前に、番組の音声データを、ネットワークNWを介して放送局3から受信してもよい。或いは、番組の音声データが記録媒体に記録されて、ドライブ装置を介してサーバ1にロードされてもよい。或いは、番組の放送中に、放送局3から発信された放送信号を図示せぬチューナが受信して、チューナにより放送信号から抽出された音声データをリアルタイムでサーバ1が取得してもよい。番組情報取得部111は、番組の音声データを番組音情報として記憶部14に記憶させてもよい。或いは、番組情報取得部111は、ユーザ端末2の検出情報取得部212と同様に、音声データから番組の音の波形情報又は特徴量を所定時間間隔で抽出してもよい。そして、番組情報取得部111は、波形情報又は特徴量の時系列で構成される時系列データを、番組音情報としてデータベース化してもよい。この番組音情報において各波形情報又は特徴量は、評価対象の番組においてこの波形情報又は特徴量に対応する音が放送される時刻と関連付けられてもよい。この放送時刻は、絶対的な時刻であってもよいし、放送開始時刻からの相対的な時刻であってもよい。 The program information acquisition unit 111 acquires content information indicating at least one of a video and a sound, which constitutes the content of the broadcast program to be evaluated. In the present embodiment, the program information acquisition unit 111 acquires program sound information indicating the sound of a broadcast program as content information. For example, when a program to be evaluated is recorded in advance, the server 1 may receive the audio data of the program from the broadcast station 3 via the network NW before the broadcast of the program starts. Alternatively, audio data of a program may be recorded on a recording medium and loaded into the server 1 via a drive device. Alternatively, a tuner (not shown) may receive a broadcast signal transmitted from the broadcast station 3 while a program is being broadcast, and the server 1 may acquire audio data extracted from the broadcast signal by the tuner in real time. The program information acquisition unit 111 may cause the storage unit 14 to store the audio data of the program as program sound information. Alternatively, the program information acquisition unit 111 may extract the waveform information or feature amount of the sound of the program from the audio data at predetermined time intervals, similarly to the detection information acquisition unit 212 of the user terminal 2. Then, the program information acquisition unit 111 may convert the time series data composed of the time series of the waveform information or the feature amount into a database as the program sound information. In the program sound information, each waveform information or feature amount may be associated with a time at which a sound corresponding to the waveform information or the feature amount is broadcasted in the program to be evaluated. This broadcast time may be an absolute time or a relative time from the broadcast start time.
 評価情報・検出情報受信部112は、各ユーザ端末2から、評価情報及び検出音情報を受信する。上述したように、評価情報と検出音情報はともに送信されてきてもよいし、別々のタイミングで送信されてきてもよい。 (4) The evaluation information / detection information receiving unit 112 receives the evaluation information and the detection sound information from each user terminal 2. As described above, the evaluation information and the detected sound information may be transmitted together, or may be transmitted at different timings.
 ここで、複数のユーザ端末2から一斉に評価情報や検出音情報が送信されてくると、サーバ1の処理負荷や通信負荷、更にはネットワーク負荷が増大する可能性がある。そこで、評価情報・検出情報受信部112は、複数のユーザ端末2のうち少なくとも一のユーザ端末2による評価情報及び検出音情報の少なくとも何れか一方の送信タイミングが、複数のユーザ端末2のうち他の少なくとも一のユーザ端末2による評価情報及び検出音情報の少なくとも何れか一方の送信タイミングと相違するように、各ユーザ端末2について、評価情報及び検出音情報の少なくとも何れか一方の送信タイミングを決定してもよい。そして、サーバ1は、決定されたタイミングを示すタイミング情報を、複数のユーザ端末2それぞれに送信してもよい。例えば、評価情報・検出情報受信部112は、複数のユーザ端末2における少なくとも一のユーザ端末2による送信タイミングの間隔の間に、他の少なくとも一のユーザ端末2による送信タイミングを決定してもよい。これにより、単位時間当たりにサーバ装置が受信する検出情報の数が均一化される。この場合、サーバ1は、各ユーザ端末2により定期的に検出音情報を送信させることになる。 Here, if the evaluation information and the detected sound information are transmitted from the plurality of user terminals 2 at the same time, the processing load, the communication load, and the network load of the server 1 may increase. Therefore, the evaluation information / detection information receiving unit 112 determines that the transmission timing of at least one of the evaluation information and the detection sound information by at least one user terminal 2 among the plurality of user terminals 2 The transmission timing of at least one of the evaluation information and the detected sound information is determined for each user terminal 2 so as to be different from the transmission timing of at least one of the evaluation information and the detected sound information by at least one user terminal 2. May be. Then, the server 1 may transmit timing information indicating the determined timing to each of the plurality of user terminals 2. For example, the evaluation information / detection information receiving unit 112 may determine the transmission timing of at least one other user terminal 2 during the interval of the transmission timing of at least one user terminal 2 among the plurality of user terminals 2. . Thereby, the number of pieces of detection information received by the server device per unit time is made uniform. In this case, the server 1 causes each user terminal 2 to periodically transmit the detected sound information.
 例えば、各ユーザ端末2による情報(評価情報及び検出音情報の少なくとも何れか一方)の送信周期をP秒とし、情報の送信タイミングをN個の送信タイミングに分散させるとする。また、番組放送開始からi番目の送信周期の開始時刻をTiとする。この場合、i番目の周期における情報の送信タイミングは、例えばTi+0秒、Ti+1×P/N秒、Ti+2×P/N秒、・・・Ti+(N-1)×P/N秒となる。0、1×P/N、2×P/N、・・・(N-1)×P/Nは、それぞれ開始時刻からのオフセットである。各ユーザ端末2は、例えば、番組の評価を開始するための操作がユーザにより行われたときに、サーバ1に対して通知を行ってもよい。評価情報・検出情報受信部112は、ユーザ端末2からの通知に応答して、複数の送信タイミングの中から一の送信タイミングを、その順序に従って循環的に又はランダムに決定してもよい。複数のユーザ端末2による送信タイミングが全体的として分散されればよいので、一部のユーザ端末2同士で送信タイミングが重なることは問題ない。評価情報・検出情報受信部112は、決定した送信タイミングを示すタイミング情報を、通知を送信してきたユーザ端末2へ送信する。ユーザ端末2の評価情報・検出情報送信部213は、タイミング情報に示されるタイミングに従って、情報を送信する。タイミング情報は、例えば送信周期の開始時刻(例えば毎分0秒等)、及びオフセット(例えば0秒、20秒、40秒等)等を含んでもよい。 For example, suppose that the transmission cycle of the information (at least one of the evaluation information and the detected sound information) by each user terminal 2 is P seconds, and the transmission timing of the information is distributed to N transmission timings. Also, the start time of the i-th transmission cycle from the start of program broadcasting is set to Ti. In this case, the information transmission timing in the i-th cycle is, for example, Ti + 0 seconds, Ti + 1 × P / N seconds, Ti + 2 × P / N seconds,... Ti + (N−1) × P / N seconds. 0, 1 × P / N, 2 × P / N,... (N−1) × P / N are offsets from the start time. For example, each user terminal 2 may notify the server 1 when an operation for starting evaluation of a program is performed by a user. In response to the notification from the user terminal 2, the evaluation information / detection information receiving unit 112 may determine one transmission timing from among a plurality of transmission timings cyclically or randomly according to the order. Since the transmission timings of the plurality of user terminals 2 need only be dispersed as a whole, there is no problem that the transmission timings of some of the user terminals 2 overlap. The evaluation information / detection information receiving unit 112 transmits timing information indicating the determined transmission timing to the user terminal 2 that has transmitted the notification. The evaluation information / detection information transmitting unit 213 of the user terminal 2 transmits information according to the timing indicated in the timing information. The timing information may include, for example, a start time of a transmission cycle (for example, 0 seconds per minute), an offset (for example, 0 seconds, 20 seconds, 40 seconds, and the like).
 図6は、複数のユーザ端末2による情報の送信タイミングの一例を示す図である。この例では、Pを60秒とし、Nを3としている。図6において、ユーザ端末2-1は、サーバ1から「毎分0秒時」を示すタイミング情報を受信する。ユーザ端末2-2は、サーバ1から「毎分0秒から20秒経過後」を示すタイミング情報を受信する。ユーザ端末2-3は、サーバ1から「毎分秒から40秒経過後」を示すタイミング情報を受信する。時刻Tで或る送信周期が開始するとする。この場合、ユーザ端末2-1は、時刻T~T+20秒の間に情報を送信する。ユーザ端末2-2は、時刻T+20~T+40秒の間に情報を送信する。ユーザ端末2-3は、時刻T+40~T+60秒の間に情報を送信する。更に、ユーザ端末2-1は、時刻T+60~T+80秒の間に情報を送信する。ユーザ端末2-2は、時刻T+80~T+100秒の間に情報を送信する。ユーザ端末2-3は、時刻T+100~T+120秒の間に情報を送信する。 FIG. 6 is a diagram illustrating an example of information transmission timings by a plurality of user terminals 2. In this example, P is 60 seconds and N is 3. In FIG. 6, the user terminal 2-1 receives timing information indicating “0 second per minute” from the server 1. The user terminal 2-2 receives timing information indicating “after 20 seconds from 0 seconds every minute” from the server 1. The user terminal 2-3 receives timing information indicating “after a lapse of 40 seconds from every minute” from the server 1. It is assumed that a certain transmission cycle starts at time T. In this case, the user terminal 2-1 transmits information during a period from time T to T + 20 seconds. The user terminal 2-2 transmits information between times T + 20 and T + 40 seconds. The user terminal 2-3 transmits information between times T + 40 and T + 60 seconds. Further, the user terminal 2-1 transmits information between times T + 60 and T + 80 seconds. The user terminal 2-2 transmits information between times T + 80 and T + 100 seconds. The user terminal 2-3 transmits information between times T + 100 and T + 120 seconds.
 比較部113は、番組情報取得部111により取得された番組音情報と、評価情報・検出情報受信部112により受信された検出音情報とを比較する。例えば番組音情報及び検出音情報が波形情報である場合、比較部113は、検出音情報に示される、ユーザ端末2により検出された音の波形情報又は特徴量と、番組音情報に示される番組の音の波形情報の時系列の各波形情報とを比較して、検出された音の波形情報と番組の音の波形情報との一致度を算出してもよい。例えば、振幅帯域ごとのサンプル数の一致及び不一致等に基づいて、一致度が算出されてもよい。比較部113は、番組音情報に示される波形情報の時系列のうち、ユーザ端末2において音が検出された時刻から前後所定時間内に放送される波形情報のみを、検出音情報と比較してもよい。番組音情報及び検出音情報が特徴量である場合も基本的には同様である。特徴量の場合の一致度は、例えばコサイン類似度等であってもよい。 The comparing unit 113 compares the program sound information acquired by the program information acquiring unit 111 with the detected sound information received by the evaluation information / detected information receiving unit 112. For example, when the program sound information and the detected sound information are waveform information, the comparing unit 113 compares the waveform information or feature amount of the sound detected by the user terminal 2 indicated by the detected sound information with the program indicated by the program sound information. The degree of coincidence between the detected waveform information of the sound and the waveform information of the sound of the program may be calculated by comparing the time-series waveform information of the waveform information of the sound. For example, the degree of coincidence may be calculated based on the coincidence and non-coincidence of the number of samples for each amplitude band. The comparing unit 113 compares only the waveform information broadcasted within a predetermined time before and after the time when the sound is detected in the user terminal 2 with the detected sound information, out of the time series of the waveform information indicated in the program sound information. Is also good. The same applies to the case where the program sound information and the detected sound information are feature amounts. The coincidence in the case of the feature amount may be, for example, a cosine similarity.
 評価対象の番組が複数存在し、それら複数の番組が同時間帯に放送される場合がある。この場合、例えばユーザが評価を行う番組を、ユーザ端末2を用いて選択可能なように、番組評価アプリケーションがプログラムされてもよい。比較部113は、選択された番組をユーザが視聴していると仮定し、選択された番組の番組音情報と、検出音情報とを比較する。或いは、比較部113は、複数の番組の番組音情報それぞれと、検出音情報とを比較して、一致度が最も高い番組を、ユーザが視聴している番組であると判定してもよい。 複数 There may be multiple programs to be evaluated and these multiple programs are broadcast in the same time zone. In this case, for example, the program evaluation application may be programmed so that the user can select a program to be evaluated using the user terminal 2. The comparison unit 113 compares the program sound information of the selected program with the detected sound information, assuming that the user is viewing the selected program. Alternatively, the comparing unit 113 may compare the program sound information of each of the plurality of programs with the detected sound information, and determine that the program with the highest degree of coincidence is the program being watched by the user.
 評価情報利用決定部114は、比較部113による比較結果により、番組音情報と検出音情報との間に所定の一致がある場合、評価情報・検出情報受信部112により受信された評価情報を、番組の評価に用いるよう選択する。例えば、評価情報利用決定部114は、図4Aに示すように、ユーザ端末2により検出された音と、番組の一部の音とが一致すると判定した場合には、評価情報を評価に用いると決定し、一致しないと判定した場合には、評価情報を評価に用いないと決定してもよい。ユーザ端末2により検出された音と、番組の一部の音とが一致する場合、ユーザは、評価対象の番組を視聴して評価を行っている可能性が高いので、評価情報の信頼性が高い。一方、それらが一致しない場合、ユーザは、評価対象の番組を視聴している可能性が低いので、評価情報の信頼性が低い。具体的に、評価情報利用決定部114は、番組音情報が、検出音情報と一致すると判定される部分を有するか否かを判定する。例えば番組音情報及び検出音情報が波形情報である場合、評価情報利用決定部114は、番組音情報に含まれる波形情報のうち、ユーザ端末2により検出された音の波形情報との一致度が所定の閾値を超える波形情報が存在するか否かを判定する。評価情報利用決定部114は、そのような波形情報が存在する場合、検出された音と番組の音とが一致すると判定してもよい。本実施形態においては、正確な音声認識は不要であり、音声波形の一致性の判定を行うことができれば、ユーザが実際に番組を視聴しているか否かの判定が可能である。番組音情報及び検出音情報が特徴量である場合も評価情報利用決定部114の処理は基本的には同様である。ユーザ端末2のマイク28により、番組の音とともに、ユーザ等の話し声や環境音等が検出される可能性がある。従って、閾値は、低めに設定されてもよい。この場合であっても、例えば所定数以上のタイミングでそれぞれ検出された音と番組の音との一致度を考慮することで、一致判定の正確度を高めることは可能である。例えば、評価情報利用決定部114は、全ての一致度が閾値を超える場合に、検出された音と番組の音とが一致すると判定してもよいし、一致度の平均値が閾値を超え且つ一致度の標準偏差が所定値未満である場合に、検出された音と番組の音とが一致すると判定してもよい。なお、ユーザ端末2のマイク28により、ユーザ等の話し声や環境音等が検出される状況においては、前処理を行いノイズとして除去してもよい。ユーザ端末2が、所定時間間隔を置いて繰り返して、マイク28による音の検出を実行して検出音情報を送信する場合、評価情報利用決定部114は、番組の音と検出された音との一致判定を所定時間間隔で実行することができる。この間隔を短くすることに従い、ユーザが評価情報を入力したときにユーザが番組を視聴しているか否かをリアルタイムに且つ的確に判定することができる。番組に対するユーザの興味は番組の進行に従って秒単位で変化し得るものであり、番組が面白くなければユーザはその番組の途中でもその番組の視聴を止め又は別の番組を視聴する。従って、番組に対する秒単位での評価情報は重要である。そのリアルタイムの評価情報の信頼性を確保することができる。 Based on the comparison result by the comparing unit 113, when there is a predetermined match between the program sound information and the detected sound information, the evaluation information use determining unit 114 converts the evaluation information received by the evaluation information / detected information receiving unit 112 into Select to use for program evaluation. For example, as shown in FIG. 4A, when it is determined that the sound detected by the user terminal 2 and a part of the sound of the program match, the evaluation information use determining unit 114 uses the evaluation information for evaluation. If it is determined and it is determined that they do not match, it may be determined that the evaluation information is not used for the evaluation. If the sound detected by the user terminal 2 and a part of the sound of the program match, the user is highly likely to view and evaluate the program to be evaluated, and the reliability of the evaluation information is low. high. On the other hand, if they do not match, it is unlikely that the user is viewing the program to be evaluated, so the reliability of the evaluation information is low. Specifically, the evaluation information use determining unit 114 determines whether or not the program sound information has a portion that is determined to match the detected sound information. For example, when the program sound information and the detected sound information are waveform information, the evaluation information use determining unit 114 determines that the degree of coincidence with the waveform information of the sound detected by the user terminal 2 among the waveform information included in the program sound information. It is determined whether or not there is waveform information exceeding a predetermined threshold. When such waveform information exists, the evaluation information use determining unit 114 may determine that the detected sound matches the sound of the program. In the present embodiment, accurate voice recognition is not required, and if it is possible to determine the consistency of voice waveforms, it is possible to determine whether the user is actually watching a program. When the program sound information and the detected sound information are feature amounts, the processing of the evaluation information use determining unit 114 is basically the same. There is a possibility that the microphone 28 of the user terminal 2 may detect a voice of the user or the like, an environmental sound, or the like, together with the sound of the program. Therefore, the threshold may be set lower. Even in this case, it is possible to increase the accuracy of the match determination by considering the degree of coincidence between the sound detected at a predetermined number or more and the sound of the program, for example. For example, the evaluation information use determination unit 114 may determine that the detected sound matches the sound of the program when all the degrees of coincidence exceed the threshold, or that the average value of the degree of coincidence exceeds the threshold and If the standard deviation of the degree of coincidence is smaller than a predetermined value, it may be determined that the detected sound matches the sound of the program. Note that in a situation where the microphone 28 of the user terminal 2 detects a voice or environmental sound of the user or the like, preprocessing may be performed to remove the noise as noise. When the user terminal 2 repeatedly performs the detection of the sound by the microphone 28 and transmits the detected sound information at predetermined time intervals, the evaluation information use determining unit 114 determines whether the sound of the program and the detected sound are different. The match determination can be performed at predetermined time intervals. By shortening the interval, it is possible to accurately determine in real time whether or not the user is watching the program when the user inputs the evaluation information. A user's interest in a program may change in seconds as the program progresses, and if the program is not interesting, the user may stop watching the program or watch another program in the middle of the program. Therefore, the evaluation information in seconds for the program is important. The reliability of the real-time evaluation information can be secured.
 評価情報利用決定部114は、ユーザにより評価情報が入力された時刻と同時刻又は近い時刻にマイク28により検出された音と、番組の音との比較に基づいて、その評価情報を番組の評価に用いるか否かを決定してもよい。評価対象の番組が放送されている間に、ユーザがその番組を視聴しているか否かが変化する場合がある。評価情報利用決定部114は、この視聴状況の変化に応じた決定を行う。図4Bは、評価情報を番組の評価に用いるか否かを決定する方法が実施される場合に生じ得るユーザ行動の一例を示す図である。図4Bに示すように、或るユーザが、例えば番組の放送が開始してから5分経過時に評価情報を入力した。この頃にマイク28により検出された音と、番組の音が一致した。この場合、その評価情報は評価に用いられる。その後、ユーザは、テレビ受像機4がある部屋から離れ、番組の放送が開始してから30分経過時に、評価情報を入力した。この頃に検出された音と、番組の音が不一致であった。この場合、その評価情報は評価に用いられない。その後、ユーザは部屋に戻り、番組の放送が開始してから50分経過時に、評価情報を入力した。この頃に検出された音と、番組の音が一致した。この場合、その評価情報は評価に用いられる。 The evaluation information use determining unit 114 evaluates the evaluation information of the program based on a comparison between the sound detected by the microphone 28 at the same time as or near the time when the evaluation information is input by the user and the sound of the program. May be determined. While the program to be evaluated is being broadcast, whether or not the user is watching the program may change. The evaluation information use determining unit 114 makes a determination according to the change in the viewing situation. FIG. 4B is a diagram illustrating an example of a user action that may occur when a method of determining whether to use evaluation information for evaluating a program is performed. As shown in FIG. 4B, a certain user inputs the evaluation information, for example, five minutes after the start of the broadcast of the program. At this time, the sound detected by the microphone 28 coincided with the sound of the program. In this case, the evaluation information is used for evaluation. Thereafter, the user leaves the room where the television receiver 4 is located, and inputs evaluation information 30 minutes after the broadcast of the program has started. The sound detected at this time did not match the sound of the program. In this case, the evaluation information is not used for evaluation. Thereafter, the user returned to the room, and input evaluation information 50 minutes after the broadcast of the program started. The sound detected around this time coincided with the sound of the program. In this case, the evaluation information is used for evaluation.
 ユーザ端末2が評価情報とともに検出音情報がサーバ1へ送信する場合、評価情報利用決定部114は、その検出音情報と番組音情報との比較に基づいて、その評価情報を番組の評価に用いるか否かを判定してもよい。ユーザ端末2は所定時間間隔を置いて繰り返し検出音情報が送信する一方で、評価情報が入力されるたびにその評価情報をサーバ1へ送信する場合、評価情報利用決定部114は、定期的に受信される検出音情報のうち、評価情報の入力時刻以前又は以後でその入力時刻に最も近い時刻にユーザ端末2のマイク28により検出される音を示す検出音情報と番組音情報との比較結果に少なくとも基づいて、その評価情報を番組の評価に用いるか否かを決定してもよい。すなわち、評価情報を評価に用いるか否かの決定に、評価情報の入力時刻以前又は以後でその入力時刻に最も近い時刻の検出情報が少なくとも用いられる。なお、評価情報が入力される時間幅と検出音情報が検出される時間幅とが少なくとも一部で重複する場合に、その評価情報を番組の評価に用いるか否かを決定してもよい。 When the user terminal 2 transmits the detected sound information to the server 1 together with the evaluation information, the evaluation information use determining unit 114 uses the evaluation information for program evaluation based on a comparison between the detected sound information and the program sound information. It may be determined whether or not there is. When the user terminal 2 repeatedly transmits the detected sound information at predetermined time intervals, and transmits the evaluation information to the server 1 each time the evaluation information is input, the evaluation information use determination unit 114 receives the evaluation information periodically. Among the detected sound information, the comparison result between the detected sound information indicating the sound detected by the microphone 28 of the user terminal 2 at the time before or after the input time of the evaluation information and the time closest to the input time and the program sound information. At least based on this, it may be determined whether or not the evaluation information is used for evaluating the program. That is, in determining whether to use the evaluation information for the evaluation, at least the detection information at the time closest to the input time before or after the input time of the evaluation information is used. When the time width during which the evaluation information is input and the time width during which the detected sound information is detected at least partially overlap, it may be determined whether or not to use the evaluation information for evaluating the program.
 ユーザ端末2は、評価情報の入力時刻を評価情報とともにサーバ1へ送信してもよいし、評価情報を送信する時刻を、評価情報の入力時刻としてサーバ1へ送信してもよい。或いは、サーバ1が、ユーザ端末2から評価情報を受信した時刻を、評価情報の入力時刻として用いてもよい。また、ユーザ端末2は、マイク28により音が検出された時刻を検出音情報とともにサーバ1へ送信してもよいし、検出音情報を送信する時刻を、音が検出された時刻としてサーバ1へ送信してもよい。或いは、サーバ1が、ユーザ端末2から検出音情報を受信した時刻を、音が検出された時刻として用いてもよい。 The user terminal 2 may transmit the input time of the evaluation information to the server 1 together with the evaluation information, or may transmit the time at which the evaluation information is transmitted to the server 1 as the input time of the evaluation information. Alternatively, the time at which the server 1 receives the evaluation information from the user terminal 2 may be used as the input time of the evaluation information. Further, the user terminal 2 may transmit the time at which the sound is detected by the microphone 28 to the server 1 together with the detected sound information, or the time at which the detected sound information is transmitted to the server 1 as the time at which the sound is detected. May be sent. Alternatively, the time at which the server 1 receives the detected sound information from the user terminal 2 may be used as the time at which the sound is detected.
 評価情報利用決定部114は、所定時間間隔を置いて繰り返し受信される検出音情報のうち、評価情報の入力時刻から相対的に近い時刻にマイク28により検出された音をそれぞれ示す2以上の検出音情報検出音情報と番組音情報との比較結果に基づいて、その評価情報を番組の評価に用いるか否かを決定してもよい。すなわち、評価情報を評価に用いるか否かの決定に、2以上の検出音情報が用いられる。これにより、ユーザが評価対象の番組を視聴しているか否かの判定精度を高めることができる。例えば、評価情報利用決定部114は、評価情報の入力時刻の前又は後、検出された時刻が入力時刻から近い順に所定数の検出音情報を用いてもよいし、検出された時刻が入力時刻から所定時間内である検出音情報を用いてもよい。或いは、評価情報利用決定部114は、評価情報の入力時刻の前後それぞれ、検出された時刻が入力時刻から近い順に所定数の検出音情報を用いてもよいし、検出された時刻が入力時刻から所定時間内である検出音情報を用いてもよい。評価情報利用決定部114は、例えば2以上の検出音情報全てについての一致度が閾値を超える場合に、評価情報を評価に用いると決定してもよいし、一致度の平均値が閾値を超え且つ一致度の標準偏差が所定値未満である場合に、評価情報を評価に用いると決定してもよい。 The evaluation information use determination unit 114 detects two or more detection sounds indicating the sounds detected by the microphone 28 at a time relatively close to the input time of the evaluation information among the detection sound information repeatedly received at predetermined time intervals. Based on the comparison result between the sound information detected sound information and the program sound information, it may be determined whether or not the evaluation information is used for evaluating the program. That is, two or more pieces of detected sound information are used to determine whether to use the evaluation information for evaluation. This makes it possible to increase the accuracy of determining whether or not the user is watching the program to be evaluated. For example, the evaluation information use determination unit 114 may use a predetermined number of pieces of detected sound information in order of the detected time being closer to the input time before or after the input time of the evaluation information, or the detected time may be the input time. Alternatively, detection sound information within a predetermined time from may be used. Alternatively, the evaluation information use determination unit 114 may use a predetermined number of pieces of detected sound information in order of the detected time being closer to the input time before and after the input time of the evaluation information, or the detected time may be changed from the input time. Detection sound information within a predetermined time may be used. The evaluation information use determining unit 114 may, for example, determine that the evaluation information is used for evaluation when the degree of coincidence for all of the two or more pieces of detected sound information exceeds the threshold, or the average value of the degree of coincidence may exceed the threshold. When the standard deviation of the degree of coincidence is smaller than a predetermined value, it may be determined that the evaluation information is used for the evaluation.
 なお、本実施形態においては、番組の音のみに基づいて、評価情報を番組の評価に用いるか否かを決定するが、評価情報利用決定部114は、番組の映像のみに基づいて、又は映像及び音の両方に基づいて、評価情報を番組の評価に用いるか否かを決定してもよい。映像を用いる場合、ユーザは、ユーザ端末2のカメラ29のレンズをテレビ受像機4に向ける。ユーザ端末2の検出情報取得部212は、例えば定期的に又は評価情報が入力されたときに、カメラ29により映像を検出させる。検出情報取得部212は、例えばカメラ29から出力される映像データから映像の特徴量を抽出して、検出映像情報を生成してもよい。例えば、SIFT(Scale Invariant Feature Transform)、SURF(Speeded Up Robust Features)等のアルゴリズムを用いて特徴量が抽出されてもよい。評価情報・検出情報送信部213は、検出映像情報をサーバ1へ送信する。サーバ1の番組情報取得部111は、放送局からの放送信号をチューナにより受信させ又はネットワークNWを介して映像データを取得して、番組の映像の特徴量を抽出して、番組映像情報を生成してもよい。比較部113は、番組映像情報と検出映像情報とを比較し、評価情報利用決定部114は、この比較に基づいて、評価情報を番組の評価に用いるか否かを決定する。映像を用いる場合の詳細及び変形例は、音を用いる場合と同様であってもよい。映像及び音の両方を用いる場合、評価情報利用決定部114は、例えば映像が一致し且つ音が一致すると判定した場合に、評価情報を番組の評価に用いると決定してもよい。 In the present embodiment, whether or not to use the evaluation information for evaluating the program is determined based on only the sound of the program. However, the evaluation information use determining unit 114 determines whether or not to use the evaluation information based on only the video of the program. Whether to use the evaluation information for evaluating the program may be determined based on both the sound and the sound. When using a video, the user points the lens of the camera 29 of the user terminal 2 to the television receiver 4. The detection information acquisition unit 212 of the user terminal 2 causes the camera 29 to detect an image, for example, periodically or when evaluation information is input. The detection information obtaining unit 212 may generate the detected video information by extracting the feature amount of the video from the video data output from the camera 29, for example. For example, the feature amount may be extracted using an algorithm such as SIFT (Scale Invariant Feature Transform) or SURF (Speeded Up Robust Feature). The evaluation information / detection information transmitting unit 213 transmits the detected video information to the server 1. The program information acquisition unit 111 of the server 1 causes a tuner to receive a broadcast signal from a broadcast station or acquires video data via a network NW, extracts a feature amount of a video of a program, and generates program video information. May be. The comparing unit 113 compares the program video information with the detected video information, and the evaluation information use determining unit 114 determines whether to use the evaluation information for evaluating the program based on the comparison. The details and modifications when using video may be the same as when using sound. When both video and sound are used, the evaluation information use determination unit 114 may determine that the evaluation information is used for evaluating the program, for example, when it is determined that the video matches and the sound matches.
 評価部115は、評価情報・検出情報受信部112により受信された評価情報のうち、評価情報利用決定部114により、用いると決定された評価情報に基づいて、番組を評価する処理を実行する。例えば、評価情報・検出情報受信部112は、評価情報を集計したり分析したりする。評価部115は、「つまらない」、「おもしろい」等の各項目についての評価の累計や、番組放送開始から終了までの間の各時刻に入力された評価数を計算してもよい。また、評価部115は、全体としての評価の累計や、各時刻に入力された評価数をカウントしてもよい。評価部115は、評価数の推移を示す情報や、評価数が多い時刻のランキングを示す情報等を生成してもよい。また、評価部115は、番組の評価に参加したユーザの人数や、少なくとも一の評価情報について評価情報利用決定部114により番組の評価に用いられると判定されたユーザの人数(有効な評価を少なくとも1回行ったユーザの人数)をカウントしてもよい。また、評価部115は、有効な評価を行ったユーザの属性の分布を生成してもよい。また、評価部115は、コメントのリストを生成してもよい。評価部115は、番組の評価結果としてレポートを生成してもよい。図7は、生成されたレポートの一例を示す図である。レポートの形式は、例えばHTML(HyperText Markup Language)、PDF(Portable Document Format)等であってもよい。サーバ1は、例えば放送局3における図示せぬ端末装置からの要求に応じて、生成されたレポートをネットワークNWを介して送信する。なお、番組を評価する処理は、例えば番組評価システムSの管理者が利用する端末装置等が実行してもよい。また、評価部115は、評価を行ったユーザのIDに対して、ポイント等の特典を付与してもよい。これにより、ユーザのモチベーションアップを利用して評価情報を多く集めることができ、母集団を大きくすることで評価結果の信頼性を高めることができる。 (4) The evaluation unit 115 executes a process of evaluating a program based on the evaluation information determined to be used by the evaluation information use determining unit 114 among the evaluation information received by the evaluation information / detection information receiving unit 112. For example, the evaluation information / detection information receiving unit 112 counts and analyzes the evaluation information. The evaluation unit 115 may calculate the total evaluation of each item such as “uninteresting” and “interesting”, and the number of evaluations input at each time from the start to the end of the broadcast of the program. In addition, the evaluation unit 115 may count the total number of evaluations as a whole or the number of evaluations input at each time. The evaluation unit 115 may generate information indicating a transition of the number of evaluations, information indicating a ranking at a time when the number of evaluations is large, and the like. In addition, the evaluation unit 115 determines the number of users who have participated in the evaluation of the program, and the number of users who have been determined by the evaluation information use determination unit 114 to use at least one piece of evaluation information (the number of effective evaluations is at least one). (The number of users who performed once) may be counted. In addition, the evaluation unit 115 may generate a distribution of the attributes of the user who has performed the effective evaluation. Further, the evaluation unit 115 may generate a list of comments. The evaluation unit 115 may generate a report as a program evaluation result. FIG. 7 is a diagram illustrating an example of the generated report. The format of the report may be, for example, HTML (HyperText Markup Language), PDF (Portable Document Format), or the like. The server 1 transmits the generated report via the network NW, for example, in response to a request from a terminal device (not shown) in the broadcasting station 3. The process of evaluating a program may be executed by, for example, a terminal device used by a manager of the program evaluation system S. In addition, the evaluation unit 115 may give a privilege such as a point to the ID of the user who has performed the evaluation. As a result, a large amount of evaluation information can be collected by using the motivation of the user, and the reliability of the evaluation result can be increased by increasing the population.
[1-5.番組評価システムの動作] [1-5. Operation of the program evaluation system]
 次に、番組評価システムSの動作について、図8及び9を用いて説明する。以下に説明する動作例において、サーバ1は、番組音情報を予め取得して記憶部14に記憶させているものとする。ユーザ端末2は、音を定期的に検出して検出音情報をサーバ1へ送信するものとする。サーバ1は、評価情報の受信時刻以前でその受信時刻から最も近い時刻に受信された検出音情報を用いて、その評価情報を番組の評価に用いるか否かを決定するものとする。番組音情報及び検出音情報として、波形情報が用いられるものとする。 Next, the operation of the program evaluation system S will be described with reference to FIGS. In the operation example described below, it is assumed that the server 1 acquires program sound information in advance and stores it in the storage unit 14. It is assumed that the user terminal 2 periodically detects a sound and transmits detected sound information to the server 1. The server 1 uses the detected sound information received before the evaluation information reception time and at the time closest to the reception time, and determines whether or not to use the evaluation information for program evaluation. It is assumed that waveform information is used as the program sound information and the detected sound information.
 図8は、ユーザ端末2のシステム制御部21により実行される端末処理の一例を示すフローチャートである。例えば、ユーザは、番組評価アプリケーションを起動して番組調査の開始操作を行う。これに応じて、システム制御部21は、番組評価アプリケーションに従って端末処理を実行する。 FIG. 8 is a flowchart illustrating an example of a terminal process executed by the system control unit 21 of the user terminal 2. For example, the user activates the program evaluation application and performs a program survey start operation. In response, the system control unit 21 executes terminal processing according to the program evaluation application.
 先ず、評価情報・検出情報送信部213は、調査開始通知を、ユーザ端末2を利用するユーザのユーザIDとともにサーバ1へ送信する(ステップS1)。次いで、評価情報・検出情報送信部213は、サーバ1からタイミング情報を受信してRAM21cに記憶させる(ステップS2)。 First, the evaluation information / detection information transmitting unit 213 transmits a survey start notification to the server 1 together with the user ID of the user who uses the user terminal 2 (step S1). Next, the evaluation information / detection information transmission unit 213 receives the timing information from the server 1 and stores it in the RAM 21c (Step S2).
 次いで、検出情報取得部212は、現在時刻に基づいて、音の検出タイミングが到来したか否かを判定する(ステップS3)。例えば、検出情報取得部212は、タイミング情報が示す検出音情報の送信タイミング、及び音の検出を継続させる時間等に基づいて、送信タイミングに間に合うように検出タイミングを決定する。検出情報取得部212は、検出タイミングが到来したと判定した場合には(ステップS2:YES)、処理をステップS4に進める。ステップS4において、検出情報取得部212は、マイク28により音を検出させる。次いで、検出情報取得部212は、マイク28から出力される音声信号から波形情報を検出音情報として抽出する(ステップS5)。 Next, the detection information acquisition unit 212 determines whether or not the sound detection timing has arrived based on the current time (step S3). For example, the detection information acquisition unit 212 determines the detection timing in time for the transmission timing based on the transmission timing of the detected sound information indicated by the timing information, the time for continuing the sound detection, and the like. If the detection information acquisition unit 212 determines that the detection timing has come (step S2: YES), the process proceeds to step S4. In step S4, the detection information acquisition unit 212 causes the microphone 28 to detect a sound. Next, the detection information acquisition unit 212 extracts waveform information from the audio signal output from the microphone 28 as detection sound information (step S5).
 ステップS3において、検出情報取得部212は、検出タイミングが到来していないと判定した場合には(ステップS2:NO)、処理をステップS6に進める。ステップS6において、評価情報・検出情報送信部213は、現在時刻に基づいて、タイミング情報が示す送信タイミングが到来したか否かを判定する。評価情報・検出情報送信部213は、送信タイミングが到来したと判定した場合には(ステップS6:YES)、処理をステップS7に進める。ステップS7において、評価情報・検出情報送信部213は、RAM21cに記憶された検出音情報をユーザIDとともにサーバ1へ送信する。 In step S3, when the detection information acquisition unit 212 determines that the detection timing has not arrived (step S2: NO), the detection information acquisition unit 212 proceeds with the process to step S6. In step S6, the evaluation information / detection information transmitting unit 213 determines whether the transmission timing indicated by the timing information has arrived based on the current time. When it is determined that the transmission timing has arrived (step S6: YES), the evaluation information / detection information transmitting unit 213 advances the process to step S7. In step S7, the evaluation information / detection information transmission unit 213 transmits the detection sound information stored in the RAM 21c to the server 1 together with the user ID.
 ステップS6において、評価情報・検出情報送信部213は、送信タイミングが到来していないと判定した場合には(ステップS6:YES)、処理をステップS8に進める。ステップS8において、評価情報取得部211は、操作入力部26からの信号に基づいて、評価情報が入力されたか否かを判定する。評価情報取得部211は、評価情報が入力されたと判定した場合には(ステップS8:YES)、処理をステップS9に進める。ステップS9において、評価情報取得部211は、入力された評価情報を、ユーザIDとともにサーバ1へ送信する。 評 価 In step S6, when the evaluation information / detection information transmitting unit 213 determines that the transmission timing has not arrived (step S6: YES), the process proceeds to step S8. In step S8, the evaluation information acquisition unit 211 determines whether evaluation information has been input based on a signal from the operation input unit 26. When determining that the evaluation information has been input (step S8: YES), the evaluation information acquisition unit 211 advances the processing to step S9. In step S9, the evaluation information acquisition unit 211 transmits the input evaluation information to the server 1 together with the user ID.
 ステップS5、S7若しくはS9を終えた場合、又はステップS8において評価情報が入力されなかったと判定された場合(ステップS8:NO)、システム制御部21は、評価対象の番組の放送終了時刻が到来したか否かを判定する(ステップS10)。システム制御部21は、終了時刻が到来していないと判定した場合には(ステップS10:NO)、処理をステップS3に進める。一方、システム制御部21は、終了時刻が到来したと判定した場合には(ステップS10:YES)、端末処理を終了させる。 When step S5, S7 or S9 is completed, or when it is determined in step S8 that the evaluation information has not been input (step S8: NO), the system control unit 21 arrives at the broadcast end time of the program to be evaluated. It is determined whether or not (Step S10). If it is determined that the end time has not arrived (step S10: NO), the system control unit 21 advances the processing to step S3. On the other hand, when it is determined that the end time has arrived (step S10: YES), the system control unit 21 ends the terminal processing.
 図9は、サーバ1のシステム制御部11により実行されるサーバ処理の一例を示すフローチャートである。システム制御部11は、サーバプログラムに従って、例えば評価対象の番組の放送開始から所定時間前にサーバ処理を開始させる。 FIG. 9 is a flowchart illustrating an example of a server process executed by the system control unit 11 of the server 1. The system control unit 11 starts the server processing, for example, a predetermined time before the start of the broadcast of the program to be evaluated, according to the server program.
 先ず、評価情報・検出情報受信部112は、何れかのユーザ端末2から調査開始通知を受信したか否かを判定する(ステップS21)。評価情報・検出情報受信部112は、調査開始通知を受信したと判定した場合には(ステップS21:YES)、処理をステップS22に進める。ステップS22において、評価情報・検出情報受信部112は、調査開始通知とともに受信されたユーザIDに対応付けて、FALSEに設定された視聴フラグをRAM11cに記憶させる。視聴フラグは、ユーザが評価対象の番組を視聴しているか否かを示す情報である。次いで、評価情報・検出情報受信部112は、予め定められた複数の送信タイミングの中から何れか1つの送信タイミングを、例えばランダムに決定する(ステップS23)。評価情報・検出情報受信部112は、決定された送信タイミングを示すタイミング情報を、調査開始通知を送信してきたユーザ端末2へ送信する(ステップS24)。 First, the evaluation information / detection information receiving unit 112 determines whether a survey start notification has been received from any of the user terminals 2 (step S21). When the evaluation information / detection information receiving unit 112 determines that the investigation start notification has been received (step S21: YES), the process proceeds to step S22. In step S22, the evaluation information / detection information receiving unit 112 stores the viewing flag set to FALSE in the RAM 11c in association with the user ID received together with the survey start notification. The viewing flag is information indicating whether or not the user is viewing the program to be evaluated. Next, the evaluation information / detection information receiving unit 112 determines any one of a plurality of predetermined transmission timings, for example, at random (Step S23). The evaluation information / detection information receiving unit 112 transmits timing information indicating the determined transmission timing to the user terminal 2 that has transmitted the investigation start notification (step S24).
 ステップS21において、評価情報・検出情報受信部112は、調査開始通知を受信しなかったと判定した場合には(ステップS21:NO)、処理をステップS25に進める。ステップS25において、評価情報・検出情報受信部112は、何れかのユーザ端末2から検出音情報を受信したか否かを判定する。評価情報・検出情報受信部112は、検出音情報を受信したと判定した場合には(ステップS25:YES)、処理をステップS26に進める。ステップS26において、比較部113は、受信された検出音情報に示される波形情報と、番組情報に含まれる各波形情報との一致度を算出する。比較部113は、番組情報に含まれるに含まれる波形情報のうち、算出された一致度が閾値を超えるものがあるか否かを判定する。比較部113は、一致度が閾値を超える波形情報があると判定した場合には(ステップS26:YES)、検出音情報とともに受信されたユーザIDに対応付けられた視聴フラグをTRUEに設定する(ステップS27)。一方、比較部113は、一致度が閾値を超える波形情報がないと判定した場合には(ステップS26:NO)、検出音情報とともに受信されたユーザIDに対応付けられた視聴フラグをFALSEに設定する(ステップS28)。 In step S21, when the evaluation information / detection information receiving unit 112 determines that the investigation start notification has not been received (step S21: NO), the process proceeds to step S25. In step S25, the evaluation information / detection information receiving unit 112 determines whether the detection sound information has been received from any of the user terminals 2. When the evaluation information / detection information receiving unit 112 determines that the detection sound information has been received (step S25: YES), the process proceeds to step S26. In step S26, the comparing unit 113 calculates the degree of coincidence between the waveform information indicated in the received detected sound information and each piece of waveform information included in the program information. The comparing unit 113 determines whether or not there is any of the pieces of waveform information included in the program information whose calculated coincidence exceeds a threshold. When the comparing unit 113 determines that there is waveform information whose degree of coincidence exceeds the threshold (step S26: YES), it sets the viewing flag associated with the user ID received with the detected sound information to TRUE (step S26). Step S27). On the other hand, when the comparing unit 113 determines that there is no waveform information whose matching degree exceeds the threshold value (step S26: NO), the viewing flag associated with the user ID received together with the detected sound information is set to FALSE. (Step S28).
 ステップS25において、評価情報・検出情報受信部112は、検出音情報を受信しなかったと判定した場合には(ステップS25:NO)、処理をステップS29に進める。ステップS29において、評価情報・検出情報受信部112は、何れかのユーザ端末2から評価情報を受信したか否かを判定する。評価情報・検出情報受信部112は、評価情報を受信したと判定した場合には(ステップS29:YES)、処理をステップS30に進める。ステップS30において、評価情報利用決定部114は、評価情報とともに受信されたユーザIDに対応付けられた視聴フラグがTRUEであるか否かを判定する。評価情報利用決定部114は、視聴フラグがTRUEであると判定した場合には(ステップS30:YES)、処理をステップS31に進める。ステップS31において、評価情報利用決定部114は、受信された評価情報の入力時刻を、評価情報の受信時刻に設定する。次いで、評価情報利用決定部114は、評価情報、入力時刻及びユーザIDを関連付けて記憶部14に記憶させる(ステップS32)。一方、評価情報利用決定部114は、視聴フラグがFALSEであると判定した場合には(ステップS30:NO)、受信した評価情報を破棄する(ステップS33)。 In step S25, when the evaluation information / detection information receiving unit 112 determines that the detection sound information has not been received (step S25: NO), the process proceeds to step S29. In step S29, the evaluation information / detection information receiving unit 112 determines whether evaluation information has been received from any of the user terminals 2. When the evaluation information / detection information receiving unit 112 determines that the evaluation information has been received (step S29: YES), the process proceeds to step S30. In step S30, the evaluation information use determining unit 114 determines whether the viewing flag associated with the user ID received along with the evaluation information is TRUE. When the evaluation information use determining unit 114 determines that the viewing flag is TRUE (step S30: YES), the process proceeds to step S31. In step S31, the evaluation information use determining unit 114 sets the input time of the received evaluation information to the reception time of the evaluation information. Next, the evaluation information use determining unit 114 stores the evaluation information, the input time, and the user ID in the storage unit 14 in association with each other (Step S32). On the other hand, when it is determined that the viewing flag is FALSE (step S30: NO), the evaluation information use determining unit 114 discards the received evaluation information (step S33).
 ステップS24、S27、S28、S32又はS33を終えたとき、又はステップS29において評価情報が受信されなかったと判定されたとき(ステップS29:NO)、評価部115は、評価対象の番組の放送終了時刻が到来したか否かを判定する(ステップS34)。評価部115は、終了時刻が到来していないと判定した場合には(ステップS34:NO)、処理をステップS21に進める。一方、評価部115は、終了時刻が到来したと判定した場合には(ステップS34:YES)、評価情報を用いて番組の評価処理を実行する(ステップS35)。評価部115は、記憶部14に記憶された評価情報を集計、分析する。例えば、評価部115は、入力時刻を用いて、評価項目ごと評価数の推移等を算出したり、評価数の多い時刻のランキングを決定したりしてもよい。評価部115は、評価結果を示すレポートを生成して記憶部14に記憶させる。評価処理を終えると、評価部115は、サーバ処理を終了させる。 When step S24, S27, S28, S32 or S33 is completed, or when it is determined in step S29 that the evaluation information has not been received (step S29: NO), the evaluation unit 115 determines the broadcast end time of the program to be evaluated. Is determined (step S34). When determining that the end time has not arrived (step S34: NO), the evaluation unit 115 advances the processing to step S21. On the other hand, when the evaluation unit 115 determines that the end time has arrived (step S34: YES), the evaluation unit 115 executes a program evaluation process using the evaluation information (step S35). The evaluation unit 115 counts and analyzes the evaluation information stored in the storage unit 14. For example, the evaluation unit 115 may use the input time to calculate the transition of the number of evaluations for each evaluation item, or determine the ranking of the time with the highest number of evaluations. The evaluation unit 115 generates a report indicating the evaluation result, and causes the storage unit 14 to store the report. When the evaluation processing ends, the evaluation unit 115 ends the server processing.
 以上説明したように、本実施形態によれば、ユーザ端末2が、番組を構成する映像及び音を出力するテレビ受像機4により出力された音を検出する。また、ユーザ端末2が、ユーザにより入力された評価情報と、検出された音を示す検出音情報と、をサーバ1へ送信する。サーバ1が、番組を構成する音を示す番組音情報を取得する。また、サーバ1が、ユーザ端末2から評価情報及び検出音情報を受信する。また、サーバ1が、取得された番組音情報と受信された検出音情報とを比較する。また、サーバ1が、この比較結果により、番組音情報と受信された検出音情報との間に所定の一致がある場合、受信された評価情報を番組の評価に用いるよう選択する。そのため、評価対象の番組を実際にユーザが視聴しているか否かを推定することが可能となる。従って、ユーザが番組を視聴していると推定される評価情報のみを、番組の評価に用いることができる。そのため、評価情報の信頼性を確保することが可能となる。 As described above, according to the present embodiment, the user terminal 2 detects the sound output by the television receiver 4 that outputs the video and the sound constituting the program. Also, the user terminal 2 transmits the evaluation information input by the user and the detected sound information indicating the detected sound to the server 1. The server 1 acquires program sound information indicating the sound that constitutes the program. Further, the server 1 receives the evaluation information and the detected sound information from the user terminal 2. Further, the server 1 compares the acquired program sound information with the received detected sound information. If the server 1 determines from the comparison result that there is a predetermined match between the program sound information and the received detected sound information, the server 1 selects the received evaluation information to be used for evaluating the program. Therefore, it is possible to estimate whether the user is actually watching the program to be evaluated. Therefore, only the evaluation information that is estimated that the user is watching the program can be used for evaluating the program. Therefore, it is possible to ensure the reliability of the evaluation information.
 また、コンテンツは、放送番組のコンテンツであってもよい。また、サーバ1が、番組の放送中における複数のユーザ端末2のうち少なくとも一のユーザ端末2による評価情報及び検出音情報の少なくとも何れか一方の送信タイミングが、複数のユーザ端末2のうち他の少なくとも一のユーザ端末2による送信タイミングと相違するように、複数のユーザ端末2それぞれについて、送信タイミングを決定してもよい。また、サーバ1が、決定された送信タイミングを示すタイミング情報を、複数のユーザ端末2それぞれに送信してもよい。各ユーザ端末2が、サーバ1からタイミング情報を受信してもよい。また、各ユーザ端末2が、受信されたタイミング情報に示される送信タイミングに従って、評価情報及び検出音情報の少なくとも何れか一方を送信してもよい。この場合、番組の放送中に複数のユーザ端末2それぞれからサーバ1へ検出音情報が送信される態様であっても、評価情報及び検出音情報の少なくとも何れか一方の送信タイミングが複数に分散される。従って、サーバ1の処理負荷が同時点に集中することを抑制し負荷分散させることができる。 コ ン テ ン ツ Also, the content may be a content of a broadcast program. In addition, the server 1 determines that at least one of the evaluation information and the detected sound information transmitted by at least one of the user terminals 2 among the plurality of user terminals 2 during the broadcast of the program is transmitted to another of the plurality of user terminals 2 The transmission timing may be determined for each of the plurality of user terminals 2 so as to be different from the transmission timing by at least one user terminal 2. Further, the server 1 may transmit timing information indicating the determined transmission timing to each of the plurality of user terminals 2. Each user terminal 2 may receive timing information from the server 1. Further, each user terminal 2 may transmit at least one of the evaluation information and the detected sound information according to the transmission timing indicated in the received timing information. In this case, even when the detected sound information is transmitted from each of the plurality of user terminals 2 to the server 1 during the broadcast of the program, the transmission timing of at least one of the evaluation information and the detected sound information is dispersed to a plurality. You. Therefore, it is possible to suppress the processing load of the server 1 from being concentrated at the same point and to distribute the load.
 また、サーバ1が、複数のユーザ端末2における少なくとも一のユーザ端末2による送信タイミングの間隔の間に、他の少なくとも一のユーザ端末2による送信タイミングを決定してもよい。この場合、単位時間当たりにサーバ1が受信する検出情報の数が均一化されるので、サーバ装置の処理負荷を更に分散させることができる。 The server 1 may determine the transmission timing of at least one other user terminal 2 during the interval of the transmission timing of at least one user terminal 2 among the plurality of user terminals 2. In this case, the number of pieces of detection information received by the server 1 per unit time is made uniform, so that the processing load on the server device can be further dispersed.
 また、ユーザ端末2が、所定時間間隔を置いて繰り返し、音を検出して検出音情報を送信してもよい。サーバ1は、所定時間間隔を置いて繰り返し受信される検出音情報のうち、評価情報の入力時刻以前又は以後で入力時刻に最も近い時刻にユーザ端末2により検出される音を示す検出音情報と番組音情報との比較結果に少なくとも基づいて、評価情報を番組の評価に用いるか否かを決定してもよい。この場合、ユーザ端末2が音の検出及び検出音情報の送信を定期的に実行する態様においても、評価情報を入力した時点においてユーザが番組を視聴しているか否かの推定精度を高めることができる。 (4) The user terminal 2 may repeatedly detect the sound at predetermined time intervals and transmit the detected sound information. The server 1 includes, among the detected sound information repeatedly received at predetermined time intervals, detected sound information indicating a sound detected by the user terminal 2 at a time before or after the input time of the evaluation information and closest to the input time. Whether to use the evaluation information for evaluating the program may be determined based at least on the result of comparison with the program sound information. In this case, even in a mode in which the user terminal 2 periodically performs sound detection and transmission of the detected sound information, it is possible to increase the estimation accuracy of whether or not the user is watching the program at the time when the evaluation information is input. it can.
 また、サーバ1が、定期的に受信される検出音情報のうち、評価情報の入力時刻から相対的に近い時刻に検出された音をそれぞれ示す2以上の検出音情報と番組音情報との比較結果に基づいて、評価情報を番組の評価に用いるか否かを決定してもよい。この場合、評価情報を入力した時点においてユーザが番組を視聴しているか否かの推定精度を更に高めることができる。 In addition, the server 1 compares the program sound information with two or more pieces of detected sound information indicating sounds detected at a time relatively close to the input time of the evaluation information among the detected sound information received periodically. Based on the result, it may be determined whether to use the evaluation information for evaluating the program. In this case, the accuracy of estimating whether or not the user is watching the program at the time when the evaluation information is input can be further increased.
 また、ユーザ端末2が、評価情報が入力されたときに音を検出してもよい。サーバ1は、番組音情報が、検出音情報と一致すると判定される部分を有する場合、検出音情報とともに受信された評価情報を、番組の評価に用いると決定してもよい。この場合、評価情報を入力した時点においてユーザが番組を視聴しているか否かの推定精度を高めることができる。 The user terminal 2 may detect a sound when the evaluation information is input. When the program sound information includes a portion determined to match the detected sound information, the server 1 may determine that the evaluation information received together with the detected sound information is used for evaluating the program. In this case, the accuracy of estimating whether the user is watching the program at the time when the evaluation information is input can be improved.
[2.第2実施形態] [2. Second Embodiment]
[2-1.機能概要] [2-1. Functional overview]
 次に、図10を用いて、第2実施形態におけるサーバ1のシステム制御部11、及びユーザ端末2のシステム制御部21の機能概要を説明する。本実施形態において、ユーザ端末2は、評価情報が入力されたタイミングで音を検出して検出音情報を取得する。評価情報及び検出音情報の送信タイミングは、検出音情報が検出された時点であってもよいし、定期的であってもよい。 Next, an outline of functions of the system control unit 11 of the server 1 and the system control unit 21 of the user terminal 2 in the second embodiment will be described with reference to FIG. In the present embodiment, the user terminal 2 detects sound at the timing when the evaluation information is input, and acquires the detected sound information. The transmission timing of the evaluation information and the detected sound information may be a point in time when the detected sound information is detected, or may be a regular period.
 サーバ1は、検出音情報に基づいて、ユーザ端末2に評価情報が入力された時刻を特定する。評価情報利用決定部114は、番組音情報が、検出音情報と一致すると判定される部分を有する場合、ユーザ端末2からその検出音情報とともに受信された評価情報を番組の評価に用いると決定する。このとき、評価情報利用決定部114は、評価対象の番組において検出情報と一致すると判定された部分により示される音のテレビ受像機4による出力時刻(放送時刻)を、その検出音情報とともに受信された評価情報の入力時刻として特定してもよい。この時刻は、絶対的な時刻であってもよいし、放送開始時刻からの相対的な時刻であってもよい。これにより、正確度の高い入力時刻を用いて、番組の評価を行うことが可能である。この入力時刻特定方法は、配信時間帯が予め定められていないオンデマンド配信にも有効である。 The server 1 specifies the time at which the evaluation information was input to the user terminal 2 based on the detected sound information. When the program sound information has a portion determined to match the detected sound information, the evaluation information use determining unit 114 determines that the evaluation information received together with the detected sound information from the user terminal 2 is used for evaluating the program. . At this time, the evaluation information use determining unit 114 receives the output time (broadcast time) of the sound indicated by the portion determined to match the detection information in the program to be evaluated by the television receiver 4 together with the detected sound information. May be specified as the input time of the evaluation information. This time may be an absolute time or a relative time from the broadcast start time. As a result, it is possible to evaluate the program using the input time with high accuracy. This input time specifying method is also effective for on-demand distribution in which the distribution time zone is not predetermined.
 サーバ1における一致判定には、例えば音の特徴量が用いられてもよいが、この特徴量を特定することが可能な特徴量特定情報が用いられてもよい。この特徴量特定情報は、特徴量よりも情報量が少ない情報である。この特徴量特定情報は、少なくとも一の番組において、基本的に特徴量ごとに異なるものである。評価の対象となる番組が複数存在する場合、番組間においても特徴量特定情報が異なることが好ましい。特徴量特定情報は、例えば特徴量の要約を示すハッシュ値であってもよいし、所定の基準に基づいて特徴量に付与される識別情報であってもよい。或いは、特徴量特定情報は、特徴量に対応する音の放送時刻を含んでもよい。なお、特徴量及び特徴量特定情報の代わりに、波形情報及びこの波形情報を特定することが可能な情報であって且つ波形情報よりも情報量が少ない情報が用いられてもよい。 For the match determination in the server 1, for example, a feature value of a sound may be used, but feature value specifying information capable of specifying the feature value may be used. This feature amount specifying information is information having an information amount smaller than the feature amount. This characteristic amount specifying information is basically different for each characteristic amount in at least one program. When there are a plurality of programs to be evaluated, it is preferable that the feature amount specifying information differs between the programs. The characteristic amount specifying information may be, for example, a hash value indicating a summary of the characteristic amount, or may be identification information given to the characteristic amount based on a predetermined criterion. Alternatively, the characteristic amount specifying information may include a broadcast time of a sound corresponding to the characteristic amount. Instead of the characteristic amount and the characteristic amount specifying information, waveform information and information capable of specifying the waveform information and information having a smaller information amount than the waveform information may be used.
 以降においては、特徴量特定情報として特徴量のハッシュ値を用いた場合について説明する。図10は、番組評価システムSにおける処理概要の一例を示す図である。番組情報取得部111は、評価対象の番組の放送が開始される以前に、ネットワークNWを介して取得された番組の音声データから、番組の音の特徴量の時系列で構成される特徴量時系列データを生成する。言い換えれば、番組の音声データを複数に区分し時系列に並べたものを各々特徴量に変換した状態で記憶部24に記憶しておく。また、番組情報取得部111は、生成された特徴量時系列データを特定情報時系列データに変換する。例えば、番組情報取得部111は、特徴量時系列データの各特徴量のハッシュ値を、所定のハッシュ関数により生成する。そして、番組情報取得部111は、ハッシュ値の時系列で構成される特定情報時系列データを番組音情報として生成して記憶部14に記憶させる。この番組音情報において各ハッシュ値は、評価対象の番組においてこのハッシュ値に対応する音が放送される時刻と関連付けられてもよい。 Hereinafter, the case where the hash value of the feature amount is used as the feature amount specifying information will be described. FIG. 10 is a diagram illustrating an example of a processing outline in the program evaluation system S. Before the broadcast of the program to be evaluated is started, the program information acquisition unit 111 extracts a feature time from the audio data of the program acquired via the network NW in the time series of the feature amount of the sound of the program. Generate series data. In other words, the audio data of the program is divided into a plurality of pieces and those arranged in chronological order are stored in the storage unit 24 in a state where they are each converted into a feature amount. Further, the program information acquisition unit 111 converts the generated feature amount time series data into specific information time series data. For example, the program information acquisition unit 111 generates a hash value of each feature amount of the feature amount time-series data using a predetermined hash function. Then, the program information acquisition unit 111 generates specific information time-series data composed of a time series of hash values as program sound information and causes the storage unit 14 to store the information. In the program sound information, each hash value may be associated with a time at which a sound corresponding to the hash value is broadcast in the program to be evaluated.
 ユーザ端末2は、事前に、特徴量時系列データを、番組音特徴量情報として取得する。例えば、ユーザは、評価対象の番組の放送が開始されるよりも前にユーザ端末2を操作して、番組調査に予めエントリーする。サーバ1は、ユーザ端末2からのエントリーの通知を受信すると、番組音特徴量情報を送信する。ユーザ端末2は、番組音特徴量情報を記憶部24に記憶させる。 (4) The user terminal 2 acquires the feature amount time-series data as program sound feature amount information in advance. For example, the user operates the user terminal 2 before the broadcast of the program to be evaluated starts, and preliminarily enters a program survey. Upon receiving the entry notification from the user terminal 2, the server 1 transmits the program sound feature information. The user terminal 2 causes the storage unit 24 to store the program sound feature information.
 評価対象の番組の放送中に、ユーザは、評価情報をユーザ端末2に入力する。このとき、検出情報取得部212は、マイク28により音を検出させて、マイク28から音声信号を受信する。検出情報取得部212は、音声信号から特徴量を抽出して、この特徴量と番組音特徴量情報における各特徴量とを比較して、一致度を算出する。検出情報取得部212は、番組音特徴量情報の中に、一致度が所定の閾値を超える特徴量が存在する場合、その特徴量のハッシュ値を、サーバ1で用いられるハッシュ関数と同一のハッシュ関数(例えば番組評価アプリケーションに含まれている。)により生成する。検出情報取得部212は、このハッシュ値を、検出音情報として取得する。評価情報・検出情報送信部213は、検出音情報としてのハッシュ値を、評価情報とともにサーバ1へ送信する。 (4) During the broadcast of the program to be evaluated, the user inputs the evaluation information to the user terminal 2. At this time, the detection information acquisition unit 212 causes the microphone 28 to detect a sound, and receives an audio signal from the microphone 28. The detection information acquisition unit 212 extracts a feature amount from the audio signal, compares the feature amount with each feature amount in the program sound feature amount information, and calculates a degree of coincidence. When the program sound feature amount information includes a feature amount having a matching degree exceeding a predetermined threshold, the detection information acquisition unit 212 converts the hash value of the feature amount into the same hash function as the hash function used in the server 1. Generated by a function (for example, included in the program evaluation application). The detection information acquisition unit 212 acquires this hash value as detection sound information. The evaluation information / detection information transmitting unit 213 transmits a hash value as detected sound information to the server 1 together with the evaluation information.
 サーバ1の比較部113は、ユーザ端末2から受信された検出音情報と、番組音情報としてのハッシュ値の時系列の各ハッシュ値とを比較する。比較の結果、評価情報・検出情報送信部213は、ハッシュ値の時系列中に検出音情報と一致するハッシュ値が存在する場合、評価情報を番組の評価に用いると決定する。評価情報・検出情報送信部213は、評価対象の番組において、検出音情報と一致するハッシュ値に対応する音が放送される時刻を、評価情報の入力時刻として特定する。ハッシュ値等の特徴量特定情報を用いることにより、評価に用いられる入力時刻の正確度が高まるとともに、番組の放送中におけるサーバ1とユーザ端末2との間の通信量を削減することができる。この通信量の削減は、検出音情報が特徴量から特徴量特定情報に置き換えられることによる検出音情報の情報量の削減により実現される。また、マイク28により検出された音と番組の音とが一致しなかった場合、ユーザ端末2は検出音情報及び評価情報の何れもサーバ1へ送信しなくてもよいので、通信回数が削減される。 The comparison unit 113 of the server 1 compares the detected sound information received from the user terminal 2 with each time-series hash value of the hash value as the program sound information. As a result of the comparison, when there is a hash value that matches the detected sound information in the time series of the hash values, the evaluation information / detection information transmitting unit 213 determines that the evaluation information is used for evaluating the program. The evaluation information / detection information transmitting unit 213 specifies a time at which a sound corresponding to a hash value matching the detected sound information is broadcasted in the program to be evaluated as an input time of the evaluation information. By using the feature amount specifying information such as the hash value, the accuracy of the input time used for the evaluation is increased, and the communication amount between the server 1 and the user terminal 2 during the broadcast of the program can be reduced. This reduction in the communication amount is realized by reducing the information amount of the detected sound information by replacing the detected sound information from the characteristic amount to the characteristic amount specifying information. Further, when the sound detected by the microphone 28 does not match the sound of the program, the user terminal 2 does not need to transmit any of the detected sound information and the evaluation information to the server 1, so that the number of times of communication is reduced. You.
 例えば、図10において、ユーザ端末2は、番組音特徴量情報中の特徴量102が検出音情報と一致すると判定した。そこで、ユーザ端末2は、特徴量102のハッシュ値をサーバ1へ送信する。サーバ1側では、番組音情報中において元々特徴量102から生成されたハッシュ値102と、ユーザ端末2から受信されたハッシュ値とが一致する。従って、ハッシュ値102に対応付けられた時刻T102が、評価情報の入力時刻となる。 For example, in FIG. 10, the user terminal 2 has determined that the feature amount 102 in the program sound feature amount information matches the detected sound information. Therefore, the user terminal 2 transmits the hash value of the feature amount 102 to the server 1. On the server 1 side, the hash value 102 originally generated from the feature amount 102 in the program sound information matches the hash value received from the user terminal 2. Therefore, the time T102 associated with the hash value 102 is the input time of the evaluation information.
[2-2.番組評価システムの動作] [2-2. Operation of the program evaluation system]
 次に、番組評価システムSの動作について、図11及び12を用いて説明する。以下に説明する動作例において、ユーザ端末2は、評価情報が入力されたタイミングで、検出音情報を評価情報とともにサーバ1へ送信するものとする。また、ユーザ端末2は、番組音特徴量情報を予め取得して記憶部24に記憶しているものとする。 Next, the operation of the program evaluation system S will be described with reference to FIGS. In the operation example described below, it is assumed that the user terminal 2 transmits the detected sound information to the server 1 together with the evaluation information at the timing when the evaluation information is input. In addition, it is assumed that the user terminal 2 previously acquires the program sound feature amount information and stores it in the storage unit 24.
 図11は、ユーザ端末2のシステム制御部21により実行される端末処理の一例を示すフローチャートである。図11において、図8と同様の処理については同様の符号が付されている。 FIG. 11 is a flowchart illustrating an example of a terminal process executed by the system control unit 21 of the user terminal 2. 11, the same processes as those in FIG. 8 are denoted by the same reference numerals.
 先ず、評価情報取得部211は、評価情報が入力されたか否かを判定する(ステップS8)。評価情報取得部211は、評価情報が入力されたと判定した場合には(ステップS8:YES)、処理をステップ4に進める。ステップS4において、検出情報取得部212は、マイク28により音を検出させる。次いで、検出情報取得部212は、マイク28から出力される音声信号から特徴量を抽出する(ステップS41)。次いで、検出情報取得部212は、生成された特徴量と、番組音特徴量情報中における各特徴量とを比較する。評価情報取得部211は、生成された特徴量との間の一致度が閾値を超える特徴量が番組音特徴量情報中に存在するか否かを判定する(ステップS42)。検出情報取得部212は、一致度が閾値を超える特徴量が存在すると判定した場合には(ステップS42:YES)、処理をステップS43に進める。ステップS43において、検出情報取得部212は、番組音特徴量情報中において、生成された特徴量の一致度が閾値を超える特徴量のうち、一致度が最大である特徴量のハッシュ値を生成する。次いで、評価情報・検出情報送信部213は、評価情報と生成されたハッシュ値とをユーザIDとともにサーバ1へ送信する(ステップS44)。 First, the evaluation information acquisition unit 211 determines whether the evaluation information has been input (step S8). When it is determined that the evaluation information has been input (step S8: YES), the evaluation information acquisition unit 211 proceeds to step S4. In step S4, the detection information acquisition unit 212 causes the microphone 28 to detect a sound. Next, the detection information acquisition unit 212 extracts a feature amount from the audio signal output from the microphone 28 (Step S41). Next, the detection information acquisition unit 212 compares the generated feature amount with each feature amount in the program sound feature amount information. The evaluation information acquisition unit 211 determines whether or not a feature amount whose degree of coincidence with the generated feature amount exceeds a threshold exists in the program sound feature amount information (step S42). If the detection information acquisition unit 212 determines that there is a feature amount whose coincidence exceeds the threshold (step S42: YES), the process proceeds to step S43. In step S43, the detection information acquisition unit 212 generates a hash value of the feature amount having the highest matching degree among the feature amounts whose matching degree of the generated feature amount exceeds the threshold value in the program sound feature amount information. . Next, the evaluation information / detection information transmitting unit 213 transmits the evaluation information and the generated hash value to the server 1 together with the user ID (Step S44).
 ステップS44を終えたとき、ステップS8において評価情報が入力されていないと判定されたとき(ステップS8:NO)、又はステップS42において一致度が閾値を超える特徴量が存在しないと判定されたとき(ステップS42:NO)、システム制御部21は、評価対象の番組の放送終了時刻が到来したか否かを判定する(ステップS10)。システム制御部21は、終了時刻が到来していないと判定した場合には(ステップS10:NO)、処理をステップS8に進める。一方、システム制御部21は、終了時刻が到来したと判定した場合には(ステップS10:YES)、端末処理を終了させる。 When step S44 is completed, when it is determined in step S8 that the evaluation information has not been input (step S8: NO), or when it is determined in step S42 that there is no feature amount whose matching degree exceeds the threshold value ( (Step S42: NO), the system control unit 21 determines whether or not the broadcast end time of the evaluation target program has arrived (step S10). If it is determined that the end time has not arrived (step S10: NO), the system control unit 21 advances the processing to step S8. On the other hand, when it is determined that the end time has arrived (step S10: YES), the system control unit 21 ends the terminal processing.
 図12は、サーバ1のシステム制御部11により実行されるサーバ処理の一例を示すフローチャートである。図12において、図9と同様の処理については同様の符号が付されている。 FIG. 12 is a flowchart illustrating an example of a server process executed by the system control unit 11 of the server 1. In FIG. 12, the same processes as those in FIG. 9 are denoted by the same reference numerals.
 先ず、評価情報・検出情報受信部112は、何れかのユーザ端末2から評価情報及びハッシュ値を受信したか否かを判定する(ステップS51)。評価情報・検出情報受信部112は、評価情報及びハッシュ値を受信したと判定した場合には(ステップS51:YES)、処理をステップS52に進める。ステップS52において、比較部52は、受信されたハッシュ値と一致するハッシュ値が番組音情報中に存在するか否かを判定する。評価情報・検出情報受信部112は、受信されたハッシュ値と一致するハッシュ値が存在すると判定した場合には(ステップS52:YES)、処理をステップS53に進める。ステップS53において、評価情報利用決定部114は、受信された評価情報の入力時刻を、受信されたハッシュ値と一致するハッシュ値に対応付けられた放送時刻に設定する。次いで、評価情報利用決定部114は、評価情報、入力時刻及びユーザIDを関連付けて記憶部14に記憶させる(ステップS32)。一方、評価情報利用決定部114は、受信されたハッシュ値と一致するハッシュ値が存在しないと判定した場合には(ステップS52:NO)、受信した評価情報を破棄する(ステップS33)。 First, the evaluation information / detection information receiving unit 112 determines whether evaluation information and a hash value have been received from any of the user terminals 2 (step S51). When the evaluation information / detection information receiving unit 112 determines that the evaluation information and the hash value have been received (step S51: YES), the process proceeds to step S52. In step S52, the comparison unit 52 determines whether a hash value that matches the received hash value exists in the program sound information. If the evaluation information / detection information receiving unit 112 determines that there is a hash value that matches the received hash value (step S52: YES), the process proceeds to step S53. In step S53, the evaluation information use determining unit 114 sets the input time of the received evaluation information to the broadcast time associated with the hash value that matches the received hash value. Next, the evaluation information use determining unit 114 stores the evaluation information, the input time, and the user ID in the storage unit 14 in association with each other (Step S32). On the other hand, when it is determined that there is no hash value that matches the received hash value (step S52: NO), the evaluation information use determining unit 114 discards the received evaluation information (step S33).
 ステップS32若しくはS32を終えたとき、又はステップS51におい評価情報及びハッシュ値が受信されなかったと判定されたとき(ステップS51:NO)、評価部115は、評価対象の番組の放送終了時刻が到来したか否かを判定する(ステップS34)。評価部115は、終了時刻が到来していないと判定した場合には(ステップS34:NO)、処理をステップS51に進める。一方、評価部115は、終了時刻が到来したと判定した場合には(ステップS34:YES)、評価情報を用いて番組の評価処理を実行して(ステップS35)、サーバ処理を終了させる。 When step S32 or S32 is completed, or when it is determined in step S51 that the evaluation information and the hash value have not been received (step S51: NO), the evaluation unit 115 arrives at the broadcast end time of the program to be evaluated. It is determined whether or not (Step S34). If the evaluation unit 115 determines that the end time has not arrived (step S34: NO), the processing proceeds to step S51. On the other hand, when the evaluation unit 115 determines that the end time has arrived (step S34: YES), the evaluation unit 115 executes a program evaluation process using the evaluation information (step S35), and ends the server process.
 以上説明したように、本実施形態によれば、サーバ1が、番組音情報のうち、検出音情報と一致すると判定された部分により示される音のテレビ受像機4による出力時刻を、検出音情報とともに受信された評価情報の入力時刻として特定する。従って、番組の何れの場面に対して入力された評価情報であるかを適切に特定することが可能であるので、番組に対して適切な評価を行うことができる。 As described above, according to the present embodiment, the server 1 determines the output time of the sound indicated by the portion of the program sound information determined to match the detected sound information by the television receiver 4 by using the detected sound information. It is specified as the input time of the evaluation information received together with. Therefore, it is possible to appropriately specify which scene of the program the evaluation information is input to, so that the program can be appropriately evaluated.
 また、サーバ1が、番組のコンテンツを構成する音の特徴量の時系列における特徴量それぞれを特定する特徴量特定情報の時系列を、番組音情報として取得してもよい。ユーザ端末2は、音の特徴量の時系列を示す番組音特徴量情報を事前に取得してもよい。また、ユーザ端末2が、番組音特徴量情報の中に、検出された音の特徴量との間の一致度が閾値を超える特徴量が存在する場合、その特徴量を特定する特徴量特定情報を、検出音情報として送信してもよい。サーバ1は、番組音情報のうち、ユーザ端末2から受信された特徴量特定情報と一致する特徴量特定情報に対応する音のテレビ受像機4による出力時刻を、評価情報の入力時刻として特定してもよい。この場合、検出音情報の情報量が削減されるので、ユーザ端末2及びサーバ1の通信負荷を削減することができる。 The server 1 may acquire, as the program sound information, the time series of the feature amount specifying information for specifying each of the feature amounts in the time series of the feature amounts of the sounds constituting the program content. The user terminal 2 may acquire in advance program sound feature amount information indicating a time series of sound feature amounts. Further, when the user terminal 2 includes a feature amount whose degree of coincidence with the detected sound feature amount exceeds a threshold value in the program sound feature amount information, the feature amount identification information for identifying the feature amount May be transmitted as detected sound information. The server 1 specifies, as the input time of the evaluation information, the output time of the sound corresponding to the characteristic amount specifying information that matches the characteristic amount specifying information received from the user terminal 2 in the program sound information. You may. In this case, since the amount of detected sound information is reduced, the communication load on the user terminal 2 and the server 1 can be reduced.
1 サーバ
2 ユーザ端末
11 システム制御部
12 システムバス
13 入出力インターフェース
14 記憶部
15 通信部
111 番組情報取得部
112 評価情報・検出情報受信部
113 比較部
114 評価情報利用決定部
115 評価部
21 システム制御部
22 システムバス
23 入出力インターフェース
24 記憶部
25 通信部
26 操作入力部
27 表示部
28 マイク
29 カメラ
211 評価情報取得部
212 検出情報取得部
213 情評価情報・検出情報送信部
NW ネットワーク
1 server 2 user terminal 11 system control unit 12 system bus 13 input / output interface 14 storage unit 15 communication unit 111 program information acquisition unit 112 evaluation information / detection information reception unit 113 comparison unit 114 evaluation information use determination unit 115 evaluation unit 21 system control Unit 22 system bus 23 input / output interface 24 storage unit 25 communication unit 26 operation input unit 27 display unit microphone 29 camera 211 evaluation information acquisition unit 212 detection information acquisition unit 213 information evaluation information / detection information transmission unit NW network

Claims (16)

  1.  端末装置と、前記端末装置とネットワークを介して接続されるサーバ装置と、を含む評価システムにおいて、
     前記端末装置は、
     映像及び音の少なくとも何れか一方で構成されるコンテンツに対する評価情報がユーザにより入力される入力手段と、
     前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を出力する出力装置により出力された映像又は音を検出する検出手段と、
     前記入力された評価情報と、前記検出された映像又は音を示す検出情報と、を前記サーバ装置へ送信する送信手段と、
     を備え、
     前記サーバ装置は、
     前記コンテンツを構成する、前記映像又は前記音を示すコンテンツ情報を取得する取得手段と、
     前記端末装置から前記評価情報及び前記検出情報を受信する受信手段と、
     前記取得されたコンテンツ情報と前記受信された検出情報とを比較する比較手段と、
     前記比較手段による比較結果により、前記コンテンツ情報と前記検出情報との間に所定の一致がある場合、前記受信された評価情報を前記コンテンツの評価に用いるよう選択する選択手段と、
     を備えることを特徴とする評価システム。
    In an evaluation system including a terminal device and a server device connected to the terminal device via a network,
    The terminal device,
    Input means by which a user inputs evaluation information for content composed of at least one of video and sound,
    Detecting means that detects the video or sound output by the output device that outputs at least one of the video and the sound that constitutes the content,
    A transmitting unit that transmits the input evaluation information and detection information indicating the detected video or sound to the server device,
    With
    The server device,
    Acquisition means for acquiring the content information indicating the video or the sound, which constitutes the content,
    Receiving means for receiving the evaluation information and the detection information from the terminal device,
    Comparing means for comparing the acquired content information with the received detection information,
    Selecting means for selecting the received evaluation information to be used for evaluating the content, when there is a predetermined match between the content information and the detection information according to the comparison result by the comparing means,
    An evaluation system comprising:
  2.  前記コンテンツは、放送されるコンテンツであり、
     前記コンテンツの放送中に、前記端末装置として複数の端末装置それぞれから前記サーバ装置へ前記評価情報及び前記検出情報が送信され、
     前記サーバ装置は、
     前記複数の端末装置のうち少なくとも一の端末装置による前記評価情報及び前記検出情報の少なくとも何れか一方の送信タイミングが、前記複数の端末装置のうち他の少なくとも一の端末装置による送信タイミングと相違するように、前記複数の端末装置それぞれについて、前記送信タイミングを決定するタイミング決定手段と、
     前記決定された送信タイミングを示すタイミング情報を、前記複数の端末装置それぞれに送信するタイミング情報送信手段と、
     を更に備え、
     前記端末装置は、
     前記サーバ装置から前記タイミング情報を受信するタイミング情報受信手段を更に備え、
     前記送信手段は、前記受信されたタイミング情報に従って、前記評価情報及び前記検出情報の少なくとも何れか一方を送信することを特徴とする請求項1に記載の評価システム。
    The content is broadcast content,
    During the broadcasting of the content, the evaluation information and the detection information are transmitted from each of the plurality of terminal devices to the server device as the terminal device,
    The server device,
    A transmission timing of at least one of the evaluation information and the detection information by at least one terminal device of the plurality of terminal devices is different from a transmission timing of at least one other terminal device of the plurality of terminal devices. As described above, for each of the plurality of terminal devices, timing determining means for determining the transmission timing,
    Timing information transmitting means for transmitting the determined transmission timing to each of the plurality of terminal devices,
    Further comprising
    The terminal device,
    The apparatus further includes timing information receiving means for receiving the timing information from the server device,
    2. The evaluation system according to claim 1, wherein the transmission unit transmits at least one of the evaluation information and the detection information according to the received timing information. 3.
  3.  前記タイミング情報送信手段は、前記少なくとも一の端末装置による送信タイミングの間隔の間に、前記他の少なくとも一の端末装置による送信タイミングを決定することを特徴とする請求項2に記載の評価システム。 3. The evaluation system according to claim 2, wherein the timing information transmitting unit determines a transmission timing by the at least one other terminal device during an interval between transmission timings by the at least one terminal device.
  4.  前記検出手段及び前記送信手段は、所定時間間隔を置いて繰り返し前記映像又は前記音を検出して前記検出情報を送信し、
     前記選択手段は、前記受信手段により前記所定時間間隔を置いて繰り返し受信される前記検出情報のうち、前記評価情報の入力時刻以前又は以後で前記入力時刻に最も近い時刻に前記端末装置により検出される映像又は音を示す検出情報と前記コンテンツ情報との比較結果に少なくとも基づいて、前記選択を行うことを特徴とする請求項1乃至3の何れか1項に記載の評価システム。
    The detecting means and the transmitting means repeatedly detect the video or the sound at predetermined time intervals and transmit the detection information,
    The selection unit is detected by the terminal device at a time closest to the input time before or after the input time of the evaluation information, among the detection information repeatedly received at the predetermined time interval by the reception unit. 4. The evaluation system according to claim 1, wherein the selection is performed based on at least a comparison result between detection information indicating video or sound and the content information. 5.
  5.  前記選択手段は、前記受信手段により所定時間間隔を置いて繰り返し受信される前記検出情報のうち、前記評価情報の入力時刻から相対的に近い時刻に検出された映像又は音をそれぞれ示す2以上の検出情報と前記コンテンツ情報との比較結果に基づいて、前記選択を行うことを特徴とする請求項4に記載の評価システム。 The selection means is, among the detection information repeatedly received at predetermined time intervals by the reception means, two or more of each of the video or sound detected at a time relatively close to the input time of the evaluation information The evaluation system according to claim 4, wherein the selection is performed based on a comparison result between detection information and the content information.
  6.  前記検出手段及び前記送信手段は、前記評価情報が入力されたときに前記映像又は前記音を検出して、前記評価情報とともに前記検出情報をサーバ装置へ送信し、
     前記選択手段は、前記コンテンツ情報が、前記検出情報と一致すると判定される部分を有する場合、前記検出情報とともに受信された前記評価情報を、前記コンテンツの評価に用いるよう選択することを特徴とする請求項1乃至3の何れか1項に記載の評価システム。
    The detection means and the transmission means, when the evaluation information is input, detects the video or the sound, transmits the detection information together with the evaluation information to a server device,
    When the content information has a portion determined to match the detection information, the selection unit selects the evaluation information received together with the detection information to be used for evaluating the content. The evaluation system according to claim 1.
  7.  前記選択手段は、前記コンテンツ情報のうち、前記検出情報と一致すると判定された前記部分により示される映像又は音の前記出力装置による出力時刻を、前記検出情報とともに受信された前記評価情報の入力時刻として特定することを特徴とする請求項6に記載の評価システム。 The selecting means sets the output time of the video or sound indicated by the portion determined to match the detection information in the content information by the output time of the evaluation information received together with the detection information. The evaluation system according to claim 6, wherein the evaluation system is specified as:
  8.  前記取得手段は、前記コンテンツを構成する前記映像又は前記音の特徴を示す特徴情報の時系列で構成される特徴情報時系列データから変換された特定情報の時系列で構成される特定情報時系列データであって、各前記特定情報に基づいて対応する前記特徴情報が特定可能であり、且つ各前記特定情報の情報量は、対応する前記特徴情報の情報量よりも少ない特定情報時系列データを、前記コンテンツ情報として取得し、
     前記端末装置は、
     前記特徴情報時系列データを事前に取得する特徴情報時系列データ取得手段と、
     前記検出された映像又は音の特徴を示す特徴情報を抽出する抽出手段と、
     前記取得された特徴情報時系列データに含まれる前記特徴情報のうち、前記抽出された特徴情報との間の一致度が所定値を超える特徴情報を特定する特定情報であって、該特徴情報の情報量よりも少ない情報量の特定情報を生成する生成手段と、
     を更に備え、
     前記送信手段は、前記生成された特定情報を、前記検出情報として送信し、
     前記選択手段は、前記特徴情報時系列データに含まれる前記特定情報のうち、前記検出情報として受信された前記特定情報と一致する特定情報に対応する前記出力時刻を、前記入力時刻として特定することを特徴とする請求項7に記載の評価システム。
    The acquisition unit is a specific information time series composed of a time series of specific information converted from characteristic information time series data composed of a time series of characteristic information indicating characteristics of the video or the sound constituting the content. Data, the corresponding characteristic information can be specified based on each of the specific information, and the information amount of each of the specific information is the specific information time-series data smaller than the information amount of the corresponding characteristic information. , Obtained as the content information,
    The terminal device,
    Feature information time-series data acquisition means for acquiring the feature information time-series data in advance,
    Extraction means for extracting feature information indicating the feature of the detected video or sound,
    Among the feature information included in the acquired feature information time-series data, identification information that identifies feature information whose degree of coincidence with the extracted feature information exceeds a predetermined value, Generating means for generating specific information having an information amount smaller than the information amount;
    Further comprising
    The transmitting means transmits the generated specific information as the detection information,
    The selecting unit specifies, as the input time, the output time corresponding to specific information that matches the specific information received as the detection information, among the specific information included in the feature information time-series data. The evaluation system according to claim 7, wherein:
  9.  前記特定情報は、前記特徴情報のハッシュ値であることを特徴とする請求項8に記載の評価システム。 9. The evaluation system according to claim 8, wherein the specific information is a hash value of the feature information.
  10.  映像及び音の少なくとも何れか一方で構成されるコンテンツを構成する、前記映像又は前記音を示すコンテンツ情報を取得する取得手段と、
     前記コンテンツに対する評価情報がユーザにより入力され、且つ、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を出力する出力装置により出力された映像又は音を検出する端末装置から、前記評価情報と、前記検出された映像又は音を示す検出情報と、を受信する受信手段と、
     前記取得されたコンテンツ情報と前記受信された検出情報とを比較する比較手段と、
     前記比較手段による比較結果により、前記コンテンツ情報と前記検出情報との間に所定の一致がある場合、前記受信された評価情報を前記コンテンツの評価に用いるよう選択する選択手段と、
     を備えることを特徴とするサーバ装置。
    Acquisition means for acquiring content information indicating the video or the sound, which constitutes a content configured of at least one of a video and a sound,
    The evaluation information for the content is input by a user, and the terminal device that detects the video or sound output from the output device that outputs at least one of the video and the sound, which constitutes the content, Information, and detection information indicating the detected video or sound, receiving means for receiving,
    Comparing means for comparing the acquired content information with the received detection information,
    Selecting means for selecting the received evaluation information to be used for the evaluation of the content, when there is a predetermined match between the content information and the detection information according to the comparison result by the comparing means,
    A server device comprising:
  11.  映像及び音の少なくとも何れか一方で構成されるコンテンツに対する評価情報がユーザにより入力される入力手段と、
     前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を出力する出力装置により出力された映像又は音を検出する検出手段と、
     前記入力された評価情報と、前記検出された映像又は音を示す検出情報とを、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を示すコンテンツ情報と、前記検出情報と、の比較結果比較結果により、前記コンテンツ情報と前記検出情報との間に所定の一致がある場合、前記評価情報を前記コンテンツの評価に用いるよう選択するサーバ装置へ送信する送信手段と、
     を備えることを特徴とする端末装置。
    Input means by which a user inputs evaluation information for content composed of at least one of video and sound,
    Detecting means that detects the video or sound output by the output device that outputs at least one of the video and the sound that constitutes the content,
    The input evaluation information and the detection information indicating the detected video or sound, constituting the content, content information indicating at least one of the video and the sound, and the detection information, According to the comparison result, when there is a predetermined match between the content information and the detection information, a transmission unit that transmits the evaluation information to a server device that selects to be used for evaluating the content,
    A terminal device comprising:
  12.  端末装置と、前記端末装置とネットワークを介して接続されるサーバ装置と、を含む評価システムにおける情報処理方法において、
     前記サーバ装置が、映像及び音の少なくとも何れか一方で構成されるコンテンツを構成する、前記映像又は前記音を示すコンテンツ情報を取得する取得ステップと、
     前記端末装置が、前記端末装置が備える入力手段にユーザにより入力された、前記コンテンツに対する評価情報を取得する評価情報取得ステップと、
     前記端末装置が、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を出力する出力装置により出力された映像又は音を検出する検出ステップと、
     前記端末装置が、前記取得された評価情報と、前記検出された映像又は音を示す検出情報と、を前記サーバ装置へ送信する送信ステップと、
     前記サーバ装置が、前記端末装置から前記評価情報及び前記検出情報を受信する受信ステップと、
     前記サーバ装置が、前記取得されたコンテンツ情報と前記受信された検出情報とを比較する比較ステップと、
     前記サーバ装置が、前記比較ステップによる比較結果により、前記コンテンツ情報と前記検出情報との間に所定の一致があることが示される場合、前記受信された評価情報を前記コンテンツの評価に用いるよう選択するする選択ステップと、
     を含むことを特徴とする情報処理方法。
    In an information processing method in an evaluation system including a terminal device and a server device connected to the terminal device via a network,
    An acquisition step in which the server device configures a content configured of at least one of a video and a sound, and acquires content information indicating the video or the sound,
    The terminal device, input by the user to the input means provided in the terminal device, evaluation information obtaining step of obtaining evaluation information for the content,
    The terminal device, constituting the content, a detection step of detecting a video or sound output by an output device that outputs at least one of the video and the sound,
    A transmission step of transmitting the obtained evaluation information and the detection information indicating the detected video or sound, to the server device,
    A receiving step in which the server device receives the evaluation information and the detection information from the terminal device;
    A comparison step in which the server device compares the acquired content information with the received detection information,
    The server apparatus selects the received evaluation information to be used for evaluating the content when the comparison result in the comparing step indicates that there is a predetermined match between the content information and the detection information. A selection step to do;
    An information processing method comprising:
  13.  サーバ装置のコンピュータにより実行される情報処理方法において、
     映像及び音の少なくとも何れか一方で構成されるコンテンツを構成する、前記映像又は前記音を示すコンテンツ情報を取得する取得ステップと、
     前記コンテンツに対する評価情報がユーザにより入力され、且つ、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を出力する出力装置により出力された映像又は音を検出する端末装置から、前記評価情報と、前記検出された映像又は音を示す検出情報と、を受信する受信ステップと、
     前記取得されたコンテンツ情報と前記受信された検出情報とを比較する比較ステップと、
     前記比較ステップによる比較結果により、前記コンテンツ情報と前記検出情報との間に所定の一致がある場合、前記受信された評価情報を前記コンテンツの評価に用いるよう選択する選択ステップと、
     を含む情報処理方法。
    In an information processing method executed by a computer of a server device,
    Acquisition step of acquiring content information indicating the video or the sound, which constitutes a content configured of at least one of a video and a sound,
    The evaluation information for the content is input by a user, and the terminal device that detects the video or sound output from the output device that outputs at least one of the video and the sound, which constitutes the content, Information, and detection information indicating the detected video or sound, a receiving step of receiving,
    A comparing step of comparing the obtained content information with the received detection information,
    A selecting step of selecting the received evaluation information to be used for evaluating the content, when there is a predetermined match between the content information and the detection information according to the comparison result obtained by the comparing step;
    An information processing method including:
  14.  端末装置のコンピュータにより実行される情報処理方法において、
     前記端末装置が備える入力手段にユーザにより入力された、映像及び音の少なくとも何れか一方で構成されるコンテンツに対する評価情報を取得する評価情報取得ステップと、
     前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を出力する出力装置により出力された映像又は音を検出する検出ステップと、
     前記取得された評価情報と、前記検出された映像又は音を示す検出情報とを、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を示すコンテンツ情報と、前記検出情報と、の比較結果により、前記コンテンツ情報と前記検出情報との間に所定の一致があることが示される場合、前記評価情報を前記コンテンツの評価に用いるよう選択するサーバ装置へ送信する送信ステップと、
     を含む情報処理方法。
    In an information processing method executed by a computer of a terminal device,
    An evaluation information obtaining step of obtaining evaluation information for content configured by at least one of a video and a sound, which is input by a user to an input unit included in the terminal device,
    Constituting the content, a detection step of detecting a video or sound output by an output device that outputs at least one of the video and the sound,
    The acquired evaluation information, and the detection information indicating the detected video or sound, constituting the content, content information indicating at least one of the video and the sound, and the detection information, When the comparison result indicates that there is a predetermined match between the content information and the detection information, a transmitting step of transmitting the evaluation information to a server device selected to be used for evaluating the content,
    An information processing method including:
  15.  サーバ装置のコンピュータを、
     映像及び音の少なくとも何れか一方で構成されるコンテンツを構成する、前記映像又は前記音を示すコンテンツ情報を取得する取得手段と、
     前記コンテンツに対する評価情報がユーザにより入力され、且つ、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を出力する出力装置により出力された映像又は音を検出する端末装置から、前記評価情報と、前記検出された映像又は音を示す検出情報と、を受信する受信手段と、
     前記取得されたコンテンツ情報と前記受信された検出情報とを比較する比較手段と、
     前記比較手段による比較結果により、前記コンテンツ情報と前記検出情報との間に所定の一致がある場合、前記受信された評価情報を前記コンテンツの評価に用いるよう選択する選択手段と、
     として機能させることを特徴とする情報処理プログラム。
    The computer of the server device
    Acquisition means for acquiring content information indicating the video or the sound, which constitutes a content configured of at least one of a video and a sound,
    The evaluation information for the content is input by a user, and the terminal device that detects the video or sound output from the output device that outputs at least one of the video and the sound, which constitutes the content, Information, and detection information indicating the detected video or sound, receiving means for receiving,
    Comparing means for comparing the acquired content information with the received detection information,
    Selecting means for selecting the received evaluation information to be used for the evaluation of the content, when there is a predetermined match between the content information and the detection information according to the comparison result by the comparing means,
    An information processing program characterized by functioning as a computer.
  16.  端末装置のコンピュータを、
     前記端末装置が備える入力手段にユーザにより入力された、映像及び音の少なくとも何れか一方で構成されるコンテンツに対する評価情報を取得する評価情報取得手段と、
     前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を出力する出力装置により出力された映像又は音を検出する検出手段から、前記検出された映像又は音を示す検出情報を取得する検出情報取得手段と、
     前記取得された評価情報と、前記取得された検出情報とを、前記コンテンツを構成する、前記映像及び前記音の少なくとも何れか一方を示すコンテンツ情報と、前記検出情報と、の比較結果により、前記コンテンツ情報と前記検出情報との間に所定の一致がある場合、前記評価情報を前記コンテンツの評価に用いるよう選択するサーバ装置へ送信する送信手段と、
     として機能させることを特徴とする情報処理プログラム。
    The terminal computer,
    Evaluation information acquisition means for acquiring evaluation information for content constituted by at least one of video and sound, which is input by a user to input means provided in the terminal device,
    A detection unit that obtains detection information indicating the detected video or sound from a detection unit that detects the video or sound output by an output device that outputs at least one of the video and the sound that constitutes the content. Information acquisition means;
    The acquired evaluation information and the acquired detection information, constituting the content, content information indicating at least one of the video and the sound, and a comparison result of the detection information, When there is a predetermined match between the content information and the detection information, a transmission unit that transmits the evaluation information to a server device that selects to use the evaluation information for evaluating the content,
    An information processing program characterized by functioning as a computer.
PCT/JP2018/034671 2018-09-19 2018-09-19 Evaluation system, server device, terminal device, information processing method, and information processing program WO2020059047A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/JP2018/034671 WO2020059047A1 (en) 2018-09-19 2018-09-19 Evaluation system, server device, terminal device, information processing method, and information processing program
JP2019503504A JP6543429B1 (en) 2018-09-19 2018-09-19 Evaluation system, server device, terminal device, information processing method, and information processing program

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2018/034671 WO2020059047A1 (en) 2018-09-19 2018-09-19 Evaluation system, server device, terminal device, information processing method, and information processing program

Publications (1)

Publication Number Publication Date
WO2020059047A1 true WO2020059047A1 (en) 2020-03-26

Family

ID=67212255

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2018/034671 WO2020059047A1 (en) 2018-09-19 2018-09-19 Evaluation system, server device, terminal device, information processing method, and information processing program

Country Status (2)

Country Link
JP (1) JP6543429B1 (en)
WO (1) WO2020059047A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002271824A (en) * 2001-03-13 2002-09-20 Ricoh Co Ltd Program evaluation acquisition system
WO2012043735A1 (en) * 2010-09-30 2012-04-05 楽天株式会社 Server device for collecting survey responses
JP2013197716A (en) * 2012-03-16 2013-09-30 Nec Corp Transmission control device, feedback information transmitter, cooperative transmission system, transmission control method and feedback information transmission system and program

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002271824A (en) * 2001-03-13 2002-09-20 Ricoh Co Ltd Program evaluation acquisition system
WO2012043735A1 (en) * 2010-09-30 2012-04-05 楽天株式会社 Server device for collecting survey responses
JP2013197716A (en) * 2012-03-16 2013-09-30 Nec Corp Transmission control device, feedback information transmitter, cooperative transmission system, transmission control method and feedback information transmission system and program

Also Published As

Publication number Publication date
JPWO2020059047A1 (en) 2020-12-17
JP6543429B1 (en) 2019-07-10

Similar Documents

Publication Publication Date Title
US11659220B2 (en) System and method for surveying broadcasting ratings
KR101818986B1 (en) Method, device, and system for obtaining information based on audio input
US9942711B2 (en) Apparatus and method for determining co-location of services using a device that generates an audio signal
CN105229629B (en) For estimating the method to the user interest of media content, electronic equipment and medium
US8578415B2 (en) Information providing method, content display terminal, portable terminal, server device, information providing system and program
US10587921B2 (en) Viewer rating calculation server, method for calculating viewer rating, and viewer rating calculation remote apparatus
KR102166423B1 (en) Display device, server and method of controlling the display device
US20140075465A1 (en) Time varying evaluation of multimedia content
KR20150104358A (en) Server apparatus and method for transmitting finger print and broadcasting receiving apparatus
CN111405363B (en) Method and device for identifying current user of set top box in home network
JP6433615B1 (en) Viewing record analysis apparatus, viewing record analysis method, and viewing record analysis program
US20180260483A1 (en) Information processing apparatus, information processing method, and program
US20150026744A1 (en) Display system, display apparatus, display method, and program
US20130132996A1 (en) System and method for displaying product information about advertisement on digital television, and recording medium thereof
WO2020059047A1 (en) Evaluation system, server device, terminal device, information processing method, and information processing program
JPWO2010119834A1 (en) Content URL notification system
CN112235592B (en) Live broadcast method, live broadcast processing method, device and computer equipment
JP6567715B2 (en) Information processing apparatus, information processing method, and program
JP2020166791A (en) Advertisement distribution system
JP2020048079A (en) Viewing record analyzer and viewing record analyzing method
JP2017092664A (en) Program analysis device and program analysis method
JP2019220826A (en) View recording analyzer, view recording analytical method, and view recording analytical method
CN112153469A (en) Multimedia resource playing method, device, terminal, server and storage medium
JP2020166788A (en) Advertisement distribution system

Legal Events

Date Code Title Description
ENP Entry into the national phase

Ref document number: 2019503504

Country of ref document: JP

Kind code of ref document: A

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18934012

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18934012

Country of ref document: EP

Kind code of ref document: A1