WO2009119063A1 - Program information display device and program information display method - Google Patents

Program information display device and program information display method Download PDF

Info

Publication number
WO2009119063A1
WO2009119063A1 PCT/JP2009/001274 JP2009001274W WO2009119063A1 WO 2009119063 A1 WO2009119063 A1 WO 2009119063A1 JP 2009001274 W JP2009001274 W JP 2009001274W WO 2009119063 A1 WO2009119063 A1 WO 2009119063A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
display
feature amount
performer
unit
Prior art date
Application number
PCT/JP2009/001274
Other languages
French (fr)
Japanese (ja)
Inventor
信裕 神戸
Original Assignee
パナソニック株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by パナソニック株式会社 filed Critical パナソニック株式会社
Publication of WO2009119063A1 publication Critical patent/WO2009119063A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/16Analogue secrecy systems; Analogue subscription systems
    • H04N7/173Analogue secrecy systems; Analogue subscription systems with two-way working, e.g. subscriber sending a programme selection signal
    • H04N7/17309Transmission or handling of upstream communications
    • H04N7/17318Direct or substantially direct transmission and handling of requests
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/441Acquiring end-user identification, e.g. using personal code sent by the remote control or by inserting a card
    • H04N21/4415Acquiring end-user identification, e.g. using personal code sent by the remote control or by inserting a card using biometric characteristics of the user, e.g. by voice recognition or fingerprint scanning
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/4722End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for requesting additional data associated with the content
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8126Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts
    • H04N21/8133Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts specifically related to the content, e.g. biography of the actors in a movie, detailed information about an article seen in a video program
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/84Generation or processing of descriptive data, e.g. content descriptors

Definitions

  • the present invention relates to a program information display device and a program information display method for displaying program information, for example, a program information display for displaying or reproducing a moving image, such as a television receiver, a DVD (Digital Versatile Disc) player, a hard disk recorder, etc.
  • the present invention relates to an apparatus and a program information display method in such an apparatus.
  • the name and photo of the performer are not always posted, so it is difficult to match the face and name. Even if the user is able to know the name of the performer from the program information, the name can be matched with the face by accessing the Internet and searching for an image with the name of the performer, for example. It may become. However, it is troublesome for the user.
  • an image list and an image are periodically acquired from an external device in order to update the image feature database to be searched for images based on the image feature to the latest information.
  • an image search device for example, Patent Document 2.
  • the image search device described in Patent Document 2 is a database for searching for an image feature quantity from a keyword and searching for a similar image based on the image feature quantity. ) Does not search. Further, when the image search apparatus searches the network for images acquired based on keywords, many images (noise) that are related but do not indicate a performer are also searched. However, since the search method is different, no consideration is given to noise.
  • the purpose of the present invention is to perform a search with less noise and high accuracy in response to a request to know a performer name in association with a face on the screen, such as when a character does not match a performer name while watching a program. It is an object to provide a program information display device and a program information display method that can be used.
  • the program information display device of the present invention includes a program acquisition unit that acquires a program that is a moving image and program information including a performer name, and a related image that acquires a related image from a network based on the performer name.
  • An acquisition unit; a first feature amount calculation unit that calculates a feature amount of a first extracted image obtained by cutting a person's region from the related image; and a feature amount of the first extracted image and a source of the related image Based on the information, a representative feature amount determining unit that determines a representative feature amount and its reliability, and the determined representative feature amount and reliability are managed as performer information in association with the performer name.
  • a performer information management unit a display image acquisition unit that acquires a display image from a frame that constitutes the moving image, and a second feature that calculates a feature amount of a second extracted image obtained by extracting a person's region from the display image
  • a quantity calculation unit and the second feature A similarity between the feature amount of the second extracted image calculated by the calculation unit and the representative feature amount held in the performer information management unit is calculated, and the representative feature amount that maximizes the similarity
  • a search unit for acquiring a performer name associated with the at least one of the reliability or the similarity, the performer name acquired by the search unit, and the region of the second extracted image.
  • a configuration is provided that includes a display information generation unit that generates display information based on the display information and a display unit that displays the display image and the display information.
  • the program information display method of the present invention includes a step of acquiring a program that is a moving image and a program including a performer name, a step of acquiring a related image from a network based on the performer name, and the related Based on a first feature amount calculating step for calculating a feature amount of a first extracted image obtained by cutting out a person's region from the image, and on the feature amount of the first extracted image and the information on the source of the related image, A step of determining a representative feature amount and its reliability, a step of managing the determined representative feature amount and reliability as performer information in association with the performer name, and a frame constituting the moving image A display image is obtained from the display image; a second feature value calculating step for calculating a feature value of a second extracted image obtained by cutting out a person region from the display image; and a second feature value calculating step.
  • the present invention it is not necessary to prepare a performer's face image database in advance, and intuitively determine the correctness of the performer's determination result when displaying the performer's name in association with the performer's area. The effect that can be obtained.
  • the search is performed with less noise and high accuracy by reflecting the reliability based on a database that is dynamically searched. Can be realized.
  • FIG. The flowchart which shows the processing operation until the representative feature-value of the program information display apparatus which concerns on this Embodiment 1 is determined.
  • FIG. 1 The figure which shows the example of display information generation at the time of paying attention only to the font size of the display information of the program information display apparatus which concerns on this Embodiment 1.
  • FIG. The figure which shows the structure of the program information display system provided with the program information display apparatus which concerns on Embodiment 2 of this invention.
  • the flowchart which shows the process until the display of the performer name of the program information display apparatus which concerns on this Embodiment 2.
  • FIG. 1 is a diagram showing a configuration of a program information display system provided with a program information display device according to Embodiment 1 of the present invention.
  • the program information display apparatus according to this embodiment is an example in which the present invention is applied to a television receiver capable of receiving digital broadcast radio waves.
  • the program information display system includes a program information display device 100, a network 200, an image search device 210, an image server 220, a program information server 230, a program guide server 240, and a broadcast station 250.
  • the program information display device 100 is a digital broadcast receiver that reproduces program information transmitted from the broadcast station 250. The detailed configuration of the program information display device 100 will be described later.
  • the network 200 is a communication network composed of the Internet or a dedicated line. More specifically, the network 200 is a network composed of a mobile communication network, a public telephone network, a LAN, the Internet, or the like. The network 200 may be either a wired system or a wireless system, and the type of protocol is not particularly limited. Further, as an access line of the network 200, a large capacity line such as FTTH (Fiber To The Home), HFC (Hybrid Fiber Coax), and ADSL (Asymmetric Digital Subscriber Line) can be used as FTTH (Fiber To The Home), HFC (Hybrid Fiber Coax), and ADSL (Asymmetric Digital Subscriber Line) can be used.
  • FTTH Fiber To The Home
  • HFC Hybrid Fiber Coax
  • ADSL Asymmetric Digital Subscriber Line
  • the image search device 210 is an image search site. More specifically, the image search device 210 is an image search site that collects and holds images held by various sites, their description information, and URLs (Uniform Resource Locator) of images via the network 200.
  • the sites to be searched by the image search device 210 include a plurality of sites including the image server 220 and the program information server 230.
  • the image search device 210 searches for information on the collected image based on the keyword, and outputs the search result.
  • the information about the image includes the image itself and the URL of the image.
  • the image server 220 is included in the collection destination of the information related to the image of the image search apparatus 210.
  • the image search device 210 attempts to improve search accuracy by devising a search algorithm.
  • the image server 220 is a general site that publishes web pages to be posted. More specifically, the image server 220 is a site that is managed by an unspecified number of general users or companies, and holds some images and publishes them to the network 200. Public sites do not always represent information properly. However, there are far more public sites than, for example, an official site of a specific broadcast program, and the network 200 as a whole has a huge amount of information.
  • the program information server 230 is an official site for programs broadcast by the broadcast station 250 (hereinafter referred to as “programs” as appropriate). More specifically, the program information server 230 is an official program site, and is generally managed by a television station or a program production company that is the right holder of the program. Therefore, it can be said that the official site is one of the sites that appropriately express the contents of the program. However, depending on the program, the official website may not exist, and even if it exists, the level of content fulfillment varies. Here, the official site is described as a program site, but the official site of a program performer can also be handled in the same manner because it is a site that appropriately expresses information about the performer.
  • the program guide server 240 is a site that publishes a program guide of programs of each station including the broadcast station 250. More specifically, the program guide server 240 is a site that provides, via the network 200, a net program guide that is a list of detailed program information of the corresponding program based on the date and time or the region.
  • the detailed program information includes a program name, a performer name, a URL of an official site, and the like.
  • the servers 220, 230, and 240 are connected via a network 200 to a control unit that includes a computer that controls the entire server.
  • the servers 220, 230, and 240 are each composed of, for example, a communication interface that transmits and receives data accessed by a URL on the Internet, and a database (DB) that stores data.
  • DB database
  • the program information display device 100 has the following configuration.
  • the program information display apparatus 100 includes a program acquisition unit 101, a display image acquisition unit 102, a display unit 103, a performer information management unit 104, a search unit 105, a display information generation unit 106, a related image acquisition unit 107, and a feature amount calculation unit 108. (First feature value calculating unit), representative feature value determining unit 109, performer information holding unit 110, and feature value calculating unit 111 (second feature value calculating unit).
  • the program information display device 100 includes an interface for connecting various devices such as an input device that accepts user input operations and an external recording device that records information.
  • the input device is a keyboard or a remote control device including numeric keys, cross keys, and the like.
  • the program acquisition unit 101 is a tuner that receives broadcasts of programs and program information and decodes the programs and program information.
  • the program consists of moving images.
  • Program information is, for example, various types of information related to programs included in an electronic program guide such as EPG.
  • the electronic program guide is composed of a program name, a channel, a broadcast date, a broadcast start time and end time, a performer name, and the like.
  • the program acquisition unit 101 extracts an electronic program guide from a broadcast signal and acquires program information from the extracted electronic program guide.
  • the program acquisition unit 101 has a memory (not shown) inside, and stores the acquired program information in this memory.
  • the program acquisition unit 101 acquires program information from EPG or the like
  • the name described as a performer in the electronic program guide is a target to be acquired as the performer name of the program. It is not subject to acquisition as a person's name.
  • FIG. 2 is a diagram showing an example of an electronic program guide. As shown in FIG. 2, this electronic program guide has set values for program names, synopsis, performers, and URL field names.
  • this electronic program guide has set values for program names, synopsis, performers, and URL field names.
  • “Good Morning Tokyo” is described as the program name, “Takeshi (Taro Matsuyama), Hiroko (Hanako Takeda), Yuri (Ryoko Umekawa)” as the performers, and the official website of the program as the URL The address is listed.
  • the display image acquisition unit 102 extracts a frame constituting a moving image to be displayed from the program acquired by the program acquisition unit 101, and an image of the program to be displayed on the TV screen (hereinafter referred to as “display image”) from the extracted frame. To get.
  • the display unit 103 is a display that displays a moving image (display image) in units of frames extracted by the display image acquisition unit 102.
  • the display unit 103 displays a display image and display information at the same time.
  • the performer information management unit 104 extracts information such as the performer name and the URL of the official site from the program information stored in the program acquisition unit 101.
  • the performer information management unit 104 stores the representative feature amount and reliability related to the related image obtained from the representative feature amount determination unit 109 in the performer information holding unit 110 in association with the performer name.
  • the related image is an image related to the performer of the program, and is an image that is likely to include the face image of the performer.
  • the representative feature amount is a value representing the image feature amount of a face image extracted from a plurality of related images related to the same performer (hereinafter referred to as “related extracted image”). This is an image feature amount that is highly likely to match the image feature amount.
  • the degree of reliability is the degree of reliability that the representative feature amount matches the image feature amount of the actual face image of the performer.
  • the search unit 105 searches for a performer name related to a face image similar to the face image extracted from the display image (hereinafter referred to as “display extracted image”) from the related extracted images, and temporarily stores the search result. To do. Specifically, the search unit 105 uses the feature amount calculation unit 111 to obtain a region of the display extracted image and an image feature amount (hereinafter referred to as “extracted image feature amount”). And the search part 105 acquires the representative feature-value of a related extraction image by the performer information management part 104. FIG. Then, the search unit 105 calculates the similarity between the extracted image feature quantity and the representative feature quantity, and acquires the performer name associated with the representative feature quantity having the maximum similarity. That is, the search part 105 acquires the performer name corresponding to the related extraction image similar to a display extraction image.
  • the display information generation unit 106 generates display information to be displayed on the display unit 103 from at least one of the reliability or similarity of the representative feature amount, the performer name acquired by the search unit 105, and the display extracted image area. .
  • An example of display information generation by the display information generation unit 106 will be described later with reference to FIG.
  • the related image acquisition unit 107 acquires a related image by communicating with the network 200.
  • the related image acquisition unit 107 is connected to the network 200 and acquires one or more related images related to the performer based on the program information.
  • Feature amount calculation units 108 and 111 calculate the image feature amount of an extracted image obtained by cutting out a human face area from an image.
  • the algorithm for calculating the image feature amount is the same between the feature amount calculation unit 108 and the feature amount calculation unit 111, but the processing target is different.
  • the feature amount calculation unit 108 calculates an image feature amount for a related extracted image obtained by cutting out a human face area from a related image acquired from the network 200.
  • the feature amount calculation unit 111 calculates an image feature amount for a display extracted image obtained by cutting out a human face area from a display image displayed on the TV screen.
  • the feature amount calculation units 108 and 111 calculate the image feature amount for each of the plurality of persons.
  • the representative feature amount determination unit 109 compares the weighted value of the image feature amount of the related extracted image to determine the representative feature amount and the reliability. Specifically, the representative feature amount determination unit 109 performs weighting on the image feature amount of the related extracted image (face region).
  • a weighting method for example, when the source of the base related image is an official site, there is a method of calculating the representative feature amount in a state where the number of the related extracted images is increased. According to this method, the number of image feature quantities of the relevant extracted image is larger than the number of image feature quantities of the other related extracted images, so that the image feature quantities of the related extracted images can be weighted.
  • a representative feature amount determination method a method using an arithmetic average or a method using a value that maximizes the number of all the image feature amounts of the relevant extracted image can be considered.
  • the performer information holding unit 110 is configured by a hard disk or a memory, and the representative feature amount determined by the representative feature determination unit 109 in association with the performer name extracted from the program information by the performer information management unit 104 and its performer Maintain confidence.
  • the performer name, the representative feature quantity, and the reliability are collectively referred to as “performer information” as appropriate.
  • FIG. 3 is a diagram illustrating an example of performer information held in the performer information holding unit 110.
  • the performer names shown in FIG. 3 are representative feature amounts of the performer names (titles) “Taro Matsuyama, Hanako Takeda (Hiroko), Ryoko Umekawa (Yuri)” of the performers in the electronic program guide of FIG. And its reliability.
  • the reliability of the representative feature amount of the performer name (title) “Hanako Takeda” is higher than the reliability of the representative feature amounts of the other two people.
  • the performer information management unit 104, the search unit 105, the display information generation unit 106, the feature amount calculation unit 108, the representative feature amount determination unit 109, and the feature amount calculation unit 111 include a CPU and the like, and execute program information display processing. Control of the entire device including The CPU includes a ROM, a RAM, and an EEPROM (electrically erasable programmable) ROM, which is an electrically rewritable nonvolatile memory, or a flash ROM.
  • the memory stores various data such as programs, communication control data, and terminal identification codes.
  • the memory stores the performer information of the performer information holding unit 110.
  • FIG. 4 and 5 are flowcharts showing the operation of the program information display apparatus 100.
  • FIG. FIG. 4 shows processing until the representative feature amount is determined.
  • FIG. 5 shows processing until the name of the performer is displayed after the representative feature amount is determined.
  • the flow shown in FIG. 4 is basically executed while the user is viewing the corresponding program.
  • the flow shown in FIG. 5 is basically executed with a user instruction as a trigger. As described above, since the execution conditions are different, the description will be divided into two flows.
  • the flow shown in FIGS. 4 and 5 is executed as a program information display program by the CPU as described above. In the figure, the symbol “S” indicates each step of the flow.
  • step S11 the program acquisition unit 101 acquires program information and a moving image of a program to be displayed.
  • the easiest way to acquire the program information is to acquire an electronic program guide (EPG) and acquire it from the acquired electronic program guide.
  • EPG electronic program guide
  • the method of performing and acquiring from the recognition result may be employ
  • the electronic program guide may include the URL of the official site of the performer on the Internet and the URL of the official site of the program. An example of the electronic program guide is shown in FIG.
  • step S12 the performer information management unit 104 repeats the process between the start and end of the following loop for “all performer names”.
  • the performer information management unit 104 extracts a performer name from the performer field of the electronic program guide based on the program information acquired by the program acquisition unit 101, and for all the performer names obtained thereafter, Process.
  • step S13 the performer information management unit 104 checks whether or not the performer name is held in the performer information holding unit 110. However, it is assumed that no performer name is registered in the performer information holding unit 110 at the start of the process.
  • step S14 when the performer name is held in the performer information holding unit 110, the performer information management unit 104 moves the process to the next performer. On the other hand, if the performer name is not held in the performer information holding unit 110, the performer management information unit 104 proceeds to step S15.
  • step S15 the related image acquisition unit 107 acquires a related image based on the performer name.
  • the related image acquisition unit 107 searches for related images in the network 200 including the Internet using the performer name as a keyword. Then, the related image acquisition unit 107 acquires and holds, as a search result, information such as the related image related to the performer name or the URL of the acquisition destination from the image search device 210, for example.
  • step S16 the feature amount calculation unit 108 repeats the process between the start and end of the following loop for “all related images”.
  • step S ⁇ b> 17 the feature amount calculation unit 108 cuts out a person's face area as a related extracted image from all the related images acquired for each performer by the related image acquisition unit 107, and sets the image feature amount for each extracted related extracted image. Ask for. Then, the feature amount calculation unit 108 temporarily stores the obtained image feature amount in association with the performer name corresponding to the base related image.
  • a method of cutting out a person's face area a method of extracting an edge from color or luminance information and performing a pattern recognition process of a face outline and a face part using the extracted edge is used.
  • the image feature amount is a value obtained by digitizing the color tone, the arrangement position of the face parts, and the like according to a predetermined reference.
  • step S18 the representative feature amount determination unit 109 determines a representative feature amount and reliability for each performer name from the image feature amounts of the related extracted images obtained by the feature amount calculation unit 108.
  • step S19 the performer information management part 104 hold
  • An example of information held in the performer information holding unit 110 is shown in FIG.
  • the image feature amount differs depending on the algorithm, accuracy, and storage format of the feature amount calculation unit 108, and this example does not indicate a general expression of the image feature amount.
  • the feature amount calculation unit 108 Prior to the processing of the representative feature amount determination unit 109, the feature amount calculation unit 108 performs noise removal processing on the image feature amount obtained from the related extracted image.
  • the noise is a characteristic value that greatly differs among the feature values constituting the image feature amount associated with the same performer name, that is, the feature value of the image feature amount of the performer's face image. Points that are not likely.
  • noise refers to an image from which such a characteristic value is acquired, that is, an image that is not likely to be a face image of the performer.
  • the program information display device 100 searches for related images from the image search device 210 via the network 200 using the performer name as a keyword.
  • the program information display device 100 extracts related images from the program information server 230 or the image server 220 according to the search algorithm of the image search device 210.
  • Such an image search is generally performed by a method of searching for an image related by a keyword.
  • the search result may include an image that is not a face image related to the performer name.
  • the image feature amount is obtained according to the algorithm of the feature amount calculation unit 108.
  • the image feature amount represents, for example, the arrangement position of the face part by a combination of a plurality of feature values (here, the feature value is 0 or more).
  • the feature amount calculation unit 108 obtains a statistical distribution of feature values of a plurality of related images related to the performer, and obtains a representative feature amount from the distribution.
  • the feature amount calculation unit 108 identifies feature values that are within a predetermined threshold from the average of all of the plurality of feature values, recalculates the average of the specified feature values, and displays the recalculation result as a representative feature. Adopt to quantity.
  • the feature amount calculation unit 108 excludes feature values (noise) whose values are greatly different from the average.
  • a specific example of the feature amount calculation by the feature amount calculation unit 108 will be described later with reference to FIG.
  • the feature amount calculation unit 108 excludes all image feature amounts obtained from images other than the target person's face screen as noise.
  • an image search is performed using a so-called search engine, an irrelevant image is searched, or an image in which a person other than the person is shown is often searched. Since it is considered that the accuracy increases as the number of searches increases and the number of feature amount calculation processes increases, it is possible to increase the accuracy by switching a plurality of search engines and repeating the search several times. For example, while the program is being received after the program has started, the related image search process may be constantly performed by background processing. If there is a program reservation, it is possible to execute the flow of FIG. 4 in advance prior to the start of the program. In this way, more related images can be searched, and the reliability of the representative feature amount can be increased.
  • the performer information management unit 104 compares the URL of the program official site held by the program acquisition unit 101 with the URL of the related image acquired by the related image acquisition unit 107, and if they are the same domain, , It is determined that the acquired related image is of the official website. Then, the representative feature amount determination unit 109 weights the feature amount of the related image to reflect the high reliability of the related image in the calculated representative feature amount. For example, the reliability weighting is calculated as follows.
  • the image feature amount F (x) for each performer calculated by the feature amount calculation unit 108 is expressed by the following equation (1) including m feature values.
  • the variable x represents a performer.
  • the representative feature amount F t is obtained by the following equation (2), where n is the number of related images after noise elimination (number of image feature amounts) and s i is a coefficient indicating whether or not the site is an official site. .
  • the reliability vector C v is defined as following equation (3), defined as the following equation (4) the reliability C.
  • the reliability vector C v is the characteristic value of the group, a vector value that indicates how much held together (or not vary) among a plurality of related images
  • reliability C is a scalar of the vector value This is a value obtained by normalizing the quantity by the number of feature values. As the ratio of the correct image (the target performer's face image) in the plurality of related extracted images for which the representative feature amount F t is calculated becomes higher, the variation in the feature value becomes smaller and the reliability C becomes higher. The value gets bigger. Therefore, the reliability C indicates the high reliability that the representative feature amount F t matches the image feature amount of the actual face image of the performer.
  • the coefficient s i for determining whether or not the site is an official site is set to 2 for an image of an official site, for example, and 1 for an image from another site.
  • the image feature amount is a combination of m feature values as shown in the above equation (1).
  • Individual image feature amounts can be obtained from color information and position information of each part of the face.
  • the feature amount calculation unit 108 calculates a hue or the like based on the color information, and uses the calculation result as a feature value.
  • the feature amount calculation unit 108 calculates the aspect ratio or relative position of each part of the face, and uses the calculation result as a feature value.
  • the position information of the facial parts can be obtained, for example, by extracting the contour in the image from the brightness, hue, etc. obtained from the color information.
  • FIG. 7 shows an example of the feature amount calculated by the feature amount calculation unit 108.
  • the image feature amount has three feature values 1 to 3 for each feature amount number.
  • An example of calculating a representative feature amount will be described using this feature amount.
  • the feature quantity number is an identification number of each feature quantity of the related images acquired corresponding to the same performer.
  • the feature amount calculation unit 108 calculates an arithmetic mean and variance in order to remove noise and to calculate reliability. Variance means the degree of variation of feature values.
  • the feature amount calculation unit 108 calculates the above-described threshold value based on the variance, and removes noise by excluding the feature value exceeding the calculated threshold value from the representative feature amount calculation target. In addition, the aspect which makes the value of dispersion
  • the feature amount calculation unit 108 obtains the arithmetic mean and variance using the five related images.
  • the related image is an image searched based on a certain keyword (performer name). Therefore, a plurality of related images are searched for one keyword (performer name).
  • the feature amount calculation unit 108 obtains a representative feature amount for the keyword (performer name) using the plurality of related images.
  • the related image searched by the keyword includes an unrelated image (noise).
  • the retrieved related image may not be an image of a target person.
  • the average value obtained with noise included is unlikely to be a reasonable value. Therefore, the feature amount calculation unit 108 recalculates the average value after removing the noise.
  • the feature amount calculation unit 108 performs the following processes (a) to (e).
  • the feature quantity calculation unit 108 obtains an arithmetic average of five feature values corresponding to the feature quantity numbers 1 to 5 for each of the feature values 1 to 3 in FIG. At that time, the feature amount calculation unit 108 weights the feature amount of the official site by a predetermined multiple (for example, five times) over the other feature amounts.
  • the feature amounts 1 to 3 with the feature amount number 1 are the feature amounts acquired from the official site, and thus are weighted five times as much as the feature amounts 1 to 3 with the feature amount numbers 2 to 5, respectively. .
  • the number of populations has increased by four.
  • the feature value calculation method will be described by taking the case of the feature value 1 among the feature values 1 to 3 shown in FIG. 7 as an example.
  • the arithmetic average f 1m of the feature value 1 is expressed as the following equation (5).
  • the variance ⁇ 2 is expressed as the following equation (6).
  • the threshold value th is expressed by the following equation (7) in the arithmetic mean f 1m of the feature value 1.
  • the feature value 1 of the feature number 4 shown in FIG. 7 does not correspond to the value f 1e that the feature value 1 can originally take, and can be regarded as noise.
  • the feature amount calculation unit 108 recalculates f 1m according to the following equation (10), excluding the data of feature amount number 4 and considering weighting. However, noise removal by excluding other feature values is ignored.
  • the value of the feature value 1 of the representative feature amount can be set to “2” in FIG.
  • a new average can be obtained by excluding the feature amount that becomes noise, and the representative feature amount can be calculated.
  • the representative feature amount determination unit 109 determines that the reliability is low when the variation of the image feature amount used to calculate the representative feature amount is large for the same keyword, and the reliability is high when the variation is small. .
  • the representative feature quantity determination unit 109 may obtain the reliability C by another method in order to simplify the calculation.
  • the representative feature amount determination unit 109 sets the reliability C as the following equation (11), where n is the number of feature amounts excluding noise, a is the number of images before noise removal, and N is the number of images on the official site. Ask for.
  • the reliability is low when the variation in the feature amount is large, and the reliability is high when the variation in the feature amount is small.
  • the representative feature quantity determination unit 109 holds the performer name, the representative feature quantity F t , and the reliability C thus obtained in the performer information holding unit 110 in an associated state.
  • the representative feature amount determination unit 109 may store the role name in the performer information holding unit 110 when the role name is obtained in addition to the performer name as the program information.
  • the representative feature quantity determining unit 109 may perform the processing until the representative feature quantity F t is determined asynchronously step by step during program display.
  • the representative feature quantity F t is calculated for each stage, and the later calculated representative feature quantity F t is used as a collation process described later. Used for. Thereby, the collation process can be started from the early stage of the program viewing start, and the reliability of the representative feature quantity F t can be increased with the passage of time.
  • the representative feature amount determination unit 109 processes the next performer name when several related images are processed for the performer name, processes all performer names, and then again performs each performer name. Is processed with several other related images. By performing the processing step by step in this way, the collation processing can be started from the early stage of the program viewing start, and the reliability of the representative feature amount F t can be increased with the passage of time. Become.
  • step S20 the display image acquisition unit 102 acquires a display image in units of frames from the moving image acquired from the program acquisition unit 101, and holds the acquired display image.
  • step S21 the display unit 103 displays the acquired display image as a moving image as it is.
  • the search unit 105 cuts out the face area of the person from the frame-by-frame display image acquired by the display image acquisition unit 102, outputs the cut out area as a display extracted image, and extracts the display extracted image ( (Hereinafter simply referred to as [position]).
  • the feature amount calculation unit 111 calculates an extracted image feature amount from the display extraction image. When a plurality of persons are shown in the display image, a plurality of display extraction images are obtained.
  • the feature amount calculation unit 111 has the same function as the feature amount calculation unit 108 that has already been described in detail. However, while the feature amount calculation unit 108 calculates the feature amount of the related extracted image, the feature amount calculation unit 111 calculates the feature amount of the display extracted image.
  • step S23 the search unit 105 repeats the process between the start and end of the following loop for “all related images”.
  • step S24 the search unit 105 acquires the representative feature amount from the performer information holding unit 110 via the performer information management unit 104 until there is no more representative feature amount that has been verified in the performer information holding unit 110.
  • step S25 the search unit 105 collates the extracted image feature quantity with the representative feature quantity, and calculates the similarity. If the representative feature amount is F t , the extracted image feature amount is F e , and the maximum value that can be taken by the feature amount is F max (however, the maximum value of each feature value is greater than 0), the similarity S is For example, it can obtain
  • the search unit 105 may acquire the representative feature amount Ft and the corresponding reliability C, and may not collate the representative feature amount having a reliability lower than a preset threshold value. In this case, the processing can be speeded up.
  • step S26 the search unit 105 extracts the maximum similarity among the calculated similarities, and specifies a representative feature amount corresponding to the extracted similarity.
  • step S27 the search unit 105 holds the performer name associated with the corresponding representative feature amount as a search result.
  • the search unit 105 matches the extracted image feature amount calculated by the feature amount calculation unit 111 with the representative feature amount related to the related image obtained by the representative feature amount determination unit 109 in the flow of FIG. Specifically, the search unit searches the performer information holding unit 110 via the performer information management unit 104 and extracts the performer name associated with the representative feature amount specified in step S26.
  • step S28 the display information generation unit 106 generates display information for displaying the search accuracy of the performer name simultaneously with the performer name from the performer name extracted by the search unit 105 and the display extracted image. Specifically, the display information generation unit 106 displays at least one of the display content determined from the performer name, the display position determined from the position of the display extracted image, and the reliability or similarity of the representative feature amount. From the matching result between the extracted image feature quantity and the representative feature quantity, the search accuracy of the performer name is calculated. Then, the display information generation unit 106 generates display information based on this search accuracy. At this time, the display information generation unit 106 basically generates display information for displaying the display contents with high search accuracy more prominently.
  • step S29 the display unit 103 displays the search result by displaying the display information generated by the display information generation unit 106.
  • the program display device 100 ends the series of processes when the processes up to the display of the names of the performers described above are performed for “all related images”.
  • the user can know the information of the person displayed on the display screen when he / she wants to know.
  • the display information generation unit 106 determines the display content from the performer name extracted by the search unit 105.
  • the display content is, for example, a description of a performer name, a role name, and a role.
  • the display information generation unit 106 determines the display position of the display content from the determined display content and the position of the display extracted image acquired by the search unit 105. For example, the display information generation unit 106 confirms whether the display content can be arranged in the order of upper, right, lower, and left in the display extracted image area, and determines the position where the display content can be arranged as the display position of the display content. “Placeable” means that the display content does not exceed the range of the display screen, does not overlap with other display content whose display position has already been determined, and is close to other display content whose display position has already been determined. It means that the conditions such as not being met.
  • the display information generation unit 106 determines the display form according to the accuracy of the searched display content.
  • the display form is a form in which display contents such as a display position, display contents, font type, font size, character color, background color, presence / absence of a border line, or border line color are displayed as a character string or the like.
  • the accuracy of the display content corresponds to at least one of the reliability of the representative feature quantity used by the search unit 105 when collating and the similarity calculated when the extracted image feature quantity and the representative feature quantity are collated. It is a fixed index.
  • the accuracy A used for determining the display form may be a value obtained by multiplying both values as shown in the following equation (13).
  • the coefficients ⁇ and ⁇ may be used as a value obtained by summing the weighted values of the reliability C and the similarity S as shown in the following equation (14).
  • the accuracy A can be reflected in the font size Fs of the performer name using the above formula (14).
  • the font size Fs of the performer name can be obtained by the following equation (15), where Fm is the maximum standard font size.
  • the degree of reflection of the reliability and similarity on the display information can be changed by changing the values of the coefficients ⁇ and ⁇ .
  • FIG. 6 is a table showing an example of display information generation in the above formula (15) when attention is paid only to the font size in the display form.
  • the display information generation example in which the display information generation unit 106 generates display information using at least one of the reliability C and the similarity S of the representative feature amount has been described.
  • the display information generation unit 106 calculates the reliability or the similarity based on the performer name acquired by the search unit 105 and the display extracted image. Therefore, which region is cut out as a display extraction image has an influence when calculating reliability and similarity. When the region of the display extraction image is inappropriate (for example, when only a part of the face is set as the region), the reliability and similarity may not be calculated correctly.
  • FIG. 8 is a diagram for explaining a display area and a display image when the display unit 103 is a television display unit.
  • a display area 140 is a television screen (display screen), and a display image 141 is a video of a program.
  • the display image 141 is displayed in the display area 140.
  • a person and a performer name 142 are displayed in the display area 140.
  • a region surrounding the human face in the display image 141 is a region cut out as the display extraction image 143.
  • FIG. 8 shows a case where the display image (program video) 141 is not displayed in the full display area (television screen) 140. In this case, the display screen can take a form in which the name of the performer is displayed around the display image 141.
  • An example of the case where the display image 141 is not displayed in the full display area 140 is a case where a broadcast video is reduced and displayed, such as a screen using a picture-in-picture function or a data broadcast display function in digital terrestrial broadcasting. .
  • the change in the display form of the performer name includes, for example, a change in transparency.
  • the character name of the performer is increased and the performer name is increased. Make it difficult to see.
  • the accuracy is high, that is, when it is determined that there is a high possibility that the search result is correct, the character name transparency is lowered to make the performer name clearly visible.
  • the display priority can be increased with respect to the name of the performer with high accuracy. If the accuracy is low, the display position of the performer name and the position of the image of the person to be displayed may be separated, and a lead line connecting the person image and the performer name may be displayed. In this case, if the accuracy is low, the lead line becomes long. Therefore, the difference in accuracy can be shown by the length of the lead line. Furthermore, when there are a plurality of search results, the difference in accuracy can be compared by a common display form. That is, when the accuracy changes due to the progress of the display of moving images, a plurality of display contents can be narrowed down to one display content.
  • the visibility is further improved. Further, the visibility is further improved by matching the frame of the area with the display color of the display content. Further, as a display form, if a role name is registered in the performer information holding unit 110, the role name may be written as the display content.
  • FIG. 9 is a diagram showing an example of the display of the search result displayed by the display unit 103.
  • FIG. 9 two persons are shown on the display screen 150 which is a television screen.
  • the performer name 151 is also written in the face area 153 of the left person.
  • the performer name 152 is also written in the face area 154 of the right person.
  • the performer name of FIG. 3 is displayed with a larger character size as the reliability is higher, and an example in which the performer name 152 has a higher reliability than the performer name 151 is shown.
  • the display example shown in FIG. 9 when the performer name 152 is displayed larger than the performer name 151, the user intuitively knows that the display content of the performer name 152 is correct. Can do. That is, the user can intuitively determine the correctness of the display content.
  • the method of displaying the performer name may be a change of the character or background color, a change of the transparency, a display position, etc. instead of or in combination with the change of the character size. In either case, the same effect can be obtained.
  • the program acquisition unit 101 acquires a program that is a moving image and program information including a performer name, and the related image acquisition unit 107 performs a performer name.
  • a related image is acquired from the network, and the feature amount calculation unit 108 calculates a feature amount of a related extracted image obtained by cutting out a human face area from the related image.
  • the representative feature amount determination unit 109 determines the representative feature amount and its reliability based on the feature amount of the related extracted image and the information on the acquisition destination of the related image, and the performer information management unit 104 determines The representative feature amount and the reliability are managed as performer information in association with the performer name.
  • the display image acquisition unit 102 acquires a display image from the frames constituting the moving image, and the feature amount calculation unit 111 calculates the feature amount of the display extracted image for the display image.
  • the search unit 105 calculates the similarity between the feature amount of the display extracted image calculated by the feature amount calculation unit 111 and the representative feature amount held in the performer information holding unit 110, and the similarity is the maximum.
  • the name of the performer associated with the representative feature amount is acquired.
  • the display information generation unit 106 generates display information based on at least one of reliability or similarity, the performer name acquired by the search unit 105, and the region of the display extracted image, and the display unit 103 The name of the person is displayed in association with the face area.
  • the program information display apparatus 100 obtains performer information from the program information of the moving image, obtains the performer's face image from the network, and obtains the image feature amount. And the program information display apparatus 100 specifies a performer from the image feature-value obtained by extracting a face image from a moving image, and superimposes and displays a performer name on a moving image. Thereby, the program information display apparatus 100 can display a performer name in association with a performer's area without preparing a performer's face image database in advance. Moreover, the program information display apparatus 100 reflects the reliability based on the database searched dynamically, when specifying a performer from the image feature-value obtained by extracting a face image from a moving image.
  • the program information display device 100 displays the correctness of the determination result of the performer according to the difference in display form.
  • the effect that a user can judge the determination result of a performer intuitively is acquired. For example, when the user does not agree with the name of the performer while watching the program, the user can respond to a request to know the name of the performer in association with the face on the screen. Moreover, it can be made to grasp
  • the representative feature amount determination unit 109 determines whether the related image is an image acquired from the official site and weights the related image, an effect of increasing the reliability of the feature amount can be obtained. Can do.
  • FIG. 10 is a diagram showing a configuration of a program information display system including a program information display device according to Embodiment 2 of the present invention.
  • the same components as those in FIG. 1 are denoted by the same reference numerals, and description of overlapping portions is omitted.
  • the program information display device 300 includes a search information holding unit 301 in addition to the program information display device 100 of FIG.
  • the search information holding unit 301 holds search results obtained by the search unit 105.
  • the search results by the search unit 105 are, as described above, the representative image feature amount, the reliability, the display extracted image region, the similarity, and the performer name.
  • FIG. 11 is a flowchart showing processing up to display of the performer name of the program information display apparatus 300. Steps that perform the same processing as the flow shown in FIG. 5 are denoted by the same step numbers, and description of overlapping portions is omitted.
  • step S23 the search unit 105 repeats the process between the start and end of the following loop for “all related images”.
  • step S31 the search unit 105 closes the area of the newly extracted display extraction image (hereinafter referred to as “new display extraction image”) and has been extracted in the past (referred to as past display extraction image). It is confirmed whether or not this area is registered in the search information holding unit 301.
  • step S ⁇ b> 32 when the past display extracted image area close to the new display extracted image area is registered in the search information holding unit 301 in step S ⁇ b> 32, the search unit 105 selects the new display extracted image area in step S ⁇ b> 33.
  • the extracted image feature quantity of the display extracted image is compared with the representative feature quantity of the corresponding past display extracted image, the similarity is calculated, and the process proceeds to step S26.
  • step S24 the search unit 105 proceeds to the following in step S24 and performs the performer name in FIG. This is the same as the process up to step S27.
  • step S27 the search unit 105 holds the performer name associated with the corresponding representative feature amount as a search result.
  • step S34 after determining the performer name, the search information retaining unit 301 retains the representative image feature value, the reliability, the region of the display extracted image, the similarity, and the performer name, and proceeds to step S28.
  • step S33 the search unit 105 calculates the maximum similarity and proceeds to step S26. This is because the past display extraction image adjacent to the area of the new display extraction image is almost always a face image of the same person. In this case, in step S ⁇ b> 26, the search unit 105 compares the similarity to the past display extracted image and the similarity to the related image registered in the search information holding unit 301.
  • the search unit 105 holds the larger similarity. Thereby, the search part 105 can make the similarity extraction process act only in the direction in which the similarity increases while capturing the face area of the same person.
  • the search unit 105 discards the registration data related to the past past display extraction area, that is, the process of registering invalid data instead of the process of holding the search result (the process of step S34). It becomes processing.
  • the search unit 105 includes the search information holding unit 301 that holds the search result.
  • the search unit 105 searches the search result held in the search information holding unit 301 only for the display extracted image close to the search result.
  • the search unit 105 does not need to compare the display extracted image with the similarity of the feature amounts of all registered representative images, and can speed up the processing.
  • FIG. 12 is a diagram showing a configuration of a program information display system including a program information display device according to Embodiment 3 of the present invention.
  • the program information display system includes a program information display device 400, a network 200, an image search device 500, an image server 220, a program information server 230, a program guide server 240, and a broadcast station 250. That is, in the program information display system according to the present embodiment, a program information display device 400 and an image search device 500 are arranged instead of the program information display device 100 and the image search device 210 of the first embodiment.
  • the program information display device 400 is a digital broadcast receiver that reproduces program information transmitted from the broadcast station 250.
  • the program information display device 400 includes a program acquisition unit 101, a display image acquisition unit 102, a display unit 103, a search unit 105, a display information generation unit 106, and a feature amount calculation unit 111. That is, the program information display device 400 has a configuration including a functional unit related to image display among the functional units of the program information display device 100 of the first embodiment.
  • the image search device 500 is an image search site.
  • the image search apparatus 500 includes a performer information management unit 104, a related image acquisition unit 107, a feature amount calculation unit 108, a representative feature amount determination unit 109, a performer information holding unit 110, an image information holding unit 501, and an image search unit 502. It is comprised with. That is, the image search device 500 has a configuration including a function unit related to related image search, an image information holding unit 501, and an image search unit 502 among the function units of the program information display device 100 of the first embodiment. is doing.
  • the image information holding unit 501 and the image search unit 502 correspond to the functional units of the image search apparatus 210 according to the first embodiment.
  • the program information display system has a configuration in which the function unit related to the related image search of the program information display device 400 of the first embodiment is moved to the image search device 210 of the first embodiment. ing.
  • the program information display device 400 and the image search device 500 may each be configured such that the CPU executes a program stored in a storage medium such as a ROM, similarly to the program information display device 100 of the first embodiment.
  • the program information display device 100 of the first embodiment or the program information display device 300 of the second embodiment may be combined with the image search device 500 of the present embodiment. .
  • the related image acquisition unit 107 of the image search apparatus 500 acquires an image via the network 200 based on a URL link or the like.
  • the related image acquisition unit 107 holds the acquired image, the keyword obtained from the description of the corresponding image or the content of the corresponding page, and the URL of the image in the image information holding unit 501.
  • the image search unit 502 When searching for an image, the image search unit 502 receives program information from the program information display device 400 via the network 200, and returns a corresponding URL as image information from the image information holding unit 501 based on the received program information. .
  • This configuration eliminates the processing from the program information display device 400 until the representative feature amount is determined. Further, the search unit 105 receives the representative feature amount and the reliability via the network 113 using the performer name as a keyword. These points are changes from the first embodiment.
  • the processing until the representative feature amount is determined is performed by the image search device 500, so that the processing of the program information display device 400 can be expected to be reduced and speeded up due to the distribution of functions.
  • the example in which the image information holding unit 501 and the performer information holding unit 110 are configured separately has been described, but these may be integrated.
  • the example in which the image search device 500 receives program information from the program information display device 400 has been described.
  • the program information may be received from the program information server 230, and the process of determining the representative feature amount may be executed in advance to receive the performer name and the program name from the program information display device 400.
  • the names of the program information display device and the program information display method are used.
  • the program information display device may be an image search device, a program playback device, and the program information display method may be a program information search method or the like.
  • each part constituting the program information display apparatus and method for example, the type, number and connection method of the program acquisition part and performer information holding part are not limited.
  • the detailed program information constituting the program guide is acquired from the program guide server via the network.
  • the detailed program information when the detailed program information is superimposed on the broadcast wave, it may be acquired from the broadcast wave.
  • the related image may be acquired from the media instead of via the network.
  • the broadcast station and the program guide server may not be included in the overall configuration of the program information display system.
  • the program information device is not limited to application to a television receiver, and can be applied to other various devices that display images on which performers are displayed.
  • the program information display device can be applied to a device that displays or reproduces a moving image, such as a BD (Blue-ray Disc) player, a DVD player, or a hard disk recorder.
  • the program information display device is also applicable to portable terminals such as mobile phones / PHS (Personal Handy-Phone System), portable information terminals (hereinafter referred to as PDA (Personal Digital Assistants)), personal computers, and portable game machines. can do.
  • the program information display device and the program information display method described above are also realized by a program for causing the program information display method to function.
  • This program is stored in a computer-readable recording medium.
  • a program information display apparatus and a program information display method extract a person's face area from a moving image and display a performer name in association with the person's face area, a mobile phone, a DVD player, It can be applied to video playback terminals such as personal computers and game machines.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Theoretical Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

A program information display device that displays, with good accuracy and little noise, the names of performers in a program by associating them with the faces on the screen. Device (100) is equipped with a program acquisition unit (101) that acquires information comprising a moving-image program, and the names of performers, a related image-acquisition unit (107) that acquires related images from a network (200), a feature-value calculation unit (108) that calculates feature values for images of persons extracted from the related images, a representative feature value determination unit (109) that determines a representative feature value and reliability thereof from the feature values and the source from which the related images were obtained, a performer information management unit (104) that associates the determined values with the names of performers and manages the same, a display image acquisition unit (102) that acquires a display image, a feature value calculation unit (111) that calculates feature values for images of persons extracted from the display image, a lookup unit (105) that retrieves the name of the performer for whom the degree of similarity of the representative feature value is the greatest, a display information generation unit (106) that generates display information based on the reliability or the degree of similarity with the area of a second extracted image of a performer, and a display unit (103) that displays the display image as well as the display information.

Description

番組情報表示装置および番組情報表示方法Program information display apparatus and program information display method
 本発明は、番組情報を表示する番組情報表示装置および番組情報表示方法に関し、例えば、テレビジョン受信機、DVD(Digital Versatile Disc)プレイヤ、ハードディスクレコーダ等の、動画像を表示または再生する番組情報表示装置およびこのような装置における番組情報表示方法に関する。 The present invention relates to a program information display device and a program information display method for displaying program information, for example, a program information display for displaying or reproducing a moving image, such as a television receiver, a DVD (Digital Versatile Disc) player, a hard disk recorder, etc. The present invention relates to an apparatus and a program information display method in such an apparatus.
 従来、テレビ等で視聴している番組の出演者が未知であった場合に、その出演者の名前を知りたいという要求は一般的なものである。この見知らぬ出演者の情報を得る方法として、例えば、新聞のテレビ欄あるいはEPG(Electrical Program Guide:電子番組表)から出演者名を得ることが一般的であった。 Conventionally, when a performer of a program being viewed on a TV or the like is unknown, a request to know the name of the performer is general. As a method for obtaining information on this unknown performer, for example, it is common to obtain the name of a performer from a television column of a newspaper or EPG (Electrical Program Guide).
 このような方法においては、出演者の名前と写真が掲載されているとは限らないため、顔と名前を一致させることが困難となる。ユーザは、番組の情報から出演者名を知ることができた場合であっても、例えばインターネットにアクセスして出演者名をキーワードにより画像を検索することで、顔と名前を一致させることが可能となる場合もある。しかし、ユーザにとってその手間は煩雑であった。 In such a method, the name and photo of the performer are not always posted, so it is difficult to match the face and name. Even if the user is able to know the name of the performer from the program information, the name can be matched with the face by accessing the Internet and searching for an image with the name of the performer, for example. It may become. However, it is troublesome for the user.
 この状況を解決する方法として、動画像から顔画像を抽出して予め登録された顔画像と固有情報を検索し、固有情報を顔画像の近傍に表示する情報表示方法がある(例えば、特許文献1)。 As a method for solving this situation, there is an information display method for extracting a face image from a moving image, searching for a pre-registered face image and unique information, and displaying the unique information in the vicinity of the face image (for example, Patent Documents). 1).
 また、別の方法として、画像特徴量を基にして、画像の検索する対象である画像特徴量データベースを最新の情報に更新するために、画像リストと画像とを外部装置から定期的に取得する画像検索装置がある(例えば、特許文献2)。
特開2006-293912号公報 特開2006-185320号公報
As another method, an image list and an image are periodically acquired from an external device in order to update the image feature database to be searched for images based on the image feature to the latest information. There is an image search device (for example, Patent Document 2).
JP 2006-293912 A JP 2006-185320 A
 しかしながら、このような従来の番組情報表示装置にあっては、以下のような課題があった。 However, such a conventional program information display device has the following problems.
 特許文献1記載の情報表示法では、画像特徴量データベースが固定的である。様々な番組に日々新たな出演者が登場する状況において、すべての出演者に関する画像特徴量を保持していることは現実的ではない。 In the information display method described in Patent Document 1, the image feature database is fixed. In a situation where new performers appear in various programs every day, it is not realistic to hold the image feature amounts for all the performers.
 この解決手段として上記特許文献2を用いることで、検索対象の画像特徴量データベースを定期的に更新することが可能となる。しかし、特許文献2記載の画像検索装置は、キーワードから画像特徴量を検索して、画像特徴量を基に類似画像を検索するためのデータベースであり、画像特徴量からキーワード(例えば、出演者名)を検索するものではない。また、上記画像検索装置は、キーワードを基に取得した画像をネットワークで検索すると、関連するが出演者を示すものではない画像(ノイズ)も多く検索される。ところが、検索方法が異なることから、ノイズに対する考慮がなされていない。 By using the above-mentioned Patent Document 2 as this solution, it is possible to periodically update the image feature amount database to be searched. However, the image search device described in Patent Document 2 is a database for searching for an image feature quantity from a keyword and searching for a similar image based on the image feature quantity. ) Does not search. Further, when the image search apparatus searches the network for images acquired based on keywords, many images (noise) that are related but do not indicate a performer are also searched. However, since the search method is different, no consideration is given to noise.
 また、上記方法を組み合わせても、画像特徴量が一致したと判断されるまで出演者名が表示されないという課題がある。 Moreover, even if the above methods are combined, there is a problem that the name of the performer is not displayed until it is determined that the image feature amounts match.
 本発明の目的は、番組視聴中、登場人物と出演者名が一致しないとき等、画面上の顔と対応付けて出演者名を知りたいという要求に対し、ノイズの少なく、精度の良い検索を可能とする番組情報表示装置および番組情報表示方法を提供することである。 The purpose of the present invention is to perform a search with less noise and high accuracy in response to a request to know a performer name in association with a face on the screen, such as when a character does not match a performer name while watching a program. It is an object to provide a program information display device and a program information display method that can be used.
 本発明の番組情報表示装置は、動画像である番組と、出演者名を含む番組情報と、を取得する番組取得部と、前記出演者名に基づいて、ネットワークから関連画像を取得する関連画像取得部と、前記関連画像から人物の領域を切り出した第1の抽出画像の特徴量を算出する第1特徴量算出部と、前記第1の抽出画像の特徴量と前記関連画像の入手先の情報とに基づいて、代表特徴量とその信頼度とを決定する代表特徴量決定部と、決定された前記代表特徴量および信頼度を、前記出演者名と対応付けて出演者情報として管理する出演者情報管理部と、前記動画像を構成するフレームから表示画像を取得する表示画像取得部と、前記表示画像から人物の領域を切り出した第2の抽出画像の特徴量を算出する第2特徴量算出部と、前記第2特徴量算出部により算出された前記第2の抽出画像の特徴量と、前記出演者情報管理部に保持された前記代表特徴量と、の類似度を算出し、前記類似度が最大となる代表特徴量に関連付けられた出演者名を取得する検索部と、前記信頼度または前記類似度の少なくとも一方と、前記検索部により取得された前記出演者名と、前記第2の抽出画像の領域と、に基づいて表示情報を生成する表示情報生成部と、前記表示画像および前記表示情報を表示する表示部と、を備える構成を採る。 The program information display device of the present invention includes a program acquisition unit that acquires a program that is a moving image and program information including a performer name, and a related image that acquires a related image from a network based on the performer name. An acquisition unit; a first feature amount calculation unit that calculates a feature amount of a first extracted image obtained by cutting a person's region from the related image; and a feature amount of the first extracted image and a source of the related image Based on the information, a representative feature amount determining unit that determines a representative feature amount and its reliability, and the determined representative feature amount and reliability are managed as performer information in association with the performer name. A performer information management unit, a display image acquisition unit that acquires a display image from a frame that constitutes the moving image, and a second feature that calculates a feature amount of a second extracted image obtained by extracting a person's region from the display image A quantity calculation unit and the second feature A similarity between the feature amount of the second extracted image calculated by the calculation unit and the representative feature amount held in the performer information management unit is calculated, and the representative feature amount that maximizes the similarity A search unit for acquiring a performer name associated with the at least one of the reliability or the similarity, the performer name acquired by the search unit, and the region of the second extracted image. A configuration is provided that includes a display information generation unit that generates display information based on the display information and a display unit that displays the display image and the display information.
 本発明の番組情報表示方法は、動画像である番組と、出演者名を含む番組と、を取得するステップと、前記出演者名に基づいて、ネットワークから関連画像を取得するステップと、前記関連画像から人物の領域を切り出した第1の抽出画像の特徴量を算出する第1特徴量算出ステップと、前記第1の抽出画像の特徴量と前記関連画像の入手先の情報とに基づいて、代表特徴量とその信頼度とを決定するステップと、決定された前記代表特徴量および信頼度を、前記出演者名と対応付けて出演者情報として管理するステップと、前記動画像を構成するフレームから表示画像を取得するステップと、前記表示画像から人物の領域を切り出した第2の抽出画像の特徴量を算出する第2特徴量算出ステップと、前記第2特徴量算出ステップにより算出された前記第2の抽出画像の特徴量と、保持された前記代表特徴量と、の類似度を算出し、前記類似度が最大となる代表特徴量に関連付けられた出演者名を取得するステップと、前記信頼度または前記類似度の少なくとも一方と、前記検索部により取得された前記出演者名と、前記第2の抽出画像の領域と、に基づいて表示情報を生成するステップと、前記表示画像および前記表示情報を表示するステップとを有する。 The program information display method of the present invention includes a step of acquiring a program that is a moving image and a program including a performer name, a step of acquiring a related image from a network based on the performer name, and the related Based on a first feature amount calculating step for calculating a feature amount of a first extracted image obtained by cutting out a person's region from the image, and on the feature amount of the first extracted image and the information on the source of the related image, A step of determining a representative feature amount and its reliability, a step of managing the determined representative feature amount and reliability as performer information in association with the performer name, and a frame constituting the moving image A display image is obtained from the display image; a second feature value calculating step for calculating a feature value of a second extracted image obtained by cutting out a person region from the display image; and a second feature value calculating step. Calculating the degree of similarity between the feature quantity of the second extracted image and the retained representative feature quantity, and obtaining a performer name associated with the representative feature quantity having the maximum similarity Generating display information based on at least one of the reliability or the similarity, the performer name acquired by the search unit, and the region of the second extracted image, and the display Displaying an image and the display information.
 本発明によれば、事前に出演者の顔画像データベースを用意する必要がなく、出演者の領域に関連付けて出演者名を表示する際に、出演者の判定結果の正しさを直観的に判断できる効果が得られる。 According to the present invention, it is not necessary to prepare a performer's face image database in advance, and intuitively determine the correctness of the performer's determination result when displaying the performer's name in association with the performer's area. The effect that can be obtained.
 また、動画像から顔画像を抽出して得られる画像特徴量から出演者を特定する際に、動的に検索するデータベースに基づいた信頼度を反映させることにより、ノイズが少なく、精度の良い検索を実現できる。 In addition, when specifying a performer from the image feature value obtained by extracting a face image from a moving image, the search is performed with less noise and high accuracy by reflecting the reliability based on a database that is dynamically searched. Can be realized.
本発明の実施の形態1に係る番組情報表示装置を備える番組情報表示システムの構成を示す図The figure which shows the structure of the program information display system provided with the program information display apparatus which concerns on Embodiment 1 of this invention. 本実施の形態1に係る番組情報表示装置の電子番組表の一例を示す図The figure which shows an example of the electronic program guide of the program information display apparatus which concerns on this Embodiment 1. 本実施の形態1に係る番組情報表示装置の出演者情報保持部に保持される出演者情報の一例を示す図The figure which shows an example of the performer information hold | maintained at the performer information holding part of the program information display apparatus which concerns on this Embodiment 1. FIG. 本実施の形態1に係る番組情報表示装置の代表特徴量を決定するまでの処理動作を示すフロー図The flowchart which shows the processing operation until the representative feature-value of the program information display apparatus which concerns on this Embodiment 1 is determined. 本実施の形態1に係る番組情報表示装置の出演者名の表示までの処理動作を示すフロー図The flowchart which shows the processing operation until the display of the performer name of the program information display apparatus which concerns on this Embodiment 1. 本実施の形態1に係る番組情報表示装置の表示情報のフォントサイズのみに着目した場合の表示情報生成例を示す図The figure which shows the example of display information generation at the time of paying attention only to the font size of the display information of the program information display apparatus which concerns on this Embodiment 1. FIG. 本実施の形態1に係る番組情報表示装置の代表特徴量の算出の例を説明するための図The figure for demonstrating the example of calculation of the representative feature-value of the program information display apparatus which concerns on this Embodiment 1. FIG. 本実施の形態1に係る番組情報表示装置の表示部がテレビの表示部である場合の表示領域と表示画像を説明する図The figure explaining the display area and display image in case the display part of the program information display apparatus which concerns on this Embodiment 1 is a display part of a television. 本実施の形態1に係る番組情報表示装置の表示部により表示される検索結果の表示の例を示す図The figure which shows the example of a display of the search result displayed by the display part of the program information display apparatus which concerns on this Embodiment 1. FIG. 本発明の実施の形態2に係る番組情報表示装置を備える番組情報表示システムの構成を示す図The figure which shows the structure of the program information display system provided with the program information display apparatus which concerns on Embodiment 2 of this invention. 本実施の形態2に係る番組情報表示装置の出演者名の表示までの処理を示すフロー図The flowchart which shows the process until the display of the performer name of the program information display apparatus which concerns on this Embodiment 2. 本発明の実施の形態3に係る番組情報表示装置を備える番組情報表示システムの構成を示す図The figure which shows the structure of the program information display system provided with the program information display apparatus which concerns on Embodiment 3 of this invention.
 以下、本発明の各実施の形態について図面を参照して詳細に説明する。 Hereinafter, embodiments of the present invention will be described in detail with reference to the drawings.
 (実施の形態1)
 図1は、本発明の実施の形態1に係る番組情報表示装置を備える番組情報表示システムの構成を示す図である。本実施の形態に係る番組情報表示装置は、本発明を、デジタル放送の電波を受信可能なテレビジョン受信機に適用した例である。
(Embodiment 1)
FIG. 1 is a diagram showing a configuration of a program information display system provided with a program information display device according to Embodiment 1 of the present invention. The program information display apparatus according to this embodiment is an example in which the present invention is applied to a television receiver capable of receiving digital broadcast radio waves.
 図1において、番組情報表示システムは、番組情報表示装置100、ネットワーク200、画像検索装置210、画像サーバ220、番組情報サーバ230、番組表サーバ240および放送局250を備えて構成される。 1, the program information display system includes a program information display device 100, a network 200, an image search device 210, an image server 220, a program information server 230, a program guide server 240, and a broadcast station 250.
 番組情報表示装置100は、放送局250から送信される番組情報を再生するデジタル放送受信機である。番組情報表示装置100の詳細な構成については、後述する。 The program information display device 100 is a digital broadcast receiver that reproduces program information transmitted from the broadcast station 250. The detailed configuration of the program information display device 100 will be described later.
 ネットワーク200は、インターネットまたは専用回線からなる通信ネットワークである。より具体的には、ネットワーク200は、移動体通信網、公衆電話網、LANまたはインターネット等から構成するネットワークである。ネットワーク200は、有線系または無線系のいずれでもよく、そのプロトコルの種類は特に問わない。また、ネットワーク200のアクセス回線としては、FTTH(Fiber To The Home)、HFC(Hybrid Fiber Coax:光同軸ケーブル)、およびADSL(Asymmetric Digital Subscriber Line)等の大容量回線が利用可能である。 The network 200 is a communication network composed of the Internet or a dedicated line. More specifically, the network 200 is a network composed of a mobile communication network, a public telephone network, a LAN, the Internet, or the like. The network 200 may be either a wired system or a wireless system, and the type of protocol is not particularly limited. Further, as an access line of the network 200, a large capacity line such as FTTH (Fiber To The Home), HFC (Hybrid Fiber Coax), and ADSL (Asymmetric Digital Subscriber Line) can be used.
 画像検索装置210は、画像検索サイトである。より具体的には、画像検索装置210は、ネットワーク200を介して、各種サイトが保持する画像、その説明情報、および画像のURL(Uniform Resource Locator)を収集し、保持する画像検索サイトである。画像検索装置210の検索の対象となるサイトには、画像サーバ220、番組情報サーバ230をはじめとする複数のサイトが含まれる。画像検索装置210は、収集した画像に関する情報を、キーワードに基づいて検索し、検索結果を出力する。 The image search device 210 is an image search site. More specifically, the image search device 210 is an image search site that collects and holds images held by various sites, their description information, and URLs (Uniform Resource Locator) of images via the network 200. The sites to be searched by the image search device 210 include a plurality of sites including the image server 220 and the program information server 230. The image search device 210 searches for information on the collected image based on the keyword, and outputs the search result.
 画像に関する情報には、画像そのものや、画像のURLが含まれる。また、画像検索装置210の画像に関する情報の収集先には、画像サーバ220も含まれる。但し、画像検索装置210は、画像サーバ220に掲載されている情報が適正でない場合もあるため、検索アルゴリズムを工夫することにより、検索精度の向上を図っている。 The information about the image includes the image itself and the URL of the image. In addition, the image server 220 is included in the collection destination of the information related to the image of the image search apparatus 210. However, since the information posted on the image server 220 may not be appropriate, the image search device 210 attempts to improve search accuracy by devising a search algorithm.
 画像サーバ220は、掲載するウェブページを公開している一般サイトである。より具体的には、画像サーバ220は、不特定多数の一般ユーザあるいは企業等により管理されており、何らかの画像を保持してネットワーク200に向けて公開しているサイトである。公開サイトは、必ずしも情報を適正に表現しているとは限らない。しかし、公開サイトは、例えば特定の放送番組の公式サイトに比べるとはるかに多く存在しており、ネットワーク200全体として膨大な数の情報を保有している。 The image server 220 is a general site that publishes web pages to be posted. More specifically, the image server 220 is a site that is managed by an unspecified number of general users or companies, and holds some images and publishes them to the network 200. Public sites do not always represent information properly. However, there are far more public sites than, for example, an official site of a specific broadcast program, and the network 200 as a whole has a huge amount of information.
 番組情報サーバ230は、放送局250が放送する番組(以下適宜「番組」という)の公式サイトである。より具体的には、番組情報サーバ230は、番組の公式サイトであり、一般的には番組の権利者であるテレビ局や番組制作会社が管理している。従って、公式サイトは、番組の内容を適正に表現しているサイトの一つであると言える。ただし、番組によっては、公式サイトが存在しないこともあり、存在している場合であっても内容の充実度は様々である。ここでは、公式サイトを番組のサイトとして説明しているが、番組の出演者の公式サイトも、出演者に関する情報を適正に表現しているサイトであることから、同様に扱うことができる。 The program information server 230 is an official site for programs broadcast by the broadcast station 250 (hereinafter referred to as “programs” as appropriate). More specifically, the program information server 230 is an official program site, and is generally managed by a television station or a program production company that is the right holder of the program. Therefore, it can be said that the official site is one of the sites that appropriately express the contents of the program. However, depending on the program, the official website may not exist, and even if it exists, the level of content fulfillment varies. Here, the official site is described as a program site, but the official site of a program performer can also be handled in the same manner because it is a site that appropriately expresses information about the performer.
 番組表サーバ240は、放送局250を含む各局の番組の番組表を掲載するサイトである。より具体的には、番組表サーバ240は、日時あるいは地域等を基に、該当する番組の番組詳細情報の一覧であるネット番組表を、ネットワーク200を介して提供するサイトである。番組詳細情報は、番組名、出演者名、公式サイトのURL等を含む。 The program guide server 240 is a site that publishes a program guide of programs of each station including the broadcast station 250. More specifically, the program guide server 240 is a site that provides, via the network 200, a net program guide that is a list of detailed program information of the corresponding program based on the date and time or the region. The detailed program information includes a program name, a performer name, a URL of an official site, and the like.
 上記各サーバ220,230,240は、サーバ全体を制御するコンピュータからなる制御部に、ネットワーク200を介して接続されている。サーバ220,230,240は、例えばインターネット上のURLにアクセスされたデータを送受信する通信インタフェース、およびデータを蓄積するデータベース(DB)等からそれぞれ構成される。 The servers 220, 230, and 240 are connected via a network 200 to a control unit that includes a computer that controls the entire server. The servers 220, 230, and 240 are each composed of, for example, a communication interface that transmits and receives data accessed by a URL on the Internet, and a database (DB) that stores data.
 番組情報表示装置100は、以下の構成を採る。 The program information display device 100 has the following configuration.
 番組情報表示装置100は、番組取得部101、表示画像取得部102、表示部103、出演者情報管理部104、検索部105、表示情報生成部106、関連画像取得部107、特徴量算出部108(第1特徴量算出部)、代表特徴量決定部109、出演者情報保持部110、および特徴量算出部111(第2特徴量算出部)から構成される。また、図示は省略されているが、番組情報表示装置100は、ユーザの入力操作を受け付ける入力装置、および情報を記録する外部記録装置等の、各種機器を接続するためのインタフェースを備えている。入力装置は、数値キーや十字キー等からなるキーボードあるいはリモコン装置である。 The program information display apparatus 100 includes a program acquisition unit 101, a display image acquisition unit 102, a display unit 103, a performer information management unit 104, a search unit 105, a display information generation unit 106, a related image acquisition unit 107, and a feature amount calculation unit 108. (First feature value calculating unit), representative feature value determining unit 109, performer information holding unit 110, and feature value calculating unit 111 (second feature value calculating unit). Although not shown, the program information display device 100 includes an interface for connecting various devices such as an input device that accepts user input operations and an external recording device that records information. The input device is a keyboard or a remote control device including numeric keys, cross keys, and the like.
 番組取得部101は、番組および番組情報の放送を受信し、番組および番組情報をデコードするチューナである。番組は、動画像からなる。番組情報は、例えば、EPG等の電子番組表に含まれる、番組に関する各種情報である。電子番組表は、番組名、チャンネル、放送日、放送の開始時刻と終了時刻、出演者名等から構成される。以下、番組取得部101は、放送信号から電子番組表を抽出し、抽出した電子番組表から番組情報を取得するものとする。 The program acquisition unit 101 is a tuner that receives broadcasts of programs and program information and decodes the programs and program information. The program consists of moving images. Program information is, for example, various types of information related to programs included in an electronic program guide such as EPG. The electronic program guide is composed of a program name, a channel, a broadcast date, a broadcast start time and end time, a performer name, and the like. Hereinafter, it is assumed that the program acquisition unit 101 extracts an electronic program guide from a broadcast signal and acquires program information from the extracted electronic program guide.
 番組取得部101は、内部にメモリ(図示略)を有し、このメモリに、取得した番組情報を保存する。番組取得部101が、例えばEPG等から番組情報を取得する場合、電子番組表に出演者として記載された名前は、番組の出演者名として取得する対象となるが、それ以外は、番組の出演者名として取得する対象とならない。 The program acquisition unit 101 has a memory (not shown) inside, and stores the acquired program information in this memory. For example, when the program acquisition unit 101 acquires program information from EPG or the like, the name described as a performer in the electronic program guide is a target to be acquired as the performer name of the program. It is not subject to acquisition as a person's name.
 図2は、電子番組表の一例を示す図である。図2に示すように、この電子番組表は、番組名、あらすじ、出演者およびURLのフィールド名について各設定値を持つ。図2の例では、番組名として「グッドモーニングトーキョー」が記述され、出演者として「タケシ(松山太郎)、ヒロコ(竹田花子)、ユリ(梅川良子)」が記載され、URLとして番組の公式サイトのアドレスが記載されている。 FIG. 2 is a diagram showing an example of an electronic program guide. As shown in FIG. 2, this electronic program guide has set values for program names, synopsis, performers, and URL field names. In the example of FIG. 2, “Good Morning Tokyo” is described as the program name, “Takeshi (Taro Matsuyama), Hiroko (Hanako Takeda), Yuri (Ryoko Umekawa)” as the performers, and the official website of the program as the URL The address is listed.
 表示画像取得部102は、番組取得部101が取得した番組から、表示する動画像を構成するフレームを抽出し、抽出したフレームから、TV画面に表示する番組の画像(以下「表示画像」という)を取得する。 The display image acquisition unit 102 extracts a frame constituting a moving image to be displayed from the program acquired by the program acquisition unit 101, and an image of the program to be displayed on the TV screen (hereinafter referred to as “display image”) from the extracted frame. To get.
 表示部103は、表示画像取得部102で抽出されたフレーム単位の動画像(表示画像)を表示するディスプレイである。表示部103は、表示画像と表示情報とを同時に表示する。 The display unit 103 is a display that displays a moving image (display image) in units of frames extracted by the display image acquisition unit 102. The display unit 103 displays a display image and display information at the same time.
 出演者情報管理部104は、番組取得部101に保存された番組情報から、出演者名、および公式サイトのURL等の情報を抽出する。出演者情報管理部104は、代表特徴量決定部109から得られる、関連画像に関する代表特徴量および信頼度を、出演者名と対応付けて出演者情報保持部110に保持する。 The performer information management unit 104 extracts information such as the performer name and the URL of the official site from the program information stored in the program acquisition unit 101. The performer information management unit 104 stores the representative feature amount and reliability related to the related image obtained from the representative feature amount determination unit 109 in the performer information holding unit 110 in association with the performer name.
 ここで、関連画像とは、番組の出演者に関連する画像であり、その出演者の顔画像が含まれている可能性が高い画像である。代表特徴量とは、同一の出演者に関連する複数の関連画像から抽出された顔画像(以下「関連抽出画像」という)の画像特徴量を代表する値であり、その出演者の顔画像の画像特徴量に一致する可能性が高い画像特徴量である。信頼度とは、代表特徴量が出演者の実際の顔画像の画像特徴量に一致していることについての信頼度である。 Here, the related image is an image related to the performer of the program, and is an image that is likely to include the face image of the performer. The representative feature amount is a value representing the image feature amount of a face image extracted from a plurality of related images related to the same performer (hereinafter referred to as “related extracted image”). This is an image feature amount that is highly likely to match the image feature amount. The degree of reliability is the degree of reliability that the representative feature amount matches the image feature amount of the actual face image of the performer.
 検索部105は、関連抽出画像のうち、表示画像から抽出された顔画像(以下「表示抽出画像」という)に類似する顔画像に関連する出演者名を検索し、検索結果を一時的に保持する。具体的には、検索部105は、特徴量算出部111により、表示抽出画像の領域および画像特徴量(以下「抽出画像特徴量」という)を求める。そして、検索部105は、出演者情報管理部104により、関連抽出画像の代表特徴量を取得する。そして、検索部105は、抽出画像特徴量と代表特徴量との類似度を算出し、類似度が最大となる代表特徴量に関連付けられた出演者名を取得する。すなわち、検索部105は、表示抽出画像に類似する関連抽出画像に対応する出演者名を取得する。 The search unit 105 searches for a performer name related to a face image similar to the face image extracted from the display image (hereinafter referred to as “display extracted image”) from the related extracted images, and temporarily stores the search result. To do. Specifically, the search unit 105 uses the feature amount calculation unit 111 to obtain a region of the display extracted image and an image feature amount (hereinafter referred to as “extracted image feature amount”). And the search part 105 acquires the representative feature-value of a related extraction image by the performer information management part 104. FIG. Then, the search unit 105 calculates the similarity between the extracted image feature quantity and the representative feature quantity, and acquires the performer name associated with the representative feature quantity having the maximum similarity. That is, the search part 105 acquires the performer name corresponding to the related extraction image similar to a display extraction image.
 表示情報生成部106は、代表特徴量の信頼度または類似度の少なくとも一方と、検索部105で取得した出演者名と表示抽出画像の領域とから、表示部103に表示する表示情報を生成する。表示情報生成部106による表示情報生成例については、図6により後述する。 The display information generation unit 106 generates display information to be displayed on the display unit 103 from at least one of the reliability or similarity of the representative feature amount, the performer name acquired by the search unit 105, and the display extracted image area. . An example of display information generation by the display information generation unit 106 will be described later with reference to FIG.
 関連画像取得部107は、ネットワーク200と通信して関連画像を取得する。関連画像取得部107は、ネットワーク200に接続して、番組情報を基に、出演者に関連する1つ以上の関連画像を取得する。 The related image acquisition unit 107 acquires a related image by communicating with the network 200. The related image acquisition unit 107 is connected to the network 200 and acquires one or more related images related to the performer based on the program information.
 特徴量算出部108,111は、画像から人物の顔領域を切り出した抽出画像の画像特徴量を算出する。特徴量算出部108と特徴量算出部111との間では、画像特徴量を算出するためのアルゴリズムは同じであるが、処理対象が異なる。特徴量算出部108は、ネットワーク200から取得した関連画像から人物の顔領域を切り出した関連抽出画像に対して、画像特徴量の算出を行う。特徴量算出部111は、TV画面に表示している表示画像から人物の顔領域を切り出した表示抽出画像に対して、画像特徴量の算出を行う。ここで、特徴量算出部108,111は、関連画像または表示画像に、複数人の人物が映っている場合には、その複数人のそれぞれについて画像特徴量を算出する。 Feature amount calculation units 108 and 111 calculate the image feature amount of an extracted image obtained by cutting out a human face area from an image. The algorithm for calculating the image feature amount is the same between the feature amount calculation unit 108 and the feature amount calculation unit 111, but the processing target is different. The feature amount calculation unit 108 calculates an image feature amount for a related extracted image obtained by cutting out a human face area from a related image acquired from the network 200. The feature amount calculation unit 111 calculates an image feature amount for a display extracted image obtained by cutting out a human face area from a display image displayed on the TV screen. Here, when a plurality of persons are shown in the related image or the display image, the feature amount calculation units 108 and 111 calculate the image feature amount for each of the plurality of persons.
 代表特徴量決定部109は、関連抽出画像の画像特徴量に重み付けをした値を比較して、代表特徴量および信頼度を決定する。具体的には、代表特徴量決定部109は、関連抽出画像(顔領域)の画像特徴量に対して重み付けを行う。重み付けの方法としては、例えば基の関連画像の取得元が公式サイトである場合に、その関連抽出画像の個数を増やした状態で代表特徴量を算出する方法がある。この方法によると、該当の関連抽出画像の画像特徴量の個数が他の関連抽出画像の画像特徴量の個数よりも増えることで、関連抽出画像の画像特徴量に対して重みを付けることができる。また、代表特徴量の決定方法としては、該当する関連抽出画像のすべての画像特徴量に対して、相加平均を用いる方法、あるいは個数が最大となる値を用いる方法が考えられる。 The representative feature amount determination unit 109 compares the weighted value of the image feature amount of the related extracted image to determine the representative feature amount and the reliability. Specifically, the representative feature amount determination unit 109 performs weighting on the image feature amount of the related extracted image (face region). As a weighting method, for example, when the source of the base related image is an official site, there is a method of calculating the representative feature amount in a state where the number of the related extracted images is increased. According to this method, the number of image feature quantities of the relevant extracted image is larger than the number of image feature quantities of the other related extracted images, so that the image feature quantities of the related extracted images can be weighted. . Further, as a representative feature amount determination method, a method using an arithmetic average or a method using a value that maximizes the number of all the image feature amounts of the relevant extracted image can be considered.
 出演者情報保持部110は、ハードディスクあるいはメモリにより構成され、出演者情報管理部104で番組情報から抽出された出演者名に対応付けて、代表特徴決定部109で決定された代表特徴量およびその信頼度を保持する。以下、出演者名、代表特徴量、および信頼度を、適宜、「出演者情報」と総称する。 The performer information holding unit 110 is configured by a hard disk or a memory, and the representative feature amount determined by the representative feature determination unit 109 in association with the performer name extracted from the program information by the performer information management unit 104 and its performer Maintain confidence. Hereinafter, the performer name, the representative feature quantity, and the reliability are collectively referred to as “performer information” as appropriate.
 図3は、出演者情報保持部110に保持される出演者情報の一例を示す図である。図3に示す出演者名は、図2の電子番組表の出演者の出演者名(役名)「松山太郎(タケシ)、竹田花子(ヒロコ)、梅川良子(ユリ)」に、それぞれ代表特徴量とその信頼度とを有する。図3の場合、出演者名(役名)「竹田花子(ヒロコ)」の代表特徴量の信頼度は、他の二名の代表特徴量の信頼度より高い。 FIG. 3 is a diagram illustrating an example of performer information held in the performer information holding unit 110. The performer names shown in FIG. 3 are representative feature amounts of the performer names (titles) “Taro Matsuyama, Hanako Takeda (Hiroko), Ryoko Umekawa (Yuri)” of the performers in the electronic program guide of FIG. And its reliability. In the case of FIG. 3, the reliability of the representative feature amount of the performer name (title) “Hanako Takeda” is higher than the reliability of the representative feature amounts of the other two people.
 上記出演者情報管理部104、検索部105、表示情報生成部106、特徴量算出部108、代表特徴量決定部109、および特徴量算出部111は、CPU等からなり、番組情報表示処理の実行を含む装置全体の制御を行う。上記CPUは、ROM、RAMおよび電気的に書換可能な不揮発性メモリであるEEPROM(electrically erasable programmable ROM)あるいはフラッシュROM等を備えている。上記メモリには、プログラム、通信制御データ、さらに端末の識別コード等の種々のデータを記憶する。また、上記メモリには、出演者情報保持部110の出演者情報が格納される。 The performer information management unit 104, the search unit 105, the display information generation unit 106, the feature amount calculation unit 108, the representative feature amount determination unit 109, and the feature amount calculation unit 111 include a CPU and the like, and execute program information display processing. Control of the entire device including The CPU includes a ROM, a RAM, and an EEPROM (electrically erasable programmable) ROM, which is an electrically rewritable nonvolatile memory, or a flash ROM. The memory stores various data such as programs, communication control data, and terminal identification codes. The memory stores the performer information of the performer information holding unit 110.
 以下、上述のように構成された番組情報表示装置100の動作を説明する。 Hereinafter, the operation of the program information display apparatus 100 configured as described above will be described.
 図4および図5は、番組情報表示装置100の動作を示すフローチャートである。図4は、代表特徴量を決定するまでの処理を示す。図5は、代表特徴量が決定された後に出演者名を表示するまでの処理を示す。図4に示すフローは、基本的にはユーザが該当番組を視聴中に実行される。図5に示すフローは、基本的にはユーザ指示をトリガとして実行される。このように、実行条件が異なるので、2つのフローに分けて説明する。図4および図5に示すフローは、上述したようにCPUにより番組情報表示プログラムとして実行される。図中、記号「S」はフローの各ステップを示す。 4 and 5 are flowcharts showing the operation of the program information display apparatus 100. FIG. FIG. 4 shows processing until the representative feature amount is determined. FIG. 5 shows processing until the name of the performer is displayed after the representative feature amount is determined. The flow shown in FIG. 4 is basically executed while the user is viewing the corresponding program. The flow shown in FIG. 5 is basically executed with a user instruction as a trigger. As described above, since the execution conditions are different, the description will be divided into two flows. The flow shown in FIGS. 4 and 5 is executed as a program information display program by the CPU as described above. In the figure, the symbol “S” indicates each step of the flow.
 まず、代表特徴量決定部109が代表特徴量を決定するまでの処理について、図4のフローチャートを用いて説明する。 First, processing until the representative feature amount determination unit 109 determines a representative feature amount will be described with reference to the flowchart of FIG.
 ステップS11において、番組取得部101は、表示対象となる番組の番組情報および動画像を取得する。番組情報の取得方法としては、電子番組表(EPG)を取得し、取得した電子番組表から取得する方法が最も容易であるが、動画像中の文字列、音声、またはクローズドキャプションの認識等を行い、その認識結果から取得する方法を採用してもよい。さらに、電子番組表には、インターネット上における出演者の公式サイトのURL、番組の公式サイトのURLが含まれていてもよい。電子番組表の一例は、図2に示す。 In step S11, the program acquisition unit 101 acquires program information and a moving image of a program to be displayed. The easiest way to acquire the program information is to acquire an electronic program guide (EPG) and acquire it from the acquired electronic program guide. However, it is possible to recognize character strings, audio, or closed captions in moving images. The method of performing and acquiring from the recognition result may be employ | adopted. Further, the electronic program guide may include the URL of the official site of the performer on the Internet and the URL of the official site of the program. An example of the electronic program guide is shown in FIG.
 ステップS12において、出演者情報管理部104は、「すべての出演者名」について以下のループの始端と終端との間の処理を繰り返す。出演者情報管理部104は、番組取得部101で取得された番組情報を基に、電子番組表の出演者のフィールドから出演者名を抽出し、得られたすべての出演者名について、以降の処理を行う。 In step S12, the performer information management unit 104 repeats the process between the start and end of the following loop for “all performer names”. The performer information management unit 104 extracts a performer name from the performer field of the electronic program guide based on the program information acquired by the program acquisition unit 101, and for all the performer names obtained thereafter, Process.
 ステップS13において、出演者情報管理部104は、出演者名が、出演者情報保持部110に保持されているか否かを確認する。ただし、処理の開始時において、出演者情報保持部110にはいずれの出演者名も登録されていないものとする。 In step S13, the performer information management unit 104 checks whether or not the performer name is held in the performer information holding unit 110. However, it is assumed that no performer name is registered in the performer information holding unit 110 at the start of the process.
 ステップS14において、出演者情報管理部104は、出演者名が出演者情報保持部110に保持されている場合は、次の出演者に処理を移す。一方、出演者管理情報部104は、出演者名が出演者情報保持部110に保持されていない場合は、ステップS15に進む。 In step S14, when the performer name is held in the performer information holding unit 110, the performer information management unit 104 moves the process to the next performer. On the other hand, if the performer name is not held in the performer information holding unit 110, the performer management information unit 104 proceeds to step S15.
 ステップS15において、関連画像取得部107は、出演者名を基に関連画像を取得する。関連画像取得部107は、インターネットをはじめとするネットワーク200において、出演者名をキーワードとして、関連画像を検索する。そして、関連画像取得部107は、検索結果として、例えば画像検索装置210から、出演者名に関連する関連画像、またはその取得先のURL等の情報を取得し、保持する。 In step S15, the related image acquisition unit 107 acquires a related image based on the performer name. The related image acquisition unit 107 searches for related images in the network 200 including the Internet using the performer name as a keyword. Then, the related image acquisition unit 107 acquires and holds, as a search result, information such as the related image related to the performer name or the URL of the acquisition destination from the image search device 210, for example.
 ステップS16では、特徴量算出部108は、「すべての関連画像」について以下のループの始端と終端との間の処理を繰り返す。 In step S16, the feature amount calculation unit 108 repeats the process between the start and end of the following loop for “all related images”.
 ステップS17において、特徴量算出部108は、関連画像取得部107で各出演者について取得した全ての関連画像から、人物の顔領域を関連抽出画像として切り出し、切り出した関連抽出画像ごとに画像特徴量を求める。そして、特徴量算出部108は、求めた画像特徴量を、基の関連画像に対応する出演者名と対応付けて一時的に保存する。人物の顔領域を切り出す方法としては、色や輝度の情報からエッジを抽出し、抽出したエッジを用いて顔の輪郭および顔パーツのパターン認識処理を行う方法等を用いる。画像特徴量は、色調や顔パーツの配置位置等を、所定の基準により数値化することによって求められる値である。 In step S <b> 17, the feature amount calculation unit 108 cuts out a person's face area as a related extracted image from all the related images acquired for each performer by the related image acquisition unit 107, and sets the image feature amount for each extracted related extracted image. Ask for. Then, the feature amount calculation unit 108 temporarily stores the obtained image feature amount in association with the performer name corresponding to the base related image. As a method of cutting out a person's face area, a method of extracting an edge from color or luminance information and performing a pattern recognition process of a face outline and a face part using the extracted edge is used. The image feature amount is a value obtained by digitizing the color tone, the arrangement position of the face parts, and the like according to a predetermined reference.
 ステップS18において、代表特徴量決定部109は、特徴量算出部108で得られた関連抽出画像の画像特徴量から、出演者名ごとに代表特徴量と信頼度を決定する。 In step S18, the representative feature amount determination unit 109 determines a representative feature amount and reliability for each performer name from the image feature amounts of the related extracted images obtained by the feature amount calculation unit 108.
 そして、ステップS19において、出演者情報管理部104は、出演者名、代表特徴量、および信頼度を、関連付けた状態で、出演者情報保持部110に保持する。出演者情報保持部110に保持される情報の一例は、図3に示される。なお、画像特徴量は、特徴量算出部108のアルゴリズムや精度、保存形式に応じて異なるものであり、この例が画像特徴量の一般的な表現を示すものではない。 And in step S19, the performer information management part 104 hold | maintains the performer name, representative feature-value, and reliability in the performer information holding | maintenance part 110 in the linked | related state. An example of information held in the performer information holding unit 110 is shown in FIG. The image feature amount differs depending on the algorithm, accuracy, and storage format of the feature amount calculation unit 108, and this example does not indicate a general expression of the image feature amount.
 次に、代表特徴量決定部109の処理について説明する。 Next, processing of the representative feature amount determination unit 109 will be described.
 なお、代表特徴量決定部109の処理に先立って、特徴量算出部108は、関連抽出画像から得られた画像特徴量に対するノイズ除去処理を行う。ここで、ノイズとは、同一の出演者名に対応付けられた画像特徴量を構成する特徴値の中で値が大きく異なるもの、つまり、その出演者の顔画像の画像特徴量の特徴値ではない可能性が高いものを指す。または、ノイズとは、このような特徴値が取得される画像、つまり、その出演者の顔画像ではない可能性が高い画像を指す。 Prior to the processing of the representative feature amount determination unit 109, the feature amount calculation unit 108 performs noise removal processing on the image feature amount obtained from the related extracted image. Here, the noise is a characteristic value that greatly differs among the feature values constituting the image feature amount associated with the same performer name, that is, the feature value of the image feature amount of the performer's face image. Points that are not likely. Or noise refers to an image from which such a characteristic value is acquired, that is, an image that is not likely to be a face image of the performer.
 番組情報表示装置100は、出演者名をキーワードとして、ネットワーク200経由で画像検索装置210から関連画像を検索する。番組情報表示装置100は、画像検索装置210の検索アルゴリズムに従って、番組情報サーバ230あるいは画像サーバ220から、関連画像を抽出する。このような画像検索は、キーワードで関係する画像を検索する手法により行われることが一般的である。また、画像サーバ220が管理する画像の中には、キーワードが一致するだけで出演者の顔画像ではない画像も存在する。したがって、検索結果には、出演者名に関連するものの顔画像ではない画像も含まれ得る。 The program information display device 100 searches for related images from the image search device 210 via the network 200 using the performer name as a keyword. The program information display device 100 extracts related images from the program information server 230 or the image server 220 according to the search algorithm of the image search device 210. Such an image search is generally performed by a method of searching for an image related by a keyword. In addition, among the images managed by the image server 220, there are images that are not face images of performers only by matching the keywords. Therefore, the search result may include an image that is not a face image related to the performer name.
 画像特徴量は、特徴量算出部108のアルゴリズムに従って得られる。画像特徴量は、例えば、顔パーツの配置位置を、複数の特徴値(ここでは特徴値を0以上とする)の組み合わせで表現する。特徴量算出部108は、出演者ごとに、その出演者に関連する複数の関連画像の特徴値の統計的な分布を求め、その分布から、代表特徴量を求める。具体的には、特徴量算出部108は、複数の特徴値全体の平均から予め規定した閾値以内にある特徴値を特定し、特定した特徴値の平均を再計算し、再計算結果を代表特徴量に採用する。これにより、特徴量算出部108は、値が平均から大きく異なる特徴値(ノイズ)を排除する。特徴量算出部108による特徴量算出の具体例については、図7を用いて後述する。 The image feature amount is obtained according to the algorithm of the feature amount calculation unit 108. The image feature amount represents, for example, the arrangement position of the face part by a combination of a plurality of feature values (here, the feature value is 0 or more). For each performer, the feature amount calculation unit 108 obtains a statistical distribution of feature values of a plurality of related images related to the performer, and obtains a representative feature amount from the distribution. Specifically, the feature amount calculation unit 108 identifies feature values that are within a predetermined threshold from the average of all of the plurality of feature values, recalculates the average of the specified feature values, and displays the recalculation result as a representative feature. Adopt to quantity. Thereby, the feature amount calculation unit 108 excludes feature values (noise) whose values are greatly different from the average. A specific example of the feature amount calculation by the feature amount calculation unit 108 will be described later with reference to FIG.
 このようなノイズ除去処理により、特徴量算出部108は、目標となる人物の顔画面以外の画像から得られた画像特徴量は、全てノイズとして排除する。いわゆる検索エンジンを用いて画像検索を行うと、関係ない画像が検索されたり、該当人物以外の人が写っている画像が検索される場合も多い。検索数が多くなり特徴量算出処理の数が多くなると精度が高くなると考えられるので、例えば複数の検索エンジンを切り替えて検索を何度か繰り返すことにより精度を上げることが可能である。例えば、番組が始まってから番組を受信している間、バックグランド処理により、常時、関連画像の検索処理を行ってもよい。番組予約がある場合には、番組開始に先立って予め、図4のフローを実行することも可能である。このようにすれば、より多くの関連画像を検索することが可能となり、代表特徴量の信頼度を高めることができる。 By such noise removal processing, the feature amount calculation unit 108 excludes all image feature amounts obtained from images other than the target person's face screen as noise. When an image search is performed using a so-called search engine, an irrelevant image is searched, or an image in which a person other than the person is shown is often searched. Since it is considered that the accuracy increases as the number of searches increases and the number of feature amount calculation processes increases, it is possible to increase the accuracy by switching a plurality of search engines and repeating the search several times. For example, while the program is being received after the program has started, the related image search process may be constantly performed by background processing. If there is a program reservation, it is possible to execute the flow of FIG. 4 in advance prior to the start of the program. In this way, more related images can be searched, and the reliability of the representative feature amount can be increased.
 ここで、出演者情報管理部104は、番組取得部101が保持する番組公式サイトのURLと、関連画像取得部107が取得した関連画像のURLとを比較し、同一のドメインである場合には、取得した関連画像が公式サイトのものであると判断する。そして、代表特徴量決定部109は、その関連画像の特徴量に重み付けすることで、当該関連画像の信頼度の高さを、算出される代表特徴量に反映させる。信頼度の重み付けの計算は、例えば次のように行う。 Here, the performer information management unit 104 compares the URL of the program official site held by the program acquisition unit 101 with the URL of the related image acquired by the related image acquisition unit 107, and if they are the same domain, , It is determined that the acquired related image is of the official website. Then, the representative feature amount determination unit 109 weights the feature amount of the related image to reflect the high reliability of the related image in the calculated representative feature amount. For example, the reliability weighting is calculated as follows.
 ここで、特徴量算出部108により算出された出演者ごとの画像特徴量F(x)は、m個の特徴値からなる次式(1)で表現されるものとする。ここで、変数xは、出演者を表す。このとき、代表特徴量Fは、ノイズ排除後の関連画像の個数(画像特徴量の個数)をn、公式サイトか否かを示す係数をsとすると、次式(2)により得られる。 Here, it is assumed that the image feature amount F (x) for each performer calculated by the feature amount calculation unit 108 is expressed by the following equation (1) including m feature values. Here, the variable x represents a performer. At this time, the representative feature amount F t is obtained by the following equation (2), where n is the number of related images after noise elimination (number of image feature amounts) and s i is a coefficient indicating whether or not the site is an official site. .
Figure JPOXMLDOC01-appb-M000001
Figure JPOXMLDOC01-appb-M000001
Figure JPOXMLDOC01-appb-M000002
Figure JPOXMLDOC01-appb-M000002
 また、信頼度ベクトルCを次式(3)のように定義し、信頼度Cを次式(4)のように定義する。 Further, the reliability vector C v is defined as following equation (3), defined as the following equation (4) the reliability C.
Figure JPOXMLDOC01-appb-M000003
Figure JPOXMLDOC01-appb-M000003
Figure JPOXMLDOC01-appb-M000004
Figure JPOXMLDOC01-appb-M000004
 すなわち、信頼度ベクトルCは、基の特徴値が、複数の関連画像の間でどれだけまとまっているか(ばらついていないか)を示すベクトル値であり、信頼度Cは、このベクトル値のスカラー量を特徴値の個数で正規化した値である。代表特徴量Fの算出の対象となる複数の関連抽出画像に占める正しい画像(目的とする出演者の顔画像)の割合が高くなればなるほど、特徴値のばらつきは小さくなり、信頼度Cの値は大きくなる。したがって、信頼度Cは、代表特徴量Fが出演者の実際の顔画像の画像特徴量に一致していることについての信頼度の高さを示す。 That is, the reliability vector C v is the characteristic value of the group, a vector value that indicates how much held together (or not vary) among a plurality of related images, reliability C is a scalar of the vector value This is a value obtained by normalizing the quantity by the number of feature values. As the ratio of the correct image (the target performer's face image) in the plurality of related extracted images for which the representative feature amount F t is calculated becomes higher, the variation in the feature value becomes smaller and the reliability C becomes higher. The value gets bigger. Therefore, the reliability C indicates the high reliability that the representative feature amount F t matches the image feature amount of the actual face image of the performer.
 上記公式サイトか否かの係数sは、例えば公式サイトの画像である場合には2、その他サイトからの画像の場合には1というように設定する。 The coefficient s i for determining whether or not the site is an official site is set to 2 for an image of an official site, for example, and 1 for an image from another site.
 次に、特徴量算出部108が行う特徴量算出の具体例について説明する。 Next, a specific example of feature amount calculation performed by the feature amount calculation unit 108 will be described.
 画像特徴量は、上記式(1)に示すように、m個の特徴値の組み合わせである。個々の画像特徴量は、色情報や、顔の各パーツの位置情報から求めることができる。例えば、特徴量算出部108は、色情報を基にした色相等を算出し、その算出結果を特徴値とする。または、特徴量算出部108は、顔の各パーツの縦横比または相対位置等を算出し、算出結果を特徴値とする。顔のパーツの位置情報は、例えば、色情報から得られる輝度や色相等から、画像の中の輪郭を抽出することによって、得ることができる The image feature amount is a combination of m feature values as shown in the above equation (1). Individual image feature amounts can be obtained from color information and position information of each part of the face. For example, the feature amount calculation unit 108 calculates a hue or the like based on the color information, and uses the calculation result as a feature value. Alternatively, the feature amount calculation unit 108 calculates the aspect ratio or relative position of each part of the face, and uses the calculation result as a feature value. The position information of the facial parts can be obtained, for example, by extracting the contour in the image from the brightness, hue, etc. obtained from the color information.
 以下、代表特徴量の算出方法の具体例について説明する。簡単のため、m=3とする。 Hereinafter, a specific example of a representative feature amount calculation method will be described. For simplicity, m = 3.
 図7は、特徴量算出部108が算出した特徴量の例を示す。図7の場合、画像特徴量は、特徴量番号毎に、3個の特徴値1~3を有する。この特徴量を用いて、代表特徴量の算出の例を説明する。ここで、特徴量番号とは、同一の出演者に対応して取得された関連画像それぞれの特徴量の識別番号である。 FIG. 7 shows an example of the feature amount calculated by the feature amount calculation unit 108. In the case of FIG. 7, the image feature amount has three feature values 1 to 3 for each feature amount number. An example of calculating a representative feature amount will be described using this feature amount. Here, the feature quantity number is an identification number of each feature quantity of the related images acquired corresponding to the same performer.
 特徴量算出部108は、ノイズを除去するためと信頼度を算出するために、相加平均および分散を計算する。分散は、特徴値のばらつき度合いを意味するものである。特徴量算出部108は、分散を基に上述の閾値を算出し、算出した閾値を超える特徴値を代表特徴量算出の対象外とすることにより、ノイズを除去する。なお、分散の値を信頼度とする態様でもよい。 The feature amount calculation unit 108 calculates an arithmetic mean and variance in order to remove noise and to calculate reliability. Variance means the degree of variation of feature values. The feature amount calculation unit 108 calculates the above-described threshold value based on the variance, and removes noise by excluding the feature value exceeding the calculated threshold value from the representative feature amount calculation target. In addition, the aspect which makes the value of dispersion | distribution a reliability may be sufficient.
 図7では、n=5の場合を例としている。この場合、特徴量算出部108は、5個の関連画像を用いて、相加平均と分散を求めることになる。関連画像は、上述の通り、あるキーワード(出演者名)を基にして検索された画像である。したがって、ひとつのキーワード(出演者名)について、複数の関連画像が検索される。特徴量算出部108は、この複数の関連画像を用いて、キーワード(出演者名)に対する代表特徴量を求める。 FIG. 7 shows an example where n = 5. In this case, the feature amount calculation unit 108 obtains the arithmetic mean and variance using the five related images. As described above, the related image is an image searched based on a certain keyword (performer name). Therefore, a plurality of related images are searched for one keyword (performer name). The feature amount calculation unit 108 obtains a representative feature amount for the keyword (performer name) using the plurality of related images.
 ここで、キーワード(出演者名)によって検索される関連画像には、まったく関係のない画像(ノイズ)が含まれる。場合によっては、検索された関連画像は、目的とする人物の画像でないこともある。ノイズを含めたまま得られた平均値は、妥当な値であるとは考えにくい。したがって、特徴量算出部108は、ノイズを除去した後で、平均値の再計算を行う。 Here, the related image searched by the keyword (performer name) includes an unrelated image (noise). In some cases, the retrieved related image may not be an image of a target person. The average value obtained with noise included is unlikely to be a reasonable value. Therefore, the feature amount calculation unit 108 recalculates the average value after removing the noise.
 複数の関連画像から代表特徴量を求めるために、特徴量算出部108は、以下(a)-(e)の処理を行う。(a)ノイズを含む母集団から仮の平均値を求める。(b)仮の平均値から分散を求める。(c)分散を基にノイズを除去する。(d)ノイズを除いた母集団の平均を求める。これが再計算である。(e)求められた平均を代表特徴量とする。 In order to obtain a representative feature amount from a plurality of related images, the feature amount calculation unit 108 performs the following processes (a) to (e). (A) A temporary average value is obtained from a population including noise. (B) The variance is obtained from the provisional average value. (C) Remove noise based on dispersion. (D) The average of the population excluding noise is obtained. This is recalculation. (E) The average obtained is used as the representative feature amount.
 まず、特徴量算出部108は、図7における特徴値1~3のそれぞれについて、特徴量番号1~5に対応する5つの特徴値の相加平均を求める。その際に、特徴量算出部108は、公式サイトの特徴量に、他の特徴量よりも所定倍(例えば5倍)の重み付けをする。図7の場合、特徴量番号1の特徴量1~3は、公式サイトから取得した特徴量であるため、それぞれ、特徴量番号2~5の特徴量1~3よりも5倍の重み付けをする。この結果、母集団の数が4増えたのと同じことになる。 First, the feature quantity calculation unit 108 obtains an arithmetic average of five feature values corresponding to the feature quantity numbers 1 to 5 for each of the feature values 1 to 3 in FIG. At that time, the feature amount calculation unit 108 weights the feature amount of the official site by a predetermined multiple (for example, five times) over the other feature amounts. In the case of FIG. 7, the feature amounts 1 to 3 with the feature amount number 1 are the feature amounts acquired from the official site, and thus are weighted five times as much as the feature amounts 1 to 3 with the feature amount numbers 2 to 5, respectively. . As a result, the number of populations has increased by four.
 図7に示す特徴値1~3のうち、特徴値1の場合を例に採り、代表特徴量の算出手法について説明する。 The feature value calculation method will be described by taking the case of the feature value 1 among the feature values 1 to 3 shown in FIG. 7 as an example.
 特徴値1の相加平均f1mは、次式(5)として示される。分散σは、次式(6)として示される。 The arithmetic average f 1m of the feature value 1 is expressed as the following equation (5). The variance σ 2 is expressed as the following equation (6).
 f1m = ((25)+3+2+8+1)/(5+4)≒2.67 …(5)
 σ2  = (25+33+22+88+11)/(5+4)-(f1m
     ≒ 10.89-7.13=3.76 …(6)
f 1m = ((2 * 5) + 3 + 2 + 8 + 1) / (5 + 4) ≈2.67 (5)
σ 2 = (2 * 2 * 5 + 3 * 3 + 2 * 2 + 8 * 8 + 1 * 1) / (5 + 4) − (f 1m ) 2
   ≒ 10.89-7.13 = 3.76 (6)
 特徴値1が正規分布になると仮定し、かつ閾値thを±標準偏差σと予め定義した場合、閾値thは、特徴値1の相加平均f1mにおいて、次式(7)で示される。 When it is assumed that the feature value 1 has a normal distribution and the threshold value th is previously defined as ± standard deviation σ, the threshold value th is expressed by the following equation (7) in the arithmetic mean f 1m of the feature value 1.
 th ≒ ±√(3.76) = ±1.94 …(7) Th ≒ ± √ (3.76) = ± 1.94 (7)
 すなわち、特徴値1が本来取り得る値f1eとしては、次式(8)および(9)となる。 That is, the values f 1e that can be originally taken by the feature value 1 are expressed by the following equations (8) and (9).
 (f1m-th) ≦ f1e ≦ (f1m+th) …(8)
 0.73 ≦ f1e ≦ 4.61 …(9)
(F 1m −th) ≦ f 1e ≦ (f 1m + th) (8)
0.73 ≦ f 1e ≦ 4.61 (9)
 上記式(9)から、図7に示す特徴量番号4の特徴値1は、特徴値1が本来取り得る値f1eに該当しないので、ノイズとみなすことができる。例えば、特徴量算出部108は、f1mについて、特徴量番号4のデータを除き、重み付けを考慮して、次式(10)により上記再計算を行う。ただし、他の特徴値の除外によるノイズ除去は無視している。 From the above equation (9), the feature value 1 of the feature number 4 shown in FIG. 7 does not correspond to the value f 1e that the feature value 1 can originally take, and can be regarded as noise. For example, the feature amount calculation unit 108 recalculates f 1m according to the following equation (10), excluding the data of feature amount number 4 and considering weighting. However, noise removal by excluding other feature values is ignored.
 f1m=((25)+3+2+1)/(4+4)=2.00 …(10) f 1m = ((2 * 5) + 3 + 2 + 1) / (4 + 4) = 2.00 (10)
 以上により、図7において、代表特徴量の特徴値1の値を「2」とすることができる。同様にして、残りの特徴値である特徴値2、3についても、ノイズとなる特徴量を除いて新たな平均を求め、代表特徴量を算出することができる。 As described above, the value of the feature value 1 of the representative feature amount can be set to “2” in FIG. Similarly, for the feature values 2 and 3 which are the remaining feature values, a new average can be obtained by excluding the feature amount that becomes noise, and the representative feature amount can be calculated.
 代表特徴量決定部109は、同一のキーワードに対し、代表特徴量の算出に使用された画像特徴量のばらつきが大きい場合は、信頼度が低くなり、ばらつきが小さければ信頼度は高くなると判断する。 The representative feature amount determination unit 109 determines that the reliability is low when the variation of the image feature amount used to calculate the representative feature amount is large for the same keyword, and the reliability is high when the variation is small. .
 代表特徴量決定部109は、計算を簡略するために、別の手法により信頼度Cを求めてもよい。例えば、代表特徴量決定部109は、ノイズを除いた特徴量の数n、ノイズ排除前の画像数をa、公式サイトの画像数をNとした場合、信頼度Cを、次式(11)により求める。この場合も同様に、特徴量のばらつきが大きい場合には信頼度は低くなり、特徴量のばらつきが小さければ信頼度は高くなる。 The representative feature quantity determination unit 109 may obtain the reliability C by another method in order to simplify the calculation. For example, the representative feature amount determination unit 109 sets the reliability C as the following equation (11), where n is the number of feature amounts excluding noise, a is the number of images before noise removal, and N is the number of images on the official site. Ask for. Similarly, in this case, the reliability is low when the variation in the feature amount is large, and the reliability is high when the variation in the feature amount is small.
Figure JPOXMLDOC01-appb-M000005
Figure JPOXMLDOC01-appb-M000005
 代表特徴量決定部109は、このようにして得られた出演者名、代表特徴量F、および信頼度Cを、関連付けた状態で出演者情報保持部110に保持する。なお、代表特徴量決定部109は、番組情報として、出演者名に加えて役名が得られた場合に、役名も併せて出演者情報保持部110に保持してもよい。ネットワーク経由で画像検索を行う場合、多くの関連画像が検索されることもある。したがって、代表特徴量決定部109は、代表特徴量Fを決定するまでの処理を、番組表示中に、段階的に非同期で行ってもよい。すなわち、代表特徴量Fの算出に用いる関連画像の数を段階的に増やしながら、段階ごとに代表特徴量Fを算出し、より後に算出された代表特徴量Fを、後述する照合処理に用いる。これにより、番組視聴開始の早い段階から照合処理を開始することができ、更に、時間の経過とともに代表特徴量Fの信頼度を高めることができる。 The representative feature quantity determination unit 109 holds the performer name, the representative feature quantity F t , and the reliability C thus obtained in the performer information holding unit 110 in an associated state. The representative feature amount determination unit 109 may store the role name in the performer information holding unit 110 when the role name is obtained in addition to the performer name as the program information. When an image search is performed via a network, many related images may be searched. Therefore, the representative feature quantity determining unit 109 may perform the processing until the representative feature quantity F t is determined asynchronously step by step during program display. That is, while increasing the number of related images used for calculating the representative feature quantity F t in stages, the representative feature quantity F t is calculated for each stage, and the later calculated representative feature quantity F t is used as a collation process described later. Used for. Thereby, the collation process can be started from the early stage of the program viewing start, and the reliability of the representative feature quantity F t can be increased with the passage of time.
 また、代表特徴量決定部109は、出演者名について数個の関連画像を処理した時点で次の出演者名の処理を行い、すべての出演者名について処理した後、再度、各出演者名について、別の数個の関連画像を処理する。このように段階的に処理を行うようにすることによって、番組視聴開始の早い段階から照合処理を開始することでき、更に、時間の経過とともに代表特徴量Fの信頼度を高めることが可能となる。 The representative feature amount determination unit 109 processes the next performer name when several related images are processed for the performer name, processes all performer names, and then again performs each performer name. Is processed with several other related images. By performing the processing step by step in this way, the collation processing can be started from the early stage of the program viewing start, and the reliability of the representative feature amount F t can be increased with the passage of time. Become.
 次に、出演者名の表示までの処理について、図5のフローチャートを用いて説明する。ユーザは、番組視聴中に出演者の人物を知りたい場合に、所定のユーザ指示操作を、番組情報表示装置に対して行う。図5のフローは、このユーザ指示をトリガとして実行される。 Next, processing up to the display of the performer name will be described using the flowchart of FIG. When the user wants to know the performer's person while viewing the program, the user performs a predetermined user instruction operation on the program information display device. The flow in FIG. 5 is executed with this user instruction as a trigger.
 ステップS20において、表示画像取得部102は、番組取得部101から取得した動画像から、表示画像をフレーム単位で取得し、取得した表示画像を保持する。 In step S20, the display image acquisition unit 102 acquires a display image in units of frames from the moving image acquired from the program acquisition unit 101, and holds the acquired display image.
 ステップS21において、表示部103は、取得した表示画像を、そのまま動画像として表示する。 In step S21, the display unit 103 displays the acquired display image as a moving image as it is.
 ステップS22において、検索部105は、表示画像取得部102が取得したフレーム単位の表示画像から、人物の顔領域を切り出し、切り出した領域を表示抽出画像として出力すると共に、表示抽出画像の切り出し位置(以下単に[位置]という)を出力する。特徴量算出部111は、表示抽出画像から抽出画像特徴量を算出する。表示画像に複数の人物が写っている場合には、複数の表示抽出画像が得られる。特徴量算出部111は、既に詳細に説明した特徴量算出部108と同一機能を有する。ただし、特徴量算出部108が関連抽出画像の特徴量を計算するのに対し、特徴量算出部111は、表示抽出画像の特徴量を計算する。 In step S <b> 22, the search unit 105 cuts out the face area of the person from the frame-by-frame display image acquired by the display image acquisition unit 102, outputs the cut out area as a display extracted image, and extracts the display extracted image ( (Hereinafter simply referred to as [position]). The feature amount calculation unit 111 calculates an extracted image feature amount from the display extraction image. When a plurality of persons are shown in the display image, a plurality of display extraction images are obtained. The feature amount calculation unit 111 has the same function as the feature amount calculation unit 108 that has already been described in detail. However, while the feature amount calculation unit 108 calculates the feature amount of the related extracted image, the feature amount calculation unit 111 calculates the feature amount of the display extracted image.
 ステップS23では、検索部105は、「すべての関連画像」について以下のループの始端と終端との間の処理を繰り返す。 In step S23, the search unit 105 repeats the process between the start and end of the following loop for “all related images”.
 ステップS24において、検索部105は、出演者情報保持部110に照合済みの代表特徴量がなくなるまで、出演者情報保持部110から、出演者情報管理部104を介して代表特徴量を取得する。 In step S24, the search unit 105 acquires the representative feature amount from the performer information holding unit 110 via the performer information management unit 104 until there is no more representative feature amount that has been verified in the performer information holding unit 110.
 ステップS25において、検索部105は、抽出画像特徴量と代表特徴量の照合を行い、類似度を算出する。検索部105は、代表特徴量をF、抽出画像特徴量をF、特徴量の取りうる最大値をFmax(但し各特徴値の最大値は0より大きい)とすると、類似度Sは、例えば次式(12)により求めることができる。 In step S25, the search unit 105 collates the extracted image feature quantity with the representative feature quantity, and calculates the similarity. If the representative feature amount is F t , the extracted image feature amount is F e , and the maximum value that can be taken by the feature amount is F max (however, the maximum value of each feature value is greater than 0), the similarity S is For example, it can obtain | require by following Formula (12).
Figure JPOXMLDOC01-appb-M000006
Figure JPOXMLDOC01-appb-M000006
 ここで、検索部105は、代表特徴量Fと該当の信頼度Cを取得し、予め設定した閾値よりも信頼度が低い代表特徴量の照合は行わないようにしてもよい。この場合、処理を高速化することができる。 Here, the search unit 105 may acquire the representative feature amount Ft and the corresponding reliability C, and may not collate the representative feature amount having a reliability lower than a preset threshold value. In this case, the processing can be speeded up.
 ステップS26において、検索部105は、算出した類似度のうち最大となる類似度を抽出し、抽出した類似度に該当する代表特徴量を特定する。 In step S26, the search unit 105 extracts the maximum similarity among the calculated similarities, and specifies a representative feature amount corresponding to the extracted similarity.
 ステップS27において、検索部105は、該当する代表特徴量に関連付けられている出演者名を、検索結果として保持する。検索部105は、特徴量算出部111により算出された抽出画像特徴量と、図4のフローにおいて代表特徴量決定部109により得られた関連画像に関する代表特徴量とをマッチングする。具体的には、検索部は、ステップS26で特定した代表特徴量に対応付けられた出演者名を、出演者情報管理部104を介して、出演者情報保持部110で検索し、抽出する。 In step S27, the search unit 105 holds the performer name associated with the corresponding representative feature amount as a search result. The search unit 105 matches the extracted image feature amount calculated by the feature amount calculation unit 111 with the representative feature amount related to the related image obtained by the representative feature amount determination unit 109 in the flow of FIG. Specifically, the search unit searches the performer information holding unit 110 via the performer information management unit 104 and extracts the performer name associated with the representative feature amount specified in step S26.
 ステップS28において、表示情報生成部106は、検索部105によって抽出された出演者名と表示抽出画像とから、出演者名の検索精度を出演者名と同時に表示する表示情報を生成する。具体的には、表示情報生成部106は、出演者名から決定される表示内容と、表示抽出画像の位置から決定される表示位置と、代表特徴量の信頼度または類似度の少なくとも一方の情報と、抽出画像特徴量と代表特徴量とのマッチング結果とから、出演者名の検索精度を算出する。そして、表示情報生成部106は、この検索精度に基づいて、表示情報を生成する。このとき、表示情報生成部106は、基本的には、検索精度が高い表示内容をより目立つように表示する表示情報を生成する。 In step S28, the display information generation unit 106 generates display information for displaying the search accuracy of the performer name simultaneously with the performer name from the performer name extracted by the search unit 105 and the display extracted image. Specifically, the display information generation unit 106 displays at least one of the display content determined from the performer name, the display position determined from the position of the display extracted image, and the reliability or similarity of the representative feature amount. From the matching result between the extracted image feature quantity and the representative feature quantity, the search accuracy of the performer name is calculated. Then, the display information generation unit 106 generates display information based on this search accuracy. At this time, the display information generation unit 106 basically generates display information for displaying the display contents with high search accuracy more prominently.
 ステップS29において、表示部103は、表示情報生成部106によって生成された表示情報を表示することにより、検索結果を表示する。 In step S29, the display unit 103 displays the search result by displaying the display information generated by the display information generation unit 106.
 番組表示装置100は、「すべての関連画像」について以上の出演者名の表示までの処理を実行すると、一連の処理を終了する。 The program display device 100 ends the series of processes when the processes up to the display of the names of the performers described above are performed for “all related images”.
 このように、番組情報表示システムを用いることにより、ユーザは、表示画面に表示されている人物の情報を、知りたいと思ったときに知ることができる。 As described above, by using the program information display system, the user can know the information of the person displayed on the display screen when he / she wants to know.
 次に、表示情報生成部106による表示情報の生成処理の具体例について説明する。 Next, a specific example of display information generation processing by the display information generation unit 106 will be described.
 表示情報生成部106は、検索部105で抽出された出演者名から、表示内容を決定する。表示内容は、例えば、出演者名、役名、および役に関する説明である。 The display information generation unit 106 determines the display content from the performer name extracted by the search unit 105. The display content is, for example, a description of a performer name, a role name, and a role.
 そして、表示情報生成部106は、決定した表示内容と、検索部105によって取得された表示抽出画像の位置とから、表示内容の表示位置を決定する。例えば、表示情報生成部106は、表示抽出画像の領域の上、右、下、左の順に表示内容が配置可能かを確認し、配置可能な位置を、表示内容の表示位置に決定する。ここで配置可能とは、表示内容が表示画面の範囲を超えないこと、既に表示位置を決定済みの他の表示内容と重ならず、既に表示位置を決定済みの他の表示内容と近接していないこと等の条件を満たすことを意味する。 Then, the display information generation unit 106 determines the display position of the display content from the determined display content and the position of the display extracted image acquired by the search unit 105. For example, the display information generation unit 106 confirms whether the display content can be arranged in the order of upper, right, lower, and left in the display extracted image area, and determines the position where the display content can be arranged as the display position of the display content. “Placeable” means that the display content does not exceed the range of the display screen, does not overlap with other display content whose display position has already been determined, and is close to other display content whose display position has already been determined. It means that the conditions such as not being met.
 表示情報生成部106は、検索された表示内容の精度に応じて、表示形態を決定する。表示形態は、表示位置、表示内容、フォントタイプ、フォントサイズ、文字色、背景色、枠線の有無、あるいは枠線の色等、表示内容を文字列等によって表示する際の形態である。表示内容の精度は、検索部105が照合の際に用いた代表特徴量の信頼度と、抽出画像特徴量と代表特徴量の照合の際に算出した類似度との、少なくとも一方に対応して定まる指標である。表示形態の決定に用いられる精度Aは、信頼度Cと類似度Sの両方を用いて計算する場合には、次式(13)に示すように、双方の値を乗算した値としてもよい。あるいは、係数α、βを用いて、次式(14)に示すように、信頼度Cと類似度Sのそれぞれに重み付けをした値を和算した値としてもよい。 The display information generation unit 106 determines the display form according to the accuracy of the searched display content. The display form is a form in which display contents such as a display position, display contents, font type, font size, character color, background color, presence / absence of a border line, or border line color are displayed as a character string or the like. The accuracy of the display content corresponds to at least one of the reliability of the representative feature quantity used by the search unit 105 when collating and the similarity calculated when the extracted image feature quantity and the representative feature quantity are collated. It is a fixed index. When calculating using both the reliability C and the similarity S, the accuracy A used for determining the display form may be a value obtained by multiplying both values as shown in the following equation (13). Alternatively, the coefficients α and β may be used as a value obtained by summing the weighted values of the reliability C and the similarity S as shown in the following equation (14).
Figure JPOXMLDOC01-appb-M000007
Figure JPOXMLDOC01-appb-M000007
Figure JPOXMLDOC01-appb-M000008
Figure JPOXMLDOC01-appb-M000008
 例えば、上記式(14)を用いて、精度Aを出演者名のフォントサイズFsに反映させることができる。この場合、演者名のフォントサイズFsは、基準となる最大のフォントサイズをFmとすると、次式(15)により求めることができる。 For example, the accuracy A can be reflected in the font size Fs of the performer name using the above formula (14). In this case, the font size Fs of the performer name can be obtained by the following equation (15), where Fm is the maximum standard font size.
Figure JPOXMLDOC01-appb-M000009
Figure JPOXMLDOC01-appb-M000009
 上記式(15)を用いて精度Aを求める場合、係数α、βの値を変えることにより、信頼度および類似度の、表示情報への反映の度合いを変えることができる。 When obtaining the accuracy A using the above equation (15), the degree of reflection of the reliability and similarity on the display information can be changed by changing the values of the coefficients α and β.
 図6は、上記式(15)において、表示形態のうちフォントサイズのみに着目した場合の表示情報の生成例を表にして示す図である。 FIG. 6 is a table showing an example of display information generation in the above formula (15) when attention is paid only to the font size in the display form.
 図6に示すように、精度Aの算出基準が「信頼度に基づく場合」は、係数α=1かつ係数β=0となり、フォントサイズFs=CFmとなる。また、算出基準が「類似度に基づく場合」は、係数α=0かつ係数β=1となり、フォントサイズFs=SFmとなる。 As shown in FIG. 6, when the calculation standard of the accuracy A is “based on reliability”, the coefficient α = 1 and the coefficient β = 0, and the font size Fs = C * Fm. When the calculation criterion is “based on similarity”, the coefficient α = 0 and the coefficient β = 1, and the font size Fs = S * Fm.
 また、上記式(13)を用いる場合は、信頼度Cまたは類似度Sのいずれか一方を固定値にすることにより、他方の値のみを反映させて、同様にフォントサイズFsを算出することができる。 In addition, when the above equation (13) is used, it is possible to calculate the font size Fs in a similar manner by reflecting only the other value by setting one of the reliability C and the similarity S to a fixed value. it can.
 以上、表示情報生成部106が、代表特徴量の信頼度Cおよび類似度Sの少なくとも一方を用いて表示情報を生成する表示情報生成例について説明した。 In the above, the display information generation example in which the display information generation unit 106 generates display information using at least one of the reliability C and the similarity S of the representative feature amount has been described.
 表示情報生成部106は、検索部105で取得した出演者名と表示抽出画像とに基づいて、信頼度または類似度を計算する。よって、どの領域が表示抽出画像として切り出されたかは、信頼度や類似度を算出する際に影響することになる。表示抽出画像の領域が不適切な場合(例えば、顔の一部分しか領域に設定されなかった場合)、信頼度や類似度は正しく計算されないことがある。 The display information generation unit 106 calculates the reliability or the similarity based on the performer name acquired by the search unit 105 and the display extracted image. Therefore, which region is cut out as a display extraction image has an influence when calculating reliability and similarity. When the region of the display extraction image is inappropriate (for example, when only a part of the face is set as the region), the reliability and similarity may not be calculated correctly.
 図8は、表示部103がテレビの表示部である場合の、表示領域および表示画像を説明する図である。図8において、表示領域140は、テレビ画面(表示画面)であり、表示画像141は、番組の映像である。表示画像141は、表示領域140に表示される。表示画像141には、人物と出演者名142とが表示される。表示画像141における人物の顔を囲む領域が、表示抽出画像143として切り出される領域である。図8は、表示領域(テレビ画面)140いっぱいに表示画像(番組の映像)141を表示しない場合である。この場合は、表示画面は、表示画像141の周辺に、出演者名を表示する態様をとることも可能である。表示領域140いっぱいに表示画像141を表示しない場合の例としては、Picture-in-Pictureの機能や地上デジタル放送でのデータ放送表示機能を用いた画面等、放送映像を縮小表示する場合が挙げられる。 FIG. 8 is a diagram for explaining a display area and a display image when the display unit 103 is a television display unit. In FIG. 8, a display area 140 is a television screen (display screen), and a display image 141 is a video of a program. The display image 141 is displayed in the display area 140. In the display image 141, a person and a performer name 142 are displayed. A region surrounding the human face in the display image 141 is a region cut out as the display extraction image 143. FIG. 8 shows a case where the display image (program video) 141 is not displayed in the full display area (television screen) 140. In this case, the display screen can take a form in which the name of the performer is displayed around the display image 141. An example of the case where the display image 141 is not displayed in the full display area 140 is a case where a broadcast video is reduced and displayed, such as a screen using a picture-in-picture function or a data broadcast display function in digital terrestrial broadcasting. .
 出演者名の表示形態の変化には、例えば透過度の変化がある。 The change in the display form of the performer name includes, for example, a change in transparency.
 表示形態は、例えば、信頼度および類似度の精度が低い場合、すなわち検索結果が正しくない可能性が高いと判断される場合には、出演者名の文字の透過度を上げて、出演者名を見えにくくする。一方、精度が高い場合は、すなわち検索結果が正しい可能性が高いと判断される場合には、出演者名の文字の透過度を下げて、出演者名をはっきり見えるようにする。 For example, when the accuracy of reliability and similarity is low, that is, when it is determined that there is a high possibility that the search result is not correct, the character name of the performer is increased and the performer name is increased. Make it difficult to see. On the other hand, when the accuracy is high, that is, when it is determined that there is a high possibility that the search result is correct, the character name transparency is lowered to make the performer name clearly visible.
 他の表示形態の変化として、精度が高い場合には文字のサイズを大きくすること、精度の違いに応じて文字の色を変化させることがある。後者の場合には、文字の色とその色が意味する精度とを対応付けたテーブルを、出演者名と同時表示することが望ましい。また、さらに他の表示形態の変化として、出演者名を表示する文字の背景色に対するコントラストを、精度に応じて変化させることがある。 Other changes in the display mode include increasing the character size when the accuracy is high and changing the color of the character according to the difference in accuracy. In the latter case, it is desirable to display a table that associates the color of the character with the accuracy that the color means together with the name of the performer. Further, as another change in the display form, the contrast of the character displaying the performer name with respect to the background color may be changed according to the accuracy.
 また、精度の高い出演者名から順に表示位置を決めることで、精度の高い出演者名に対して表示の優先度を高くすることができる。また、精度が低い場合は、出演者名の表示位置と表示対象となる人物の画像との位置を離し、人物画像と出演者名とを結ぶ引き出し線を表示してもよい。この場合、精度が低くなると、引き出し線が長くなるので、引き出し線の長さにより、精度の違いを示すことができる。さらに、検索結果が複数ある場合には、その精度の違いを、共通の表示形態により対比させることもできる。つまり、動画像の表示の進行により精度が変化する場合は、複数の表示内容がひとつの表示内容に絞り込まれるようにすることもできる。 Also, by determining the display position in order from the name of the performer with high accuracy, the display priority can be increased with respect to the name of the performer with high accuracy. If the accuracy is low, the display position of the performer name and the position of the image of the person to be displayed may be separated, and a lead line connecting the person image and the performer name may be displayed. In this case, if the accuracy is low, the lead line becomes long. Therefore, the difference in accuracy can be shown by the length of the lead line. Furthermore, when there are a plurality of search results, the difference in accuracy can be compared by a common display form. That is, when the accuracy changes due to the progress of the display of moving images, a plurality of display contents can be narrowed down to one display content.
 表示形態として、抽出した人物の顔領域を示す枠を表示に加えるとさらに視認性が向上する。また、表示形態は、領域の枠と表示内容の表示色を合わせる等すると、さらに視認性が向上する。また、表示形態として、出演者情報保持部110に役名が登録されていれば、表示内容として役名を併記してもよい。 As a display form, if a frame indicating the extracted human face area is added to the display, the visibility is further improved. Further, the visibility is further improved by matching the frame of the area with the display color of the display content. Further, as a display form, if a role name is registered in the performer information holding unit 110, the role name may be written as the display content.
 図9は、表示部103により表示される検索結果の表示の例を示す図である。 FIG. 9 is a diagram showing an example of the display of the search result displayed by the display unit 103.
 図9において、テレビ画面である表示画面150に二人の人物が映っている。左側の人物の顔領域153には、出演者名151が併記される。同様に、右側の人物の顔領域154には、出演者名152が併記される。ここでは、図3の出演者名を、信頼度が高いほど大きな文字サイズで表示し、出演者名151よりも出演者名152のほうが信頼度が高い場合の例を示している。図9に示す表示例によれば、出演者名152の方が出演者名151よりも大きく表示されることにより、ユーザは、演者名152のほうが表示内容が正しいということを直感的に知ることができる。すなわち、ユーザは、表示内容の正しさを直観的に判断することができる。また、出演者名の表示態様の方法は、文字サイズの大きさの変更に代えて、あるいは併用して、文字または背景色の変更、透過度、表示位置等の変更であってもよい。いずれの場合も、同様の効果を得ることができる。 In FIG. 9, two persons are shown on the display screen 150 which is a television screen. The performer name 151 is also written in the face area 153 of the left person. Similarly, the performer name 152 is also written in the face area 154 of the right person. Here, the performer name of FIG. 3 is displayed with a larger character size as the reliability is higher, and an example in which the performer name 152 has a higher reliability than the performer name 151 is shown. According to the display example shown in FIG. 9, when the performer name 152 is displayed larger than the performer name 151, the user intuitively knows that the display content of the performer name 152 is correct. Can do. That is, the user can intuitively determine the correctness of the display content. Also, the method of displaying the performer name may be a change of the character or background color, a change of the transparency, a display position, etc. instead of or in combination with the change of the character size. In either case, the same effect can be obtained.
 以上詳細に説明したように、番組情報表示装置100において、番組取得部101は、動画像である番組と、出演者名を含む番組情報とを取得し、関連画像取得部107は、出演者名に基づいてネットワークから関連画像を取得し、特徴量算出部108は、関連画像から人物の顔領域を切り出した関連抽出画像の特徴量を算出する。そして、代表特徴量決定部109は、関連抽出画像の特徴量と関連画像の入手先の情報とに基づいて、代表特徴量とその信頼度とを決定し、出演者情報管理部104は、決定された代表特徴量および信頼度を出演者名と対応付けて出演者情報として管理する。一方、表示画像取得部102は、動画像を構成するフレームから表示画像を取得し、特徴量算出部111は、表示画像について表示抽出画像の特徴量を算出する。そして、検索部105は、特徴量算出部111により算出された表示抽出画像の特徴量と、出演者情報保持部110に保持された代表特徴量との類似度を算出し、類似度が最大となる代表特徴量に関連付けられた出演者名を取得する。そして、表示情報生成部106は、信頼度または類似度の少なくとも一方と、検索部105により取得した出演者名と表示抽出画像の領域とに基づいて表示情報を生成し、表示部103は、出演者名を顔領域に関連付けて表示する。 As described in detail above, in the program information display device 100, the program acquisition unit 101 acquires a program that is a moving image and program information including a performer name, and the related image acquisition unit 107 performs a performer name. Based on the network, a related image is acquired from the network, and the feature amount calculation unit 108 calculates a feature amount of a related extracted image obtained by cutting out a human face area from the related image. Then, the representative feature amount determination unit 109 determines the representative feature amount and its reliability based on the feature amount of the related extracted image and the information on the acquisition destination of the related image, and the performer information management unit 104 determines The representative feature amount and the reliability are managed as performer information in association with the performer name. On the other hand, the display image acquisition unit 102 acquires a display image from the frames constituting the moving image, and the feature amount calculation unit 111 calculates the feature amount of the display extracted image for the display image. Then, the search unit 105 calculates the similarity between the feature amount of the display extracted image calculated by the feature amount calculation unit 111 and the representative feature amount held in the performer information holding unit 110, and the similarity is the maximum. The name of the performer associated with the representative feature amount is acquired. Then, the display information generation unit 106 generates display information based on at least one of reliability or similarity, the performer name acquired by the search unit 105, and the region of the display extracted image, and the display unit 103 The name of the person is displayed in association with the face area.
 このように、番組情報表示装置100は、動画像の番組情報から出演者情報を得て、ネットワークから出演者の顔画像を取得し、画像特徴量を求める。そして、番組情報表示装置100は、動画像から顔画像を抽出して得られる画像特徴量から、出演者を特定して、出演者名を動画像に重畳して表示する。これにより、番組情報表示装置100は、事前に出演者の顔画像データベースを用意することなく、出演者の領域に関連付けて出演者名を表示することができる。また、番組情報表示装置100は、動画像から顔画像を抽出して得られる画像特徴量から出演者を特定する際に、動的に検索するデータベースに基づいた信頼度を反映させる。これにより、ノイズが少なく、精度の良い検索を可能とする効果がある。また、この際に、番組情報表示装置100は、出演者の判定結果の正しさを、表示形態の違いによって表示する。これにより、ユーザが出演者の判定結果を直観的に判断することができるという効果が得られる。例えば、ユーザが、番組視聴中に、登場人物と出演者名が一致しないとき、画面上の顔と対応付けて出演者名を知りたいという要求に応えることができる。また、出演者の判定結果がどの程度信頼できるかを直感的に把握させることができる。 As described above, the program information display apparatus 100 obtains performer information from the program information of the moving image, obtains the performer's face image from the network, and obtains the image feature amount. And the program information display apparatus 100 specifies a performer from the image feature-value obtained by extracting a face image from a moving image, and superimposes and displays a performer name on a moving image. Thereby, the program information display apparatus 100 can display a performer name in association with a performer's area without preparing a performer's face image database in advance. Moreover, the program information display apparatus 100 reflects the reliability based on the database searched dynamically, when specifying a performer from the image feature-value obtained by extracting a face image from a moving image. As a result, there is an effect of enabling a search with less noise and high accuracy. At this time, the program information display device 100 displays the correctness of the determination result of the performer according to the difference in display form. Thereby, the effect that a user can judge the determination result of a performer intuitively is acquired. For example, when the user does not agree with the name of the performer while watching the program, the user can respond to a request to know the name of the performer in association with the face on the screen. Moreover, it can be made to grasp | ascertain intuitively how reliable the determination result of a performer is.
 また、本実施の形態の代表特徴量決定部109は、関連画像が公式サイトから取得した画像であるかどうかを判定して関連画像に重み付けするので、特徴量の信頼度を高める効果を得ることができる。 In addition, since the representative feature amount determination unit 109 according to the present embodiment determines whether the related image is an image acquired from the official site and weights the related image, an effect of increasing the reliability of the feature amount can be obtained. Can do.
 (実施の形態2)
 図10は、本発明の実施の形態2に係る番組情報表示装置を備える番組情報表示システムの構成を示す図である。図1と同一構成部分には同一符号を付して重複箇所の説明を省略する。
(Embodiment 2)
FIG. 10 is a diagram showing a configuration of a program information display system including a program information display device according to Embodiment 2 of the present invention. The same components as those in FIG. 1 are denoted by the same reference numerals, and description of overlapping portions is omitted.
 図10において、番組情報表示装置300は、図1の番組情報表示装置100にさらに検索情報保持部301を備えて構成される。 10, the program information display device 300 includes a search information holding unit 301 in addition to the program information display device 100 of FIG.
 検索情報保持部301は、検索部105による検索結果を保持する。検索部105による検索結果とは、上述の通り、代表画像特徴量、信頼度、表示抽出画像の領域、類似度、および出演者名である。 The search information holding unit 301 holds search results obtained by the search unit 105. The search results by the search unit 105 are, as described above, the representative image feature amount, the reliability, the display extracted image region, the similarity, and the performer name.
 図11は、番組情報表示装置300の出演者名の表示までの処理を示すフローチャートである。図5に示すフローと同一処理を行うステップには同一ステップ番号を付して重複箇所の説明を省略する。 FIG. 11 is a flowchart showing processing up to display of the performer name of the program information display apparatus 300. Steps that perform the same processing as the flow shown in FIG. 5 are denoted by the same step numbers, and description of overlapping portions is omitted.
 ステップS23において、検索部105は、「すべての関連画像」について以下のループの始端と終端との間の処理を繰り返す。 In step S23, the search unit 105 repeats the process between the start and end of the following loop for “all related images”.
 ステップS31において、検索部105は、新たに抽出された表示抽出画像(以下「新規の表示抽出画像」という)の領域と近接する、過去に抽出された表示抽出画像(過去の表示抽出画像という)の領域が、検索情報保持部301に登録されているか否かを確認する。 In step S31, the search unit 105 closes the area of the newly extracted display extraction image (hereinafter referred to as “new display extraction image”) and has been extracted in the past (referred to as past display extraction image). It is confirmed whether or not this area is registered in the search information holding unit 301.
 ステップS32において、検索部105は、新規の表示抽出画像の領域と近接する過去の表示抽出画像の領域が検索情報保持部301に登録されている場合、ステップS33において、検索部105は、新規の表示抽出画像の抽出画像特徴量と、該当する過去の表示抽出画像の代表特徴量との照合を行い、類似度を算出して、ステップS26に進む。 In step S <b> 32, when the past display extracted image area close to the new display extracted image area is registered in the search information holding unit 301 in step S <b> 32, the search unit 105 selects the new display extracted image area in step S <b> 33. The extracted image feature quantity of the display extracted image is compared with the representative feature quantity of the corresponding past display extracted image, the similarity is calculated, and the process proceeds to step S26.
 検索部105は、新規の表示抽出画像の領域と近接する過去の表示抽出画像の領域が、検索情報保持部301に登録されていない場合は、ステップS24に以下に進み、図5の出演者名を決定するステップS27までと同様である。 If the area of the past display extraction image adjacent to the area of the new display extraction image is not registered in the search information holding unit 301, the search unit 105 proceeds to the following in step S24 and performs the performer name in FIG. This is the same as the process up to step S27.
 ステップS27において、検索部105は、該当する代表特徴量に関連付けられている出演者名を、検索結果として保持する。 In step S27, the search unit 105 holds the performer name associated with the corresponding representative feature amount as a search result.
 ステップS34において、検索情報保持部301は、出演者名を決定した後に、代表画像特徴量、信頼度、表示抽出画像の領域、類似度、および出演者名を保持して、ステップS28に進む。 In step S34, after determining the performer name, the search information retaining unit 301 retains the representative image feature value, the reliability, the region of the display extracted image, the similarity, and the performer name, and proceeds to step S28.
 ここで、上記ステップS33に進んだ場合、検索部105は、最大となる類似度を算出して、ステップS26に進むことになる。これは、新規の表示抽出画像の領域と近接する過去の表示抽出画像は、ほとんどの場合、同一人物の顔画像だからである。この場合、ステップS26において、検索部105は、近接する過去の表示抽出画像に対する類似度と、検索情報保持部301に登録されている、関連画像に対する類似度とを比較することになる。 Here, when the process proceeds to step S33, the search unit 105 calculates the maximum similarity and proceeds to step S26. This is because the past display extraction image adjacent to the area of the new display extraction image is almost always a face image of the same person. In this case, in step S <b> 26, the search unit 105 compares the similarity to the past display extracted image and the similarity to the related image registered in the search information holding unit 301.
 類似度が予め定めた閾値以内でわずかに下がる場合は、特徴量の比較における誤差と考えられる。したがって、この場合、検索部105は、大きい方の類似度を保持する。これにより、検索部105は、同一人物の顔領域を捕らえている間、類似度が上がっていく方向にのみ、類似度の抽出処理を作用させることができる。 If the degree of similarity slightly falls within a predetermined threshold, it is considered an error in the feature amount comparison. Therefore, in this case, the search unit 105 holds the larger similarity. Thereby, the search part 105 can make the similarity extraction process act only in the direction in which the similarity increases while capturing the face area of the same person.
 一方で、類似度が予め定めた閾値を超えて大きく下がる場合は、場面の切り替え等により、ほぼ同一となる領域に、別の出演者が映し出されていると判断できる。したがって、この場合、検索部105は、検索結果を保持する処理(ステップS34の処理)に代えて、無効のデータを登録する処理、つまり、その近接する過去の表示抽出領域に関する登録データを破棄する処理となる。 On the other hand, if the degree of similarity falls greatly exceeding a predetermined threshold, it can be determined that another performer is projected in the almost same area by switching scenes. Therefore, in this case, the search unit 105 discards the registration data related to the past past display extraction area, that is, the process of registering invalid data instead of the process of holding the search result (the process of step S34). It becomes processing.
 このように、本実施の形態では、検索部105が検索した結果を保持する検索情報保持部301を有する。検索部105は、検索情報保持部301に保持された検索結果に対して、表示抽出画像の位置が近いものに限定して検索を行う。これにより、検索部105は、表示抽出画像と登録されているすべての代表画像の特徴量の類似度とを比較する必要がなくなり、処理を高速化することが可能となる。 As described above, in this embodiment, the search unit 105 includes the search information holding unit 301 that holds the search result. The search unit 105 searches the search result held in the search information holding unit 301 only for the display extracted image close to the search result. As a result, the search unit 105 does not need to compare the display extracted image with the similarity of the feature amounts of all registered representative images, and can speed up the processing.
 (実施の形態3)
 図12は、本発明の実施の形態3に係る番組情報表示装置を備える番組情報表示システムの構成を示す図である。本実施の形態は、図1と同一構成部分には同一符号を付して重複箇所の説明を省略する。
(Embodiment 3)
FIG. 12 is a diagram showing a configuration of a program information display system including a program information display device according to Embodiment 3 of the present invention. In the present embodiment, the same components as those in FIG.
 図12において、番組情報表示システムは、番組情報表示装置400、ネットワーク200、画像検索装置500、画像サーバ220、番組情報サーバ230、番組表サーバ240、および放送局250を備えて構成される。すなわち、本実施の形態に係る番組情報表示システムは、実施の形態1の番組情報表示装置100および画像検索装置210に代えて、番組情報表示装置400および画像検索装置500を配置している。 12, the program information display system includes a program information display device 400, a network 200, an image search device 500, an image server 220, a program information server 230, a program guide server 240, and a broadcast station 250. That is, in the program information display system according to the present embodiment, a program information display device 400 and an image search device 500 are arranged instead of the program information display device 100 and the image search device 210 of the first embodiment.
 番組情報表示装置400は、放送局250から送信される番組情報を再生するデジタル放送受信機である。番組情報表示装置400は、番組取得部101、表示画像取得部102、表示部103、検索部105、表示情報生成部106、および特徴量算出部111を備えて構成される。すなわち、番組情報表示装置400は、実施の形態1の番組情報表示装置100の機能部のうち、画像表示に関する機能部を備えた構成を有している。 The program information display device 400 is a digital broadcast receiver that reproduces program information transmitted from the broadcast station 250. The program information display device 400 includes a program acquisition unit 101, a display image acquisition unit 102, a display unit 103, a search unit 105, a display information generation unit 106, and a feature amount calculation unit 111. That is, the program information display device 400 has a configuration including a functional unit related to image display among the functional units of the program information display device 100 of the first embodiment.
 画像検索装置500は、画像検索サイトである。画像検索装置500は、出演者情報管理部104、関連画像取得部107、特徴量算出部108、代表特徴量決定部109、出演者情報保持部110、画像情報保持部501、および画像検索部502を備えて構成される。すなわち、画像検索装置500は、実施の形態1の番組情報表示装置100の機能部のうち、関連画像検索に関する機能部と、画像情報保持部501と、画像検索部502とを備えた構成を有している。画像情報保持部501および画像検索部502は、実施の形態1の画像検索装置210の機能部に対応する。 The image search device 500 is an image search site. The image search apparatus 500 includes a performer information management unit 104, a related image acquisition unit 107, a feature amount calculation unit 108, a representative feature amount determination unit 109, a performer information holding unit 110, an image information holding unit 501, and an image search unit 502. It is comprised with. That is, the image search device 500 has a configuration including a function unit related to related image search, an image information holding unit 501, and an image search unit 502 among the function units of the program information display device 100 of the first embodiment. is doing. The image information holding unit 501 and the image search unit 502 correspond to the functional units of the image search apparatus 210 according to the first embodiment.
 すなわち、本実施の形態に係る番組情報表示システムは、実施の形態1の番組情報表示装置400の関連画像検索に関する機能部を、実施の形態1の画像検索装置210に移動させた構成を有している。 That is, the program information display system according to the present embodiment has a configuration in which the function unit related to the related image search of the program information display device 400 of the first embodiment is moved to the image search device 210 of the first embodiment. ing.
 番組情報表示装置400および画像検索装置500は、それぞれ、実施の形態1の番組情報表示装置100と同様に、CPUがROM等の記憶媒体に記憶されたプログラムを実行する構成としてもよい。 The program information display device 400 and the image search device 500 may each be configured such that the CPU executes a program stored in a storage medium such as a ROM, similarly to the program information display device 100 of the first embodiment.
 なお、番組情報表示装置400に代えて、実施の形態1の番組情報表示装置100または実施の形態2の番組情報表示装置300に、本実施の形態の画像検索装置500を組み9合わせる態様でもよい。 Instead of the program information display device 400, the program information display device 100 of the first embodiment or the program information display device 300 of the second embodiment may be combined with the image search device 500 of the present embodiment. .
 以上の構成において、画像検索装置500の関連画像取得部107は、URLのリンク等を基にネットワーク200経由で画像を取得する。関連画像取得部107は、取得した画像と、該当画像の説明あるいは該当ページの内容から得られるキーワードとその画像のURLとを画像情報保持部501に保持する。 In the above configuration, the related image acquisition unit 107 of the image search apparatus 500 acquires an image via the network 200 based on a URL link or the like. The related image acquisition unit 107 holds the acquired image, the keyword obtained from the description of the corresponding image or the content of the corresponding page, and the URL of the image in the image information holding unit 501.
 画像検索部502は、画像を検索する際、ネットワーク200経由で、番組情報表示装置400から番組情報を受け取り、受け取った番組情報を基に、画像情報保持部501から画像情報として該当するURLを返す。 When searching for an image, the image search unit 502 receives program information from the program information display device 400 via the network 200, and returns a corresponding URL as image information from the image information holding unit 501 based on the received program information. .
 このように構成することで、番組情報表示装置400から、代表特徴量を決定するまでの処理がなくなる。また、検索部105は、出演者名をキーワードとして、ネットワーク113を介して代表特徴量と信頼度を受け取る。これらの点が、実施の形態1からの変更となる。 This configuration eliminates the processing from the program information display device 400 until the representative feature amount is determined. Further, the search unit 105 receives the representative feature amount and the reliability via the network 113 using the performer name as a keyword. These points are changes from the first embodiment.
 このように、本実施の形態では、代表特徴量を決定するまでの処理を画像検索装置500で行うことにより、機能の分散による番組情報表示装置400の処理の軽減および高速化が期待できる。 As described above, in this embodiment, the processing until the representative feature amount is determined is performed by the image search device 500, so that the processing of the program information display device 400 can be expected to be reduced and speeded up due to the distribution of functions.
 なお、本実施の形態では、画像情報保持部501と出演者情報保持部110が別個の構成となっている例を説明したが、これらは一体の構成であってもよい。また、画像検索装置500が番組情報を番組情報表示装置400から受け取る例について説明した。しかし、番組情報を番組情報サーバ230から受け取って、代表特徴量を決定する処理を事前に実行し、番組情報表示装置400から出演者名と番組名を受け取るようにしてもよい。 In the present embodiment, the example in which the image information holding unit 501 and the performer information holding unit 110 are configured separately has been described, but these may be integrated. Further, the example in which the image search device 500 receives program information from the program information display device 400 has been described. However, the program information may be received from the program information server 230, and the process of determining the representative feature amount may be executed in advance to receive the performer name and the program name from the program information display device 400.
 以上の説明は本発明の好適な実施の形態の例証であり、本発明の範囲はこれに限定されることはない。上記各実施の形態では、放送番組について説明したが、番組情報を用いる装置および方法であればよく、放送番組に限定されるものではない。 The above description is an illustration of a preferred embodiment of the present invention, and the scope of the present invention is not limited to this. In each of the above embodiments, a broadcast program has been described. However, any apparatus and method using program information may be used, and the present invention is not limited to a broadcast program.
 また、各実施の形態では、番組情報表示装置、および番組情報表示方法という名称を用いた。しかし、これは、説明の便宜上であり、番組情報表示装置は、画像検索装置、番組再生装置、番組情報表示方法は番組情報検索方法等であってもよい。 In each embodiment, the names of the program information display device and the program information display method are used. However, this is for convenience of explanation, and the program information display device may be an image search device, a program playback device, and the program information display method may be a program information search method or the like.
 さらに、上記番組情報表示装置および方法を構成する各部、例えば番組取得部、出演者情報保持部の種類、その数および接続方法等は限定されるものではない。 Furthermore, each part constituting the program information display apparatus and method, for example, the type, number and connection method of the program acquisition part and performer information holding part are not limited.
 また、番組表を構成する番組詳細情報は、番組表サーバからネットワークを介して取得するとしたが、番組詳細情報が放送波に重畳されている場合には、放送波から取得してもよい。 In addition, the detailed program information constituting the program guide is acquired from the program guide server via the network. However, when the detailed program information is superimposed on the broadcast wave, it may be acquired from the broadcast wave.
 番組がDVDあるいはハードディスク等のメディアに保持されている場合には、同じメディアに、出演者の画像データと出演者名とを対応付けた情報が、番組とは別個に保持されていることも多い。この場合には、ネットワーク経由ではなく、メディアから関連画像を取得するようにしてもよい。また、この場合には、放送局および番組表サーバが番組情報表示システムの全体構成に含まれなくともよい。 When a program is held on a medium such as a DVD or a hard disk, information in which the image data of the performer and the name of the performer are associated with each other is often held separately from the program. . In this case, the related image may be acquired from the media instead of via the network. In this case, the broadcast station and the program guide server may not be included in the overall configuration of the program information display system.
 また、番組情報装置は、テレビジョン受信機への適用に限定されるものではなく、出演者が表示される画像を表示する、他の各種装置にも適用することができる。例えば、番組情報表示装置は、BD(Blue-ray Disc)プレイヤ、DVDプレイヤ、ハードディスクレコーダ等の動画像を表示または再生する装置に適用することができる。さらに、番組情報表示装置は、携帯電話機/PHS(Personal Handy-Phone System)、携帯情報端末(以下、PDA(Personal Digital Assistants)という)等の携帯端末、パソコン、さらには携帯ゲーム機等にも適用することができる。 Also, the program information device is not limited to application to a television receiver, and can be applied to other various devices that display images on which performers are displayed. For example, the program information display device can be applied to a device that displays or reproduces a moving image, such as a BD (Blue-ray Disc) player, a DVD player, or a hard disk recorder. Furthermore, the program information display device is also applicable to portable terminals such as mobile phones / PHS (Personal Handy-Phone System), portable information terminals (hereinafter referred to as PDA (Personal Digital Assistants)), personal computers, and portable game machines. can do.
 以上説明した番組情報表示装置および番組情報表示方法は、この番組情報表示方法を機能させるためのプログラムでも実現される。このプログラムはコンピュータで読み取り可能な記録媒体に格納されている。 The program information display device and the program information display method described above are also realized by a program for causing the program information display method to function. This program is stored in a computer-readable recording medium.
 2008年3月24日出願の特願2008-076310の日本出願に含まれる明細書、図面および要約書の開示内容は、すべて本願に援用される。 The disclosure of the description, drawings and abstract contained in the Japanese application of Japanese Patent Application No. 2008-076310 filed on Mar. 24, 2008 is incorporated herein by reference.
 本発明に係る番組情報表示装置および番組情報表示方法は、動画像から人物の顔領域を抽出し、出演者名を人物の顔領域に関連付けて表示するテレビ受信機、携帯電話、DVD再生機、パソコン、ゲーム機等の映像再生端末に適用することができる。 A program information display apparatus and a program information display method according to the present invention extract a person's face area from a moving image and display a performer name in association with the person's face area, a mobile phone, a DVD player, It can be applied to video playback terminals such as personal computers and game machines.

Claims (5)

  1.  動画像である番組と、出演者名を含む番組情報と、を取得する番組取得部と、
     前記出演者名に基づいて、ネットワークから関連画像を取得する関連画像取得部と、
     前記関連画像から人物の領域を切り出した第1の抽出画像の特徴量を算出する第1特徴量算出部と、
     前記第1の抽出画像の特徴量と前記関連画像の入手先の情報とに基づいて、代表特徴量とその信頼度とを決定する代表特徴量決定部と、
     決定された前記代表特徴量及び信頼度を、前記出演者名と対応付けて出演者情報として管理する出演者情報管理部と、
     前記動画像を構成するフレームから表示画像を取得する表示画像取得部と、
     前記表示画像から人物の領域を切り出した第2の抽出画像の特徴量を算出する第2特徴量算出部と、
     前記第2特徴量算出部により算出された前記第2の抽出画像の特徴量と、前記出演者情報管理部に保持された前記代表特徴量と、の類似度を算出し、前記類似度が最大となる代表特徴量に関連付けられた出演者名を取得する検索部と、
     前記信頼度又は前記類似度の少なくとも一方と、前記検索部により取得された前記出演者名と、前記第2の抽出画像の領域と、に基づいて表示情報を生成する表示情報生成部と、
     前記表示画像及び前記表示情報を表示する表示部と、
     を備える番組情報表示装置。
    A program acquisition unit for acquiring a program that is a moving image and program information including a performer name;
    Based on the performer name, a related image acquisition unit that acquires a related image from the network;
    A first feature amount calculation unit that calculates a feature amount of a first extracted image obtained by cutting out a person's region from the related image;
    A representative feature amount determining unit that determines a representative feature amount and its reliability based on the feature amount of the first extracted image and the information on the source of the related image;
    Performer information management unit for managing the determined representative feature amount and reliability as performer information in association with the performer name;
    A display image acquisition unit for acquiring a display image from a frame constituting the moving image;
    A second feature amount calculating unit that calculates a feature amount of a second extracted image obtained by cutting out a person's region from the display image;
    The similarity between the feature amount of the second extracted image calculated by the second feature amount calculation unit and the representative feature amount held in the performer information management unit is calculated, and the similarity is maximum. A search unit that obtains performer names associated with representative feature quantities,
    A display information generating unit that generates display information based on at least one of the reliability or the similarity, the performer name acquired by the search unit, and the region of the second extracted image;
    A display unit for displaying the display image and the display information;
    A program information display device comprising:
  2.  前記関連画像取得部は、前記ネットワーク上のサイトから番組出演者の関連画像を取得し、
     前記代表特徴量決定部は、取得先の前記サイトに応じて前記関連画像に重み付けして前記信頼度を決定する、
     請求項1記載の番組情報表示装置。
    The related image acquisition unit acquires a related image of a program performer from a site on the network,
    The representative feature amount determination unit determines the reliability by weighting the related image according to the site of the acquisition destination.
    The program information display device according to claim 1.
  3.  前記表示情報生成部は、前記信頼度又は前記類似度に基づいて、透過度、表示色、文字サイズ、又は表示位置の少なくともいずれか一つを変更する表示情報を生成する、
     請求項1記載の番組情報表示装置。
    The display information generation unit generates display information for changing at least one of transparency, display color, character size, or display position based on the reliability or the similarity.
    The program information display device according to claim 1.
  4.  前記検索部による検索結果を保持する検索情報保持部、をさらに備え、
     前記検索部は、前記検索情報保持部に保持された検索結果に対して、前記出演者情報に基づいて検索を行う、
     請求項1記載の番組情報表示装置。
    A search information holding unit for holding a search result by the search unit;
    The search unit searches the search result held in the search information holding unit based on the performer information.
    The program information display device according to claim 1.
  5.  動画像である番組と、出演者名とを含む番組情報と、を取得するステップと、
     前記出演者名に基づいて、ネットワークから関連画像を取得するステップと、
     前記関連画像から人物の領域を切り出した第1の抽出画像の特徴量を算出する第1特徴量算出ステップと、
     前記第1の抽出画像の特徴量と前記関連画像の入手先の情報とに基づいて、代表特徴量とその信頼度とを決定するステップと、
     決定された前記代表特徴量及び信頼度を、前記出演者名と対応付けて出演者情報として管理するステップと、
     前記動画像を構成するフレームから表示画像を取得するステップと、
     前記表示画像から人物の領域を切り出した第2の抽出画像の特徴量を算出する第2特徴量算出ステップと、
     前記第2特徴量算出ステップにより算出された前記第2の抽出画像の特徴量と、保持された前記代表特徴量と、の類似度を算出し、前記類似度が最大となる代表特徴量に関連付けられた出演者名を取得するステップと、
     前記信頼度又は前記類似度の少なくとも一方と、前記検索部により取得された前記出演者名と、前記第2の抽出画像の領域と、に基づいて表示情報を生成するステップと、
     前記表示画像及び前記表示情報を表示するステップと
     を有する番組情報表示方法。
    Obtaining program information including a program that is a moving image and a performer name;
    Obtaining a related image from the network based on the performer name;
    A first feature amount calculating step of calculating a feature amount of a first extracted image obtained by cutting out a person's region from the related image;
    Determining a representative feature amount and its reliability based on the feature amount of the first extracted image and the information on the source of the related image;
    Managing the determined representative feature quantity and reliability as performer information in association with the performer name;
    Obtaining a display image from a frame constituting the moving image;
    A second feature amount calculating step of calculating a feature amount of a second extracted image obtained by cutting out a person region from the display image;
    The similarity between the feature quantity of the second extracted image calculated by the second feature quantity calculation step and the retained representative feature quantity is calculated, and is associated with the representative feature quantity having the maximum similarity. Obtaining a given performer name,
    Generating display information based on at least one of the reliability or the similarity, the performer name acquired by the search unit, and the region of the second extracted image;
    A program information display method comprising: displaying the display image and the display information.
PCT/JP2009/001274 2008-03-24 2009-03-23 Program information display device and program information display method WO2009119063A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2008-076310 2008-03-24
JP2008076310A JP2009232250A (en) 2008-03-24 2008-03-24 Program information display apparatus and program information display method

Publications (1)

Publication Number Publication Date
WO2009119063A1 true WO2009119063A1 (en) 2009-10-01

Family

ID=41113274

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2009/001274 WO2009119063A1 (en) 2008-03-24 2009-03-23 Program information display device and program information display method

Country Status (2)

Country Link
JP (1) JP2009232250A (en)
WO (1) WO2009119063A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2016167654A (en) * 2015-03-09 2016-09-15 株式会社鳥山研究室 Program reproducing apparatus and program reproducing system
CN111712807A (en) * 2018-02-16 2020-09-25 麦克赛尔株式会社 Portable information terminal, information presentation system, and information presentation method

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8718444B2 (en) 2010-06-16 2014-05-06 Panasonic Corporation Video search device, video search method, recording medium, program, and integrated circuit
JP2012231404A (en) * 2011-04-27 2012-11-22 Toshiba Corp Image processing device and image processing method
JP5834541B2 (en) * 2011-06-29 2015-12-24 三菱電機株式会社 Digital broadcast receiving apparatus and digital broadcast receiving method
JP2013073392A (en) * 2011-09-27 2013-04-22 Fujitsu Ltd Display control device, display control program, and display control method
JP2014119975A (en) * 2012-12-17 2014-06-30 Nippon Hoso Kyokai <Nhk> Video meta data application device and program
JP2016046636A (en) * 2014-08-21 2016-04-04 日本電気株式会社 Operation control device, operation control method and operation control program
JP2017033390A (en) * 2015-08-04 2017-02-09 日本放送協会 Image analysis device and program
JP6065086B2 (en) * 2015-11-04 2017-01-25 三菱電機株式会社 Digital broadcast receiving apparatus and digital broadcast receiving method
WO2017134738A1 (en) * 2016-02-02 2017-08-10 三菱電機株式会社 Recorder device and video monitoring system
JP6161224B1 (en) 2016-12-28 2017-07-12 アンバス株式会社 Person information display device, person information display method, and person information display program
CN109525877B (en) * 2018-10-18 2021-04-20 百度在线网络技术(北京)有限公司 Video-based information acquisition method and device

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2007251296A (en) * 2006-03-14 2007-09-27 Sony Corp Program receiver, program receiving method, program of program receiving method and recording medium recording program of program receiving method
JP2008263305A (en) * 2007-04-10 2008-10-30 Toshiba Corp Electronic apparatus and video recording control method

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2007251296A (en) * 2006-03-14 2007-09-27 Sony Corp Program receiver, program receiving method, program of program receiving method and recording medium recording program of program receiving method
JP2008263305A (en) * 2007-04-10 2008-10-30 Toshiba Corp Electronic apparatus and video recording control method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
"Proceedings of the Tenth IEEE International Conference on Computer Vision (ICCV'05)", vol. 2, 17 October 2005, article R.FERGUS ET AL.: "Learning object categories from Google's image search", pages: 1816 - 1823 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2016167654A (en) * 2015-03-09 2016-09-15 株式会社鳥山研究室 Program reproducing apparatus and program reproducing system
CN111712807A (en) * 2018-02-16 2020-09-25 麦克赛尔株式会社 Portable information terminal, information presentation system, and information presentation method

Also Published As

Publication number Publication date
JP2009232250A (en) 2009-10-08

Similar Documents

Publication Publication Date Title
WO2009119063A1 (en) Program information display device and program information display method
US9292519B2 (en) Signature-based system and method for generation of personalized multimedia channels
EP2541963B1 (en) Method for identifying video segments and displaying contextually targeted content on a connected television
JP5740814B2 (en) Information processing apparatus and method
EP2417767B1 (en) Apparatus and method for providing information related to broadcasting programs
US9100701B2 (en) Enhanced video systems and methods
EP2553652B1 (en) Media fingerprinting for content determination and retrieval
US9251532B2 (en) Method and apparatus for providing search capability and targeted advertising for audio, image, and video content over the internet
US20140082663A1 (en) Methods for Identifying Video Segments and Displaying Contextually Targeted Content on a Connected Television
JP2009043156A (en) Apparatus and method for searching for program
WO2009061434A1 (en) System and method for processing digital media
CN111078931B (en) Song list pushing method, device, computer equipment and storage medium
WO2010021102A1 (en) Related scene addition device and related scene addition method
TW200834355A (en) Information processing apparatus and method, and program
JP6202815B2 (en) Character recognition device, character recognition method, and character recognition program
US20020067856A1 (en) Image recognition apparatus, image recognition method, and recording medium
CN109710801A (en) A kind of video searching method, terminal device and computer storage medium
US20150020087A1 (en) System for Identifying Features in a Television Signal
JP2008129884A (en) Information retrieval system, its method, and broadcast receiver used therefor
JP5335500B2 (en) Content search apparatus and computer program
CN110942070B (en) Content display method, device, electronic equipment and computer readable storage medium
JP2008139942A (en) Information processing apparatus, information processing method and program
CN103731568B (en) Export information output apparatus and the method for the information identical with other devices
JPWO2019082606A1 (en) Content management device, content management system, and control method
JP2009103945A (en) Video content processing device and program

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 09723999

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 09723999

Country of ref document: EP

Kind code of ref document: A1