WO2020090790A1 - Information processing device - Google Patents

Information processing device Download PDF

Info

Publication number
WO2020090790A1
WO2020090790A1 PCT/JP2019/042300 JP2019042300W WO2020090790A1 WO 2020090790 A1 WO2020090790 A1 WO 2020090790A1 JP 2019042300 W JP2019042300 W JP 2019042300W WO 2020090790 A1 WO2020090790 A1 WO 2020090790A1
Authority
WO
WIPO (PCT)
Prior art keywords
keyword
user
unit
association
keywords
Prior art date
Application number
PCT/JP2019/042300
Other languages
French (fr)
Japanese (ja)
Inventor
田中 彰
翔 七尾
広樹 石塚
昇悟 池田
充弘 小形
誠 村▲崎▼
Original Assignee
株式会社Nttドコモ
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社Nttドコモ filed Critical 株式会社Nttドコモ
Priority to JP2020553924A priority Critical patent/JPWO2020090790A1/en
Publication of WO2020090790A1 publication Critical patent/WO2020090790A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/903Querying
    • G06F16/9032Query formulation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/903Querying
    • G06F16/9035Filtering based on additional data, e.g. user or group profiles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/10Speech classification or search using distance or distortion measures between unknown speech and reference templates

Definitions

  • the present invention relates to an information processing device.
  • Patent Document 1 discloses a technique in which when a user specifies an object image in an image displayed on a display device using a pointing device, a recommendation regarding the object image is displayed. Further, Patent Document 2 discloses a technique of recommending information about an object image designated by a voice of the user in response to the voice of the user.
  • the user needs to specify the object image. That is, when the user does not specify the object image, it is not possible to generate a comment such as a recommendation in response to the user's ambiguous statement.
  • an information processing apparatus includes a first keyword generating unit that generates a first keyword based on a voice of a user, and a plurality of units extracted from an image indicated by an image signal.
  • Second keyword generation unit that generates a plurality of second keywords corresponding to the object image of 1: 1, and the plurality of second keywords based on the degree of association between each of the plurality of second keywords and the first keyword.
  • comment generation unit that generates a comment related to the target keyword.
  • a comment can be generated in response to a user's ambiguous statement without the user designating an object image to be a comment target.
  • FIG. 1 is a block diagram showing the overall configuration of a service system according to the first embodiment of the present invention.
  • the service system 1 shown in FIG. 1 provides a moving image distribution service.
  • the moving image distribution service provides, for example, movies or terrestrial digital broadcasting contents.
  • the service system 1 includes user devices 20_1 to 20_m (m is an integer of 1 or more) managed by users U_1 to U_m, a network NW, and a video distribution server 10.
  • m is an integer of 1 or more
  • NW a network
  • video distribution server 10 a video distribution server 10.
  • the user device 20 is an information processing device that processes various types of information.
  • the user device 20 is, for example, a portable information processing device such as a smartphone or a tablet terminal. However, as the user device 20, any information processing device can be adopted.
  • the user device 20 may be, for example, a terminal-type information device such as a personal computer.
  • the user device 20 receives the image signal Sg transmitted from the moving image distribution server 10 and displays the image, or transmits the image signal Sg to the television receiver 30 and displays the image on the television receiver 30.
  • Can be made The user U may speak while watching a moving image. For example, the user U may make a comment or tweet about the moving image. In this case, although the statement of the user U is related to the moving image, it is not possible to uniquely specify which object included in the image of the moving picture the statement is related to because the statement is ambiguous. There are many.
  • the user device 20 has a function of generating a comment such as a recommendation in response to the vague statement of the user U.
  • FIG. 2 is a block diagram illustrating a hardware configuration of the user device 20.
  • the user device 20 is a computer system including a processing device 21, a storage device 22, a communication device 23, an output device 24, an input device 25, a short-range wireless communication device 26, a GPS (Global Positioning System) device 27, and a bus 28. Will be realized.
  • the processing device 21, the storage device 22, the communication device 23, the output device 24, the input device 25, the short-range wireless communication device 26, and the GPS device 27 are connected by a bus 28 for communicating information.
  • the bus 28 may be configured by a single bus or may be configured by different buses among devices. Note that each element of the user device 20 is configured by a single device or a plurality of devices, and some elements of the user device 20 may be omitted.
  • the processing device 21 is a processor that controls the entire user device 20, and is composed of, for example, a single chip or a plurality of chips.
  • the processing device 21 is configured by, for example, a central processing unit (CPU) including an interface with peripheral devices, an arithmetic device, a register, and the like. Note that some or all of the functions of the processing device 21 are realized by hardware such as DSP (Digital Signal Processor), ASIC (Application Specific Integrated Circuit), PLD (Programmable Logic Device), and FPGA (Field Programmable Gate Array). May be.
  • the processing device 21 executes various processes in parallel or sequentially.
  • the storage device 22 is a recording medium that can be read by the processing device 21.
  • the storage device 22 stores a plurality of programs including the control program PRa executed by the processing device 21, a keyword table TBLa, a comment table TBLb, and various data used by the processing device 21.
  • the storage device 22 is composed of one or more types of storage circuits such as a ROM (Read Only Memory), an EPROM (Erasable Programmable ROM), an EEPROM (Electrically Erasable Programmable ROM), and a RAM (Random Access Memory).
  • a plurality of words are stored in the keyword table TBLa. Multiple words are roughly divided into nouns and adjectives. Noun words correspond to keywords. The first keyword KW1 and the second keyword KW2 described later are included in the noun words stored in the keyword table TBLa.
  • the adjective word is stored in association with the noun word. Adjectives have the function of modifying nouns. The association between the adjective word and the noun word is determined according to the word modification relation. For example, the adjective word “delicious” is associated with the noun word “food and drink”.
  • FIG. 3 is an explanatory diagram showing a data structure of a noun word stored in the keyword table TBLa.
  • the data structure of the noun word has a tree structure in which a plurality of words are hierarchized according to the meaning.
  • a plurality of words are classified into the first hierarchy to the fourth hierarchy.
  • the number of layers may be four or more.
  • the keyword table TBLa stores the degree of association indicating the degree of association between the noun word and the noun word.
  • the degree of relevance is determined in consideration of the relationship between the superordinate concept and the subordinate concept, and the use and function of the object indicated by the word. For example, “sake” and “wine” are both subordinate concepts of “sake”.
  • “Ouchiguchi” is not a subordinate concept of “Sake”
  • “Inoguchi” is used to drink “Japanese sake”. For this reason, the degree of association between “sake” and “chef” is higher than the degree of association between “sake” and “wine”.
  • the communication device 23 is a device that communicates with another device via a network NW such as a mobile communication network or the Internet.
  • the communication device 23 is also described as, for example, a network device, a network controller, a network card, or a communication module.
  • the communication device 23 can communicate with the moving image distribution server 10 via the network NW.
  • the output device 24 informs the user U of various information under the control of the processing device 21.
  • the output device 24 includes a display device 241 and a speaker 242.
  • the display device 241 displays an image.
  • various display panels such as a liquid crystal display panel or an organic EL (Electro Luminescence) display panel are preferably used as the display device 241.
  • Sound data is supplied from the processing device 21 to the speaker 242.
  • the speaker 242 includes a DA converter. The sound data is converted into an analog signal by the DA converter, and the speaker 242 is driven by the analog signal.
  • the input device 25 is a device for the user U to input information for using the user device 20.
  • the input device 25 receives an input operation by the user U.
  • the input device 25 of this example includes a microphone 251 and a touch panel 252.
  • the touch panel 252 detects a touch by the user U on the display surface of the display device 241.
  • the touch panel 252 accepts an operation of inputting codes such as numbers and letters and an operation of selecting an icon displayed on the display device 241 based on the contact position.
  • the microphone 251 converts the voice of the user U into an analog electric signal and outputs the electric signal as a sound signal Sa.
  • the audio signal Sa is converted into a digital signal by an AD converter (not shown) and supplied to the processing device 21 via the bus 28.
  • the short-range wireless communication device 26 is a device that communicates with another device by short-range wireless communication. Examples of short-range wireless communication include Bluetooth (registered trademark), ZigBee (registered trademark), and WiFi (registered trademark).
  • the television receiver 30 or the like corresponds to another device.
  • the GPS device 27 receives radio waves from a plurality of satellites and generates position information from the received radio waves.
  • the position information indicates the position of the user device 20.
  • the position information may have any format as long as the position can be specified.
  • the position information indicates, for example, the latitude and longitude of the user device 20. In the present embodiment, the position information is illustrated as being obtained from the GPS device 27, but the user device 20 may acquire the position information by any other method.
  • the user device 20 may acquire the location information by using the cell ID assigned to the base station that is the communication destination of the user device 20.
  • the user device 20 uses the identification address (MAC) on the network assigned to the access point.
  • the position information may be acquired by referring to a database in which (Media Access Control) addresses) and actual addresses (positions) are associated with each other.
  • the user device 20 receives the ID information included in the advertisement packet conforming to the BLE (Bluetooth Low Energy) standard by using the short-range wireless communication device 26, and acquires the position information based on the ID information. May be.
  • FIG. 4 is a functional block diagram showing the functions of the user device 20.
  • the processing device 21 functions as the first keyword generation unit 210, the second keyword generation unit 220A, the identification unit 230A, and the comment generation unit 240 by reading and executing the control program PRa from the storage device 22.
  • the first keyword generation unit 210 generates the first keyword KW1 based on the voice of the user U indicated by the voice signal Sa. Specifically, the first keyword generation unit 210 analyzes the voice of the user U and extracts a noun and an adjective from the analysis result. When the voice of the user U includes both a noun and an adjective, the first keyword generation unit 210 identifies the noun as the attention word. For example, when the voice of the user U is “a red car is suspicious”, the noun “car” is specified as the attention word. Further, when the voice of the user U does not include a noun but does include an adjective, the first keyword generation unit 210 identifies the adjective as the attention word. For example, when the voice of the user U is “it looks delicious”, the adjective “delicious” is specified as the attention word.
  • the first keyword generation unit 210 determines whether the focused word is included in the keyword table TBLa. The first keyword generation unit 210 does not generate the first keyword KW1 when the determination result is negative. Therefore, the first keyword KW1 is limited to the keywords included in the keyword table TBLa. On the other hand, when the determination result is affirmative and the focused word is a noun, the first keyword generation unit 210 generates the focused word as the first keyword KW1. When the determination result is affirmative and the focused word is an adjective, the first keyword generation unit 210 refers to the keyword table TBLa and generates a noun word associated with the focused word as the first keyword KW1. To do. For example, when the attention word is “delicious”, the first keyword generation unit 210 generates “food and drink” as the first keyword KW1.
  • the first keyword generation unit 210 generates the first keyword KW1 related to the remark of the user U even when the remark of the user U is ambiguous.
  • the second keyword generation unit 220A generates the second keyword KW2 for each of the object images extracted from the image indicated by the image signal Sg.
  • the second keyword generation unit 220A has an extraction unit 221 and a conversion unit 222.
  • the extraction unit 221 extracts a plurality of object images from the image indicated by the image signal Sg. There are many object images in one screen image.
  • the images extracted by the extraction unit 221 are, for example, the object images OB1 to OB5.
  • the conversion unit 222 converts each of the plurality of object images OB1 to OB5 extracted by the extraction unit 221 into the second keyword KW2.
  • the conversion unit 222 converts each object image OB into the second keyword KW2 using, for example, an image recognition model learned by machine learning.
  • the second keyword KW2 is included in the keywords stored in the keyword table TBLa.
  • the object image OB1 shown in FIG. 5 is converted into “wine”
  • the object image OB2 is converted into “wine glass”
  • the object image OB3 is converted into “clock”
  • the object image OB4 is converted into “candle”
  • the object image OB5 is converted into “western food”.
  • the identifying unit 230A selects the target keyword Wx from the plurality of second keywords KW2 generated by the second keyword generating unit 220A based on the degree of association indicating the degree of association between the second keyword KW2 and the first keyword KW1. Identify. More specifically, the identifying unit 230A refers to the keyword table TBLa to determine the degree of association between the second keyword KW2 and the first keyword KW1 that indicates the degree of association between the second keyword KW2 and the first keyword KW1. Acquire for each group. The identifying unit 230A identifies the second keyword KW2 included in the set of the second keyword KW2 and the first keyword KW1 having the highest degree of association as the target keyword Wx. As described below, the comment generating unit 240 generates a comment related to the target keyword Wx specified by the specifying unit 230A.
  • the specifying unit 230A determines the degree of association between “food and drink” and “wine”, the degree of association between “food and drink” and “wine glass”, the degree of association between “food and drink” and “clock”, and “food and drink”.
  • the degree of association between “thing” and “candle” and the degree of association between “food and drink” and “western food” are acquired by referring to the keyword table TBLa.
  • the identifying unit 230A identifies the second keyword KW2 having the highest degree of association as the target keyword Wx by comparing the obtained plurality of degrees of association.
  • the comment generator 240 generates a comment related to the target keyword Wx.
  • the comment means an explanation or an explanation of the target keyword Wx.
  • a comment is a concept that includes recommendations. Therefore, the comment includes information about the product recommended to the user U and the store handling the product in relation to the target keyword Wx.
  • the comment generation unit 240 generates a comment by reading from the comment table TBLb the comment stored in association with the target keyword Wx. Further, the comment generation unit 240 may access a search site connected to the network NW, acquire information related to the target keyword Wx from the search site, and generate the acquired information as a comment. For example, when the target keyword Wx is “ramen”, the comment generation unit 240 may search for a ramen restaurant near the position information generated by the GPS device 27 or the like and output the search result as a comment.
  • FIG. 6 is a flowchart showing the operation of the user device 20.
  • the processing device 21 identifies the attention word based on the voice of the user U (step S1).
  • the processing device 21 extracts the attention word by performing a voice recognition process for converting the voice of the user U into text and a specifying process for specifying a noun and an adjective from the converted text.
  • the extraction word is a noun or an adjective if the noun is not specified.
  • the processing device 21 determines whether or not the focused word is included in the keyword table TBLa (step S2).
  • the processing device 21 returns the process to step S1 until the word of interest included in the keyword table TBLa is specified (that is, the determination result of step S2 is affirmative). Up to), the processes of steps S1 and S2 are repeated.
  • step S3 determines whether or not the focused word is a noun.
  • the processing device 21 determines whether or not the focused word is a noun (step S3).
  • the processing device 21 generates the attention word as the first keyword KW1.
  • the processing device 21 extracts a noun or an adjective as the attention word. Therefore, when the determination result of step S3 is negative, the attention word is the adjective.
  • the processing device 21 refers to the keyword table TBLa and generates the word of the noun associated with the focused word as the first keyword KW1 (step S5).
  • the processing device 21 extracts an object image from the image indicated by the image signal Sg (step S6).
  • a plurality of object images usually exist in one frame image. Therefore, the processing device 21 extracts a plurality of object images in the process of step S6. After that, the processing device 21 converts each of the plurality of extracted object images into the second keyword KW2 (step S7).
  • the processing device 21 identifies the target keyword Wx from the plurality of second keywords KW2 generated in step S7 based on the degree of association between the second keyword KW2 and the first keyword KW1.
  • the processing device 21 generates a comment related to the target keyword Wx (step S9).
  • the processing device 21 generates a comment by reading the comment stored in association with the target keyword Wx from the comment table TBLb.
  • the processing device 21 outputs the generated comment by one of the following methods.
  • the processing device 21 causes the display device 241 to display the moving image represented by the moving image data in which the generated comment image is superimposed.
  • the processing device 21 uses the short-range wireless communication device 26 to transmit the moving image data in which the generated comment image is superimposed to the television receiver 30.
  • the processing device 21 converts the generated comment into sound data, combines the sound data representing the comment with the sound data of the moving image, and causes the speaker 242 to emit the combined result.
  • the processing device 21 uses the short-range wireless communication device 26 to transmit the synthesis result of the sound data representing the comment and the sound data of the moving image to the television receiver 30.
  • the methods (a) to (d) may be combined arbitrarily.
  • the processing device 21 functions as the first keyword generation unit 210 in the processing of steps S1 to S5, functions as the extraction unit 221 in the processing of step S6, and functions as the conversion unit 222 in the processing of step S7. Further, the processing device 21 functions as the identifying unit 230A in the process of step S8, and functions as the comment generating unit 240 in the process of step S9.
  • the information processing apparatus which is an example of the user apparatus 20, includes the first keyword generation unit 210 that generates the first keyword KW1 based on the voice of the user U, and the plurality of extracted from the image indicated by the image signal Sg. Generated by the second keyword generation unit 220A based on the degree of association between the second keyword KW2 and the first keyword KW1 and the second keyword generation unit 220A that generates the second keyword KW2 for each of the object images. Also, a specifying unit 230A that specifies a target keyword Wx to be a comment target from the plurality of second keywords KW2 and a comment generating unit 240 that generates a comment related to the target keyword Wx are provided.
  • the comment can be generated in response to the vague statement of the user U without the user U designating the object image to be the comment target.
  • the specifying unit 230A specifies the second keyword KW2 that matches the first keyword KW1 as the target keyword Wx.
  • the identifying unit 230A determines whether each of the plurality of second keywords KW2 matches the first keyword KW1, and when the determination result regarding any of the plurality of second keywords KW2 is affirmative, the first keyword The second keyword KW2 that matches KW1 can be specified as the target keyword Wx. Therefore, it is not necessary to refer to the keyword table TBLa to acquire the degree of association, and the processing load can be reduced.
  • the service system 1 of the second embodiment is the same as the service system 1 of the first embodiment, except for the function of the processing device 21 in the user device 20.
  • FIG. 7 is a functional block diagram showing the functions of the processing device 21 of the second embodiment.
  • the processing device 21 of the second embodiment differs from the processing device 21 of the first embodiment in that a second keyword generation unit 220B is provided instead of the second keyword generation unit 220A.
  • the second keyword generation unit 220B includes an extraction unit 221, a conversion unit 222, and an analysis unit 223.
  • the image signal Sg is supplied to the analysis unit 223.
  • the analysis unit 223 analyzes the image signal Sg of the moving image and outputs the analysis result to the extraction unit 221.
  • the analysis unit 223, evaluates each of the plurality of object images included in an arbitrary frame of the image signal Sg using the first to fourth evaluation items, and the total of the evaluation values as the analysis result. It is output to the extraction unit 221. In this case, the extraction unit 221 extracts an object image whose total evaluation value exceeds a predetermined value.
  • the first evaluation item is the ratio of the area of the object image to the area of one screen, and the larger the ratio of the object image, the higher the evaluation value of the object image.
  • the second evaluation item is the perspective of the object image viewed from the user U, and the evaluation value of the object image is higher as the object image is closer to the front.
  • the third evaluation item is the brightness of the object image, and the brighter the brightness of the object image, the higher the evaluation value of the object image.
  • the fourth evaluation item is the position of the object image, and the closer the position of the object image is to the center of the screen, the higher the evaluation value of the object image.
  • Each of the first to fourth evaluation items is an element that attracts the interest of the user U in the image of one screen.
  • FIG. 8 is a flowchart showing the operation of the user device 20 according to the second embodiment.
  • the flowchart shown in the figure is the same as the flowchart of the first embodiment shown in FIG. 6 except that steps S6_1 and S6_2 are executed instead of step S6. The differences will be described below.
  • step S6_1 the processing device 21 functions as the analysis unit 223, acquires a plurality of evaluation values obtained by evaluating each of a plurality of object images included in a frame for each evaluation item, and calculates the sum of these evaluation values.
  • the analysis results of the object images OB1 to OB5 shown in FIG. 5 are as shown in FIG.
  • the sum of the evaluation values for each of the object images OB1 to OB5 is in the range of “11” to “16”.
  • step S6_2 the processing device 21 functions as the extraction unit 221 and extracts an object image whose total evaluation value exceeds a predetermined value. For example, it is assumed that the predetermined value is “13” and that the sum of the evaluation values shown in FIG. 9 is obtained for each object image. In this case, the processing device 21 extracts the object images OB2 and OB5. Note that the processing from step S7 onward is the same as the processing described in the first embodiment with reference to FIG. 6, so description will be omitted.
  • the second keyword generating unit 220B analyzes the image signal Sg from the image indicated by the image signal Sg based on the analysis unit 223 that analyzes the image signal Sg and the analysis result of the analyzing unit 223.
  • An extraction unit 221 that extracts a plurality of object images, and a conversion unit 222 that converts each of the plurality of object images into a second keyword KW2 are provided.
  • the number of object images to be extracted can be reduced as compared with the case where the object images are extracted without using the analysis result. You can Therefore, the processing load of the conversion unit 222 can be reduced.
  • the analysis unit 223 may analyze the image signal Sg over a plurality of frames and generate an analysis result.
  • the analysis unit 223 may adopt the fifth evaluation item regarding the movement of the object image in addition to the first evaluation item to the fourth evaluation item.
  • An example of the fifth evaluation item is the number of frames corresponding to the length of time that a moving object image exists in the screen. According to this evaluation item, the larger the number of frames is (the object image exists in the screen The evaluation value of the object image is higher (the longer the time is spent). For example, when the hero of a movie moves, the movie is often shot so as to follow the movement of the hero.
  • the moving image represented by the image signal Sg is a movie
  • the evaluation value of the object image representing the protagonist and the evaluation value of the object image representing the belongings possessed by the protagonist are increased.
  • the service system 1 of the third embodiment is the same as the service system 1 of the first embodiment, except for the function of the processing device 21 in the user device 20 and the stored contents of the storage device 22.
  • FIG. 10 is a functional block diagram showing the functions of the processing device 21 of the third embodiment.
  • the processing device 21 of the third embodiment is different from the processing device 21 of the first embodiment in that it includes a specifying unit 230B instead of the specifying unit 230A.
  • the storage device 22 of the user device 20 of the third embodiment stores the action history table TBLc.
  • the behavior history table TBLc stores the behavior history of the user U.
  • the action history includes the Internet search history of the user U, the purchase history of products and services, the activity of SNS (Social Networking Service), the bookmark of the Web browser, and the like.
  • the identifying unit 230B generates a plurality of second keywords generated by the second keyword generating unit 220A, based on the degree of association between each second keyword KW2 and the first keyword KW1 and the behavior history of the user U.
  • the target keyword Wx is specified from KW2.
  • the identifying unit 230B selects, from the plurality of second keywords KW2 generated by the second keyword generating unit 220A, the second keyword KW2 having a degree of association with the first keyword KW1 that is equal to or greater than a predetermined value.
  • the selected second keyword KW2 becomes a candidate for the target keyword Wx.
  • the identifying unit 230B identifies the target keyword Wx from the selected second keywords KW2 with reference to the action history stored in the action history table TBLc. For example, assume that the second keyword KW2 selected based on the degree of association is “wine” and “western food”. Further, it is assumed that wine purchase history is recorded in the action history table TBLc.
  • the specifying unit 230B detects that the user U has a purchase history of wine by referring to the action history table TBLc, it specifies “wine” as the target keyword Wx of “wine” and “Western food”. To do. As a result, the comment generator 240 can generate a comment regarding “wine”.
  • FIG. 11 is a flowchart showing the operation of the user device 20 according to the third embodiment.
  • the flowchart shown in the figure is the same as the flowchart of the first embodiment shown in FIG. 6 except that steps S8_1 and S8_2 are executed instead of step S8. The differences will be described below.
  • step S8_1 the processing device 21 functions as the identifying unit 230B, and selects the second keyword KW2 having a degree of association with the first keyword KW1 that is equal to or greater than a predetermined value from the plurality of second keywords KW2 generated in step S7. To do.
  • step S8_2 the processing device 21 functions as the identifying unit 230B, and the second keyword KW2 related to the action history among the second keywords KW2 selected in the process of step S8_1 is set as the target keyword Wx based on the action history. Identify.
  • the identifying unit 230B identifies the target keyword Wx from the plurality of second keywords KW2 based on the degree of association indicating the degree of association and the behavior history of the user U. According to this aspect, since the target keyword Wx is specified in consideration of the behavior history of the user U, it is possible to provide a comment of high interest to the user U, as compared with the case where the behavior history of the user U is not considered. ..
  • the identifying unit 230B is a candidate for the target keyword Wx among the plurality of second keywords KW2 generated by the second keyword generating unit 220A using the degree of association. Then, the second keyword KW2 is selected (step S8_1), and then the target keyword Wx is specified based on the action history (step S8_2), but the order may be reversed. That is, the identification unit 230B selects the second keyword KW2 that is a candidate for the target keyword Wx from the plurality of second keywords KW2 generated by the second keyword generation unit 220A based on the action history, and then uses the degree of association. Alternatively, the target keyword Wx may be specified.
  • the identifying unit 230B may identify the target keyword Wx from the plurality of second keywords KW2 by using the action history and the degree of association at the same time.
  • the specifying unit 230B adds, for example, a predetermined value to the degree of association for the second keyword KW2 related to the action history, and compares the degree of association to which the predetermined value is added between the plurality of second keywords KW2 to determine the target keyword Wx. May be specified.
  • the service system 1 of the fourth embodiment is the same as the service system 1 of the first embodiment, except for the function of the processing device 21 in the user device 20 and the stored contents of the storage device 22.
  • FIG. 12 is a functional block diagram showing the functions of the processing device 21 of the fourth embodiment.
  • the processing device 21 of the fourth embodiment is different from the processing device 21 of the first embodiment in that a specifying unit 230C is provided instead of the specifying unit 230A.
  • the storage device 22 of the user device 20 of the fourth embodiment stores the profile data DP and the evaluation table TBLd.
  • the profile data DP indicates the profile of the user U.
  • the profile means the attribute of the user U, and includes items such as age and sex.
  • the evaluation value for each profile item is stored in association with the keyword.
  • the evaluation value is a value indicating the degree of interest of the user U with respect to the keyword.
  • FIG. 13 shows an example of the stored contents of the evaluation table TBLd. For example, for the keyword “car”, the evaluation value for gender “male” is “7”, whereas the evaluation value for gender “female” is “4”. This shows that men are more interested in cars than women.
  • the identifying unit 230C uses the plurality of second keywords KW2 generated by the second keyword generating unit 220A based on the degree of association between each second keyword KW2 and the first keyword KW1 and the profile of the user U.
  • the target keyword Wx is specified from among these.
  • the identifying unit 230C selects the second keyword KW2 having a degree of association with the first keyword KW1 that is equal to or more than a predetermined value, from among the plurality of second keywords KW2 generated by the second keyword generating unit 220A.
  • the specifying unit 230C specifies the target keyword Wx from the selected second keywords KW2 using the profile data DP and the evaluation table TBLd. Specifically, for each of the selected second keywords KW2, a total evaluation value obtained by summing evaluation values corresponding to a plurality of items of the profile of the user U is calculated, and the second keyword KW2 having the highest total evaluation value is calculated. Is specified as the target keyword Wx.
  • FIG. 14 is a flowchart showing the operation of the user device 20 according to the fourth embodiment.
  • the flowchart shown in the figure is the same as the flowchart of the third embodiment shown in FIG. 12 except that step S8_3 is executed instead of step S8_2. The differences will be described below.
  • step S8_3 the processing device 21 functions as the specifying unit 230C, and the second keyword KW2 having the highest total evaluation value of the profile of the user U among the second keywords KW2 selected in the process of step S8_1 based on the profile. Is specified as the target keyword Wx.
  • the identifying unit 230C identifies the target keyword Wx from the plurality of second keywords KW2 based on the degree of association indicating the degree of association and the profile of the user U. According to this aspect, since the target keyword Wx is specified in consideration of the profile of the user U, it is possible to provide a comment of high interest to the user U, as compared with the case where the profile of the user U is not considered.
  • the identifying unit 230B determines that the target keyword Wx is a candidate of the target keyword Wx among the plurality of second keywords KW2 generated by the second keyword generating unit 220A using the degree of association.
  • the second keyword KW2 is selected (step S8_1), and then the target keyword Wx is specified based on the profile (step S8_3), but the order may be reversed. That is, the specifying unit 230C selects the second keyword KW2 that is a candidate for the target keyword Wx from the plurality of second keywords KW2 generated by the second keyword generating unit 220A based on the profile, and then uses the degree of association.
  • the target keyword Wx may be specified.
  • the specifying unit 230C may specify the target keyword Wx from the plurality of second keywords KW2 by using the profile and the degree of association at the same time.
  • the specifying unit 230C may add the total evaluation value based on the profile to the degree of association, and compare the addition results of the plurality of second keywords KW2 to specify the target keyword Wx.
  • the frame in which the extraction unit 221 extracts the object image from the image of the image signal Sg may be the following frame.
  • the extraction unit 221 may extract the object image in a frame having a high audience rating.
  • the extraction unit 221 may acquire the audience rating from the external device in real time.
  • the extraction unit 221 extracts the object image in a frame in which the acquired audience rating exceeds a predetermined audience rating. It is estimated that the user U has a higher interest in a frame having a higher audience rating than other frames. Therefore, since the object image is extracted from the image of the frame in which the user U has a high interest, a comment useful for the user U can be generated.
  • the extraction unit 221 may extract the object image in a frame in which the user U cheers based on the audio signal Sa of the user U.
  • the extraction unit 221 may extract the object image in a frame that is the subject of the program based on the program information.
  • the extraction unit 221 may use the analysis unit 223 described in the second embodiment to analyze the image signal Sg and identify the frame that is the subject of the program.
  • the analysis unit 223 may acquire the program information from the external device via the network NW.
  • the image signal Sg has been described as a signal indicating a moving image, but the image signal Sg may be a signal indicating a still image.
  • the degree of association indicating the degree of association between the second keyword KW2 and the first keyword KW1 is stored in the keyword table TBLa, but the present invention is not limited to this.
  • the identifying units 230A, 230B, and 230C obtain the degree of association according to the number of nodes identified from the keyword table TBLa (an example of keyword data) having a tree structure in which a plurality of words are layered according to meaning. Good.
  • the first keyword generation unit 210 generates a word included in the keyword table TBLa as the first keyword KW1.
  • the second keyword generation units 220A and 220B generate words included in the keyword table TBLa as the second keyword KW2.
  • the specifying units 230A, 230B, and 230C acquire the number of nodes in the route from the first keyword KW1 to the second keyword KW2 in the tree structure of the keyword table TBLa as the degree of association. More specifically, it is assumed that the data structure of the keyword table TBLa is the tree structure shown in FIG. For example, when the first keyword KW1 is “liquor” and the second keyword KW2 is “french potato”, the route from “liquor” to “french potato” is node “liquor” ⁇ node “drink” ⁇ node “Food and drink” ⁇ node “food” ⁇ node “Western food” ⁇ node “fried potato”.
  • the number of nodes on the route from the first keyword KW1 "liquor” to the second keyword KW2 "fried potatoes” is "5".
  • the route from “food and drink” to “fries” is node “food and drink” ⁇ node “food” -> Node “Western food”-> node “fries”. Therefore, the number of nodes on the route from the first keyword KW1 "food and drink” to the second keyword KW2 "fried potatoes” is "3". The smaller the number of nodes on the route connecting the first keyword KW1 and the second keyword KW2, the higher the degree of association.
  • the degree of association is higher than the degree of association between the first keyword KW1 “sake” and the second keyword KW2 “fried potato”.
  • the extraction unit 221 extracts the object image without considering the action history of the user U, but extracts the object image from the image indicated by the image signal Sg based on the action history. Good.
  • the extraction unit 221 refers to the action history table TBLc described in the third embodiment, identifies a favorite color of the user U, for example, from the purchase history of the product, and extracts the object image of the identified color. May be. According to this modification, the object images can be narrowed down, so that the processing load of the conversion unit 222 can be reduced.
  • the second keyword generation unit 220B of the second embodiment may be used instead of the second keyword generation unit 220A of the third and fourth embodiments.
  • each functional block may be realized by one device that is physically and / or logically coupled, or may be directly and / or indirectly connected to two or more devices that are physically and / or logically separated. (For example, wired and / or wireless), and may be realized by a plurality of these devices.
  • the function of the conversion unit 222 may be provided from a server device connected via the network NW.
  • the keyword table TBLa may be provided in the server device.
  • the wording “apparatus” used in the description of each of the above-described embodiments may be replaced with another term such as a circuit, a device, or a unit.
  • the input / output information and the like may be stored in a specific place (for example, a memory) or may be managed by a management table. Information that is input / output may be overwritten, updated, or added. The output information and the like may be deleted. The input information and the like may be transmitted to another device.
  • the determination may be performed by a value (0 or 1) represented by 1 bit, or by a true / false value (Boolean: true or false).
  • the comparison may be performed by comparing numerical values (for example, comparison with a predetermined value).
  • the storage device 22 is a recording medium that can be read by the processing device 21, and is exemplified by a ROM and a RAM, but a flexible disk, a magneto-optical disk (for example, a compact disk, a digital multi-disk). Purpose disk, Blu-ray (registered trademark) disk), smart card, flash memory device (for example, card, stick, key drive), CD-ROM (Compact Disc-ROM), register, removable disk, hard disk, floppy (registered Trademark disks, magnetic strips, databases, servers and other suitable storage media.
  • the program may be transmitted from the network NW. Further, the program may be transmitted from a communication network via an electric communication line.
  • Each of the above-described embodiments includes LTE (Long Term Evolution), LTE-A (LTE-Advanced), SUPER 3G, IMT-Advanced, 4G, 5G, FRA (Future Radio Access), W-CDMA (registered trademark) ), GSM (registered trademark), CDMA2000, UMB (Ultra Mobile Broadband), IEEE 802.11 (Wi-Fi), IEEE 802.11 (WiMAX), IEEE 802.20, UWB (Ultra-WideBand), Bluetooth (registered). (Trademark), other systems utilizing appropriate systems, and / or next-generation systems extended based on these.
  • the information, signals, etc. described may be represented using any of a variety of different technologies.
  • data, instructions, commands, information, signals, bits, symbols, chips, etc. that may be mentioned throughout the above description are voltage, current, electromagnetic waves, magnetic fields or magnetic particles, optical fields or photons, or any of these. May be represented by a combination of Note that the terms described in this specification and / or the terms necessary for understanding this specification may be replaced with terms having the same or similar meanings.
  • Each function illustrated in FIGS. 4, 7, 10, and 12 is realized by an arbitrary combination of hardware and software. Further, each function may be realized by a single device, or may be realized by two or more devices that are configured separately from each other.
  • the program illustrated in each of the above-described embodiments is an instruction, an instruction set, a code, or a code segment regardless of whether it is called software, firmware, middleware, microcode, a hardware description language, or another name.
  • Program code, subprograms, software modules, applications, software applications, software packages, routines, subroutines, objects, executable files, threads of execution, procedures or functions, etc. should be construed broadly. Further, software, instructions, etc. may be transmitted and received via a transmission medium.
  • the software may use a wired technology such as coaxial cable, fiber optic cable, twisted pair and digital subscriber line (DSL) and / or wireless technology such as infrared, wireless and microwave to websites, servers, or other When transmitted from a remote source, these wireline and / or wireless technologies are included within the definition of transmission medium.
  • a wired technology such as coaxial cable, fiber optic cable, twisted pair and digital subscriber line (DSL) and / or wireless technology such as infrared, wireless and microwave to websites, servers, or other
  • information, parameters, and the like may be represented by an absolute value, a relative value from a predetermined value, or another corresponding information. Good.
  • a mobile station can be a subscriber station, mobile unit, subscriber unit, wireless unit, remote unit, mobile device, wireless device, wireless communication device, remote device, mobile subscriber station, access terminal, mobile terminal, wireless, by a person skilled in the art. It may also be referred to as a terminal, remote terminal, handset, user agent, mobile client, client, or some other suitable term.
  • connection means any direct or indirect connection or coupling between two or more elements, It can include the presence of one or more intermediate elements between two elements that are “connected” to each other.
  • the connections between the elements may be physical, logical, or a combination thereof.
  • two elements are radio frequency by using one or more wires, cables and / or printed electrical connections, and as some non-limiting and non-exhaustive examples.
  • electromagnetic energy such as electromagnetic energy having wavelengths in the region, the microwave region and the light (both visible and invisible) region, it can be considered to be “connected” to each other.
  • any reference to elements using the designations "first”, “second”, etc. does not generally limit the amount or order of those elements. These designations may be used herein as a convenient way to distinguish between two or more elements. Thus, references to the first and second elements do not imply that only two elements may be employed therein, or that the first element must precede the second element in any way.

Abstract

Provided is a user device comprising: a first keyword generation unit which generates a first keyword on the basis of the voice of a user; a second keyword generation unit which generates a plurality of second keywords corresponding one-on-one to a plurality of object images extracted from an image indicated by an image signal; an identifying unit which, on the basis of the level of relevance between each of the plurality of second keywords and the first keyword, identifies a target keyword as a target for a comment from among the plurality of second keywords generated by the second keyword generation unit; and a comment generation unit which generates a comment related to the target keyword.

Description

情報処理装置Information processing equipment
 本発明は、情報処理装置に関する。 The present invention relates to an information processing device.
 特許文献1には、ユーザがポインティングデバイスを用いて表示装置に表示される画像内のオブジェクト画像を指定すると、当該オブジェクト画像に関するレコメンドを表示する技術が開示されている。また、特許文献2にはユーザの音声に応答して当該ユーザが音声で指定したオブジェクト画像に関する情報をレコメンドする技術が開示されている。 Patent Document 1 discloses a technique in which when a user specifies an object image in an image displayed on a display device using a pointing device, a recommendation regarding the object image is displayed. Further, Patent Document 2 discloses a technique of recommending information about an object image designated by a voice of the user in response to the voice of the user.
特開2017-228177号公報JP, 2017-228177, A 特開2013-88906号公報JP, 2013-88906, A
 しかしながら、従来の技術では、ユーザがオブジェクト画像を指定する必要がある。すなわち、オブジェクト画像をユーザが指定しない場合に、ユーザの曖昧な発言に応答して、レコメンドなどのコメントを生成することはできなかった。 However, in the conventional technology, the user needs to specify the object image. That is, when the user does not specify the object image, it is not possible to generate a comment such as a recommendation in response to the user's ambiguous statement.
 以上の課題を解決するために、本発明の好適な態様に係る情報処理装置は、ユーザの音声に基づいて第1キーワードを生成する第1キーワード生成部と、画像信号の示す画像から抽出した複数のオブジェクト画像に1対1で対応する複数の第2キーワードを生成する第2キーワード生成部と、前記複数の第2キーワードの各々と前記第1キーワードとの関連性の程度に基づいて、前記複数の第2キーワードの中からコメントの対象となる対象キーワードを特定する特定部と、前記対象キーワードに関連するコメントを生成するコメント生成部とを備える。 In order to solve the above problems, an information processing apparatus according to a preferred aspect of the present invention includes a first keyword generating unit that generates a first keyword based on a voice of a user, and a plurality of units extracted from an image indicated by an image signal. Second keyword generation unit that generates a plurality of second keywords corresponding to the object image of 1: 1, and the plurality of second keywords based on the degree of association between each of the plurality of second keywords and the first keyword. And a comment generation unit that generates a comment related to the target keyword.
 本発明に係る情報処理装置によれば、ユーザがコメントの対象となるオブジェクト画像を指定することなく、ユーザの曖昧な発言に応答して、コメントを生成することができる。 According to the information processing apparatus of the present invention, a comment can be generated in response to a user's ambiguous statement without the user designating an object image to be a comment target.
本発明の第1実施形態に係るサービスシステムの全体構成を示すブロック図である。It is a block diagram showing the whole service system composition concerning a 1st embodiment of the present invention. 同実施形態に用いられるユーザ装置のハードウェア構成を例示するブロック図である。It is a block diagram which illustrates the hardware constitutions of the user apparatus used for the same embodiment. 同実施形態に用いられるキーワードテーブルのデータ構造を示す説明図である。It is explanatory drawing which shows the data structure of the keyword table used for the same embodiment. 同実施形態に用いられるユーザ装置の機能を示す機能ブロック図である。It is a functional block diagram which shows the function of the user apparatus used for the same embodiment. 同実施形態のオブジェクト画像の一例を示す説明図である。It is explanatory drawing which shows an example of the object image of the same embodiment. 同実施形態に用いられるユーザ装置の動作を示すフローチャートである。It is a flowchart which shows operation | movement of the user apparatus used for the same embodiment. 第2実施形態に用いられるユーザ装置の機能を示す機能ブロック図である。It is a functional block diagram which shows the function of the user apparatus used for 2nd Embodiment. 同実施形態に用いられるユーザ装置の動作を示すフローチャートである。It is a flowchart which shows operation | movement of the user apparatus used for the same embodiment. 同実施形態におけるオブジェクト画像の評価結果を説明するための説明図である。It is an explanatory view for explaining an evaluation result of an object image in the embodiment. 第3実施形態に用いられるユーザ装置の機能を示す機能ブロック図である。It is a functional block diagram which shows the function of the user apparatus used for 3rd Embodiment. 同実施形態に用いられるユーザ装置の動作を示すフローチャートである。It is a flowchart which shows operation | movement of the user apparatus used for the same embodiment. 第4実施形態に用いられるユーザ装置の機能を示す機能ブロック図である。It is a functional block diagram which shows the function of the user apparatus used for 4th Embodiment. 同実施形態に用いられる評価テーブルの記憶内容を示す説明図である。It is explanatory drawing which shows the memory content of the evaluation table used for the same embodiment. 同実施形態に用いられるユーザ装置の動作を示すフローチャートである。It is a flowchart which shows operation | movement of the user apparatus used for the same embodiment.
[1.第1実施形態]
[1.1.サービスシステムの構成]
 図1は、本発明の第1実施形態に係るサービスシステムの全体構成を示すブロック図である。図1に示されるサービスシステム1は、動画の配信サービスを提供する。動画の配信サービスは、例えば、映画又は地上波デジタル放送のコンテンツなどを提供する。
[1. First Embodiment]
[1.1. Service system configuration]
FIG. 1 is a block diagram showing the overall configuration of a service system according to the first embodiment of the present invention. The service system 1 shown in FIG. 1 provides a moving image distribution service. The moving image distribution service provides, for example, movies or terrestrial digital broadcasting contents.
 図1に例示するように、サービスシステム1は、ユーザU_1~ユーザU_mが管理するユーザ装置20_1~20_m(mは1以上の整数)と、ネットワークNWと、動画配信サーバ10とを備える。以下の説明では、同種の要素を区別しない場合には、ユーザ装置20又はユーザUのように、参照符号のうちの共通番号だけを使用する。 As illustrated in FIG. 1, the service system 1 includes user devices 20_1 to 20_m (m is an integer of 1 or more) managed by users U_1 to U_m, a network NW, and a video distribution server 10. In the following description, when the same type of element is not distinguished, only the common number among the reference numerals is used like the user device 20 or the user U.
 ユーザ装置20は、各種の情報を処理する情報処理装置である。ユーザ装置20は、例えば、スマートフォン又はタブレット端末等の可搬型の情報処理装置である。但し、ユーザ装置20としては、任意の情報処理装置を採用することができる。ユーザ装置20は、例えば、パーソナルコンピュータ等の端末型の情報機器であってもよい。 The user device 20 is an information processing device that processes various types of information. The user device 20 is, for example, a portable information processing device such as a smartphone or a tablet terminal. However, as the user device 20, any information processing device can be adopted. The user device 20 may be, for example, a terminal-type information device such as a personal computer.
 ユーザ装置20は、動画配信サーバ10から送信される画像信号Sgを受信して画像を表示したり、あるいは、画像信号Sgをテレビジョン受像機30に送信してテレビジョン受像機30に画像を表示させることができる。
 ユーザUは、動画を見ながら発言することがある。例えば、ユーザUは動画についての感想を述べたり、つぶやくことがある。この場合、ユーザUの発言は動画に関連するものではあるが、当該発言が曖昧であることが理由で発言が動画の画像に含まれるどの物体に関連するものであるかを一意に特定できないことが多い。ユーザ装置20は、ユーザUの曖昧な発言に応答してレコメンドなどのコメントを生成する機能を有する。
The user device 20 receives the image signal Sg transmitted from the moving image distribution server 10 and displays the image, or transmits the image signal Sg to the television receiver 30 and displays the image on the television receiver 30. Can be made
The user U may speak while watching a moving image. For example, the user U may make a comment or tweet about the moving image. In this case, although the statement of the user U is related to the moving image, it is not possible to uniquely specify which object included in the image of the moving picture the statement is related to because the statement is ambiguous. There are many. The user device 20 has a function of generating a comment such as a recommendation in response to the vague statement of the user U.
[1.2.ユーザ装置の構成]
 図2は、ユーザ装置20のハードウェア構成を例示するブロック図である。ユーザ装置20は、処理装置21、記憶装置22、通信装置23、出力装置24、入力装置25、近距離無線通信装置26、GPS(Global Positioning System)装置27、及びバス28を具備するコンピュータシステムにより実現される。処理装置21、記憶装置22、通信装置23、出力装置24、入力装置25、近距離無線通信装置26及びGPS装置27は、情報を通信するためのバス28で接続される。バス28は、単一のバスで構成されてもよいし、装置間で異なるバスで構成されてもよい。なお、ユーザ装置20の各要素は、単数又は複数の機器で構成され、ユーザ装置20の一部の要素を省略してもよい。
[1.2. Configuration of user device]
FIG. 2 is a block diagram illustrating a hardware configuration of the user device 20. The user device 20 is a computer system including a processing device 21, a storage device 22, a communication device 23, an output device 24, an input device 25, a short-range wireless communication device 26, a GPS (Global Positioning System) device 27, and a bus 28. Will be realized. The processing device 21, the storage device 22, the communication device 23, the output device 24, the input device 25, the short-range wireless communication device 26, and the GPS device 27 are connected by a bus 28 for communicating information. The bus 28 may be configured by a single bus or may be configured by different buses among devices. Note that each element of the user device 20 is configured by a single device or a plurality of devices, and some elements of the user device 20 may be omitted.
 処理装置21は、ユーザ装置20の全体を制御するプロセッサであり、例えば単数又は複数のチップで構成される。処理装置21は、例えば、周辺装置とのインタフェース、演算装置及びレジスタ等を含む中央処理装置(CPU:Central Processing Unit)で構成される。なお、処理装置21の機能の一部又は全部を、DSP(Digital Signal Processor)、ASIC(Application Specific Integrated Circuit)、PLD(Programmable Logic Device)、FPGA(Field Programmable Gate Array)等のハードウェアで実現してもよい。処理装置21は、各種の処理を並列的又は逐次的に実行する。 The processing device 21 is a processor that controls the entire user device 20, and is composed of, for example, a single chip or a plurality of chips. The processing device 21 is configured by, for example, a central processing unit (CPU) including an interface with peripheral devices, an arithmetic device, a register, and the like. Note that some or all of the functions of the processing device 21 are realized by hardware such as DSP (Digital Signal Processor), ASIC (Application Specific Integrated Circuit), PLD (Programmable Logic Device), and FPGA (Field Programmable Gate Array). May be. The processing device 21 executes various processes in parallel or sequentially.
 記憶装置22は、処理装置21が読取可能な記録媒体である。記憶装置22は、処理装置21が実行する制御プログラムPRaを含む複数のプログラム、キーワードテーブルTBLa、コメントテーブルTBLb及び処理装置21が使用する各種のデータを記憶する。記憶装置22は、例えば、ROM(Read Only Memory)、EPROM(Erasable Programmable ROM)、EEPROM(Electrically Erasable Programmable ROM)、及びRAM(Random Access Memory)等の記憶回路の1種類以上で構成される。 The storage device 22 is a recording medium that can be read by the processing device 21. The storage device 22 stores a plurality of programs including the control program PRa executed by the processing device 21, a keyword table TBLa, a comment table TBLb, and various data used by the processing device 21. The storage device 22 is composed of one or more types of storage circuits such as a ROM (Read Only Memory), an EPROM (Erasable Programmable ROM), an EEPROM (Electrically Erasable Programmable ROM), and a RAM (Random Access Memory).
 キーワードテーブルTBLaには、複数の単語が記憶されている。複数の単語は、名詞と形容詞に大別される。名詞の単語はキーワードに対応する。後述する第1キーワードKW1と第2キーワードKW2とは、キーワードテーブルTBLaに記憶されている名詞の単語に含まれる。また、形容詞の単語は、名詞の単語に対応付けて記憶されている。形容詞は名詞を修飾する機能がある。形容詞の単語と名詞の単語との対応付けは、単語の修飾関係に応じて定められている。例えば、形容詞の単語である「美味しい」は、名詞の単語である「飲食物」に対応付けられている。 A plurality of words are stored in the keyword table TBLa. Multiple words are roughly divided into nouns and adjectives. Noun words correspond to keywords. The first keyword KW1 and the second keyword KW2 described later are included in the noun words stored in the keyword table TBLa. The adjective word is stored in association with the noun word. Adjectives have the function of modifying nouns. The association between the adjective word and the noun word is determined according to the word modification relation. For example, the adjective word “delicious” is associated with the noun word “food and drink”.
 図3は、キーワードテーブルTBLaに記憶される名詞の単語のデータ構造を示す説明図である。同図に示されるように、名詞の単語のデータ構造は、複数の単語が意味によって階層化された木構造となっている。この例では、複数の単語が第1階層から第4階層に分類されている。なお、階層数は、4以上であってもよい。 FIG. 3 is an explanatory diagram showing a data structure of a noun word stored in the keyword table TBLa. As shown in the figure, the data structure of the noun word has a tree structure in which a plurality of words are hierarchized according to the meaning. In this example, a plurality of words are classified into the first hierarchy to the fourth hierarchy. The number of layers may be four or more.
 また、キーワードテーブルTBLaには、名詞の単語と名詞の単語との関連性の程度を示す関連度が記憶されている。関連性の程度は、上位概念と下位概念の関係の他、単語の示す物体の用途及び機能を考慮して定められる。例えば、「日本酒」と「ワイン」とは、いずれも「酒」の下位概念である。これに対して、「お猪口」は「酒」の下位概念ではないが、「お猪口」は「日本酒」を飲むために用いられる。このため、「日本酒」と「お猪口」との関連度は、「日本酒」と「ワイン」との関連度より高くなっている。 Further, the keyword table TBLa stores the degree of association indicating the degree of association between the noun word and the noun word. The degree of relevance is determined in consideration of the relationship between the superordinate concept and the subordinate concept, and the use and function of the object indicated by the word. For example, “sake” and “wine” are both subordinate concepts of “sake”. On the other hand, although "Ouchiguchi" is not a subordinate concept of "Sake", "Inoguchi" is used to drink "Japanese sake". For this reason, the degree of association between “sake” and “chef” is higher than the degree of association between “sake” and “wine”.
 説明を図2に戻す。通信装置23は、移動体通信網又はインターネット等のネットワークNWを介して他の装置と通信する機器である。通信装置23は、例えばネットワークデバイス、ネットワークコントローラ、ネットワークカード又は通信モジュールとも表記される。通信装置23は、ネットワークNWを介して、動画配信サーバ10と通信可能である。 Return the explanation to Figure 2. The communication device 23 is a device that communicates with another device via a network NW such as a mobile communication network or the Internet. The communication device 23 is also described as, for example, a network device, a network controller, a network card, or a communication module. The communication device 23 can communicate with the moving image distribution server 10 via the network NW.
 出力装置24は、処理装置21による制御のもとで各種の情報をユーザUに知らせる。出力装置24は、表示装置241とスピーカ242とを備える。表示装置241は、画像を表示する。例えば液晶表示パネル、又は有機EL(Electro Luminescence)表示パネル等の各種の表示パネルが表示装置241として好適に利用される。
 スピーカ242には、処理装置21から音データが供給される。スピーカ242はDA変換器を備える。DA変換器によって音データはアナログ信号に変換され、アナログ信号によってスピーカ242は駆動される。
The output device 24 informs the user U of various information under the control of the processing device 21. The output device 24 includes a display device 241 and a speaker 242. The display device 241 displays an image. For example, various display panels such as a liquid crystal display panel or an organic EL (Electro Luminescence) display panel are preferably used as the display device 241.
Sound data is supplied from the processing device 21 to the speaker 242. The speaker 242 includes a DA converter. The sound data is converted into an analog signal by the DA converter, and the speaker 242 is driven by the analog signal.
 入力装置25は、ユーザ装置20を使用するための情報をユーザUが入力するための機器である。入力装置25は、ユーザUによる入力操作を受け付ける。この例の入力装置25は、マイクロフォン251及びタッチパネル252を備える。タッチパネル252は、表示装置241の表示面に対するユーザUによる接触を検出する。タッチパネル252は、接触位置に基づいて、数字及び文字等の符号を入力する操作と、表示装置241が表示するアイコンを選択する操作とを受け付ける。マイクロフォン251は、ユーザUの音声をアナログの電気信号に変換し、当該電気信号を音声信号Saとして出力する。音声信号Saは図示せぬAD変換部によりデジタル信号に変換されバス28を介して処理装置21に供給される。 The input device 25 is a device for the user U to input information for using the user device 20. The input device 25 receives an input operation by the user U. The input device 25 of this example includes a microphone 251 and a touch panel 252. The touch panel 252 detects a touch by the user U on the display surface of the display device 241. The touch panel 252 accepts an operation of inputting codes such as numbers and letters and an operation of selecting an icon displayed on the display device 241 based on the contact position. The microphone 251 converts the voice of the user U into an analog electric signal and outputs the electric signal as a sound signal Sa. The audio signal Sa is converted into a digital signal by an AD converter (not shown) and supplied to the processing device 21 via the bus 28.
 近距離無線通信装置26は、近距離無線通信によって他の装置と通信する機器である。近距離無線通信には、例えばBluetooth(登録商標)、ZigBee(登録商標)、又は、WiFi(登録商標)等が挙げられる。他の装置としては、テレビジョン受像機30等が該当する。
 GPS装置27は複数の衛星からの電波を受信し、受信した電波から位置情報を生成する。位置情報は、ユーザ装置20の位置を示す。位置情報は、位置を特定できるのであれば、どのような形式であってもよい。位置情報は、例えば、ユーザ装置20の緯度と経度とを示す。本実施形態では、位置情報はGPS装置27から得られることを例示するが、ユーザ装置20は、他の任意の方法で位置情報を取得してもよい。例えば、ユーザ装置20は、ユーザ装置20の通信先である基地局に割り当てられたセルIDを用いて位置情報を取得してもよい。あるいは、ユーザ装置20が近距離無線通信装置26を用いて無線LAN(Local Area Network)のアクセスポイントと通信する場合には、ユーザ装置20は、アクセスポイントに割り当てられたネットワーク上の識別アドレス(MAC(Media Access Control)アドレス)と実際の住所(位置)とを互いに対応付けたデータベースを参照して位置情報を取得してもよい。あるいは、ユーザ装置20は、近距離無線通信装置26を用いてBLE(Bluetooth Low Energy) 規格に準拠したアドバタイズメント・パケットに含まれるID情報を受信し、当該ID情報に基づいて位置情報を取得してもよい。
The short-range wireless communication device 26 is a device that communicates with another device by short-range wireless communication. Examples of short-range wireless communication include Bluetooth (registered trademark), ZigBee (registered trademark), and WiFi (registered trademark). The television receiver 30 or the like corresponds to another device.
The GPS device 27 receives radio waves from a plurality of satellites and generates position information from the received radio waves. The position information indicates the position of the user device 20. The position information may have any format as long as the position can be specified. The position information indicates, for example, the latitude and longitude of the user device 20. In the present embodiment, the position information is illustrated as being obtained from the GPS device 27, but the user device 20 may acquire the position information by any other method. For example, the user device 20 may acquire the location information by using the cell ID assigned to the base station that is the communication destination of the user device 20. Alternatively, when the user device 20 communicates with an access point of a wireless LAN (Local Area Network) using the short-range wireless communication device 26, the user device 20 uses the identification address (MAC) on the network assigned to the access point. The position information may be acquired by referring to a database in which (Media Access Control) addresses) and actual addresses (positions) are associated with each other. Alternatively, the user device 20 receives the ID information included in the advertisement packet conforming to the BLE (Bluetooth Low Energy) standard by using the short-range wireless communication device 26, and acquires the position information based on the ID information. May be.
[1.3.ユーザ装置20の機能]
 図4は、ユーザ装置20の機能を示す機能ブロック図である。処理装置21は記憶装置22から制御プログラムPRaを読み取り実行することによって、第1キーワード生成部210、第2キーワード生成部220A、特定部230A、及びコメント生成部240として機能する。
[1.3. Functions of User Device 20]
FIG. 4 is a functional block diagram showing the functions of the user device 20. The processing device 21 functions as the first keyword generation unit 210, the second keyword generation unit 220A, the identification unit 230A, and the comment generation unit 240 by reading and executing the control program PRa from the storage device 22.
 第1キーワード生成部210は、音声信号Saによって示されるユーザUの音声に基づいて第1キーワードKW1を生成する。
 具体的には、第1キーワード生成部210は、ユーザUの音声を解析し、解析結果から名詞と形容詞とを抽出する。第1キーワード生成部210は、ユーザUの音声に名詞と形容詞との両方が含まれている場合には注目ワードとして名詞を特定する。例えば、ユーザUの音声が「赤い車が怪しい」である場合、名詞である「車」が注目ワードとして特定される。また、第1キーワード生成部210は、ユーザUの音声に名詞が含まれておらず形容詞が含まれている場合、形容詞を注目ワードとして特定する。例えば、ユーザUの音声が「美味しそうだな」である場合、形容詞である「美味しい」が注目ワードとして特定される。
The first keyword generation unit 210 generates the first keyword KW1 based on the voice of the user U indicated by the voice signal Sa.
Specifically, the first keyword generation unit 210 analyzes the voice of the user U and extracts a noun and an adjective from the analysis result. When the voice of the user U includes both a noun and an adjective, the first keyword generation unit 210 identifies the noun as the attention word. For example, when the voice of the user U is “a red car is suspicious”, the noun “car” is specified as the attention word. Further, when the voice of the user U does not include a noun but does include an adjective, the first keyword generation unit 210 identifies the adjective as the attention word. For example, when the voice of the user U is “it looks delicious”, the adjective “delicious” is specified as the attention word.
 また、第1キーワード生成部210は、注目ワードがキーワードテーブルTBLaに含まれるかを判定する。第1キーワード生成部210は、判定結果が否定である場合、第1キーワードKW1を生成しない。従って、第1キーワードKW1はキーワードテーブルTBLaに含まれるキーワードに限定される。一方、判定結果が肯定であり、かつ注目ワードが名詞である場合、第1キーワード生成部210は、注目ワードを第1キーワードKW1として生成する。判定結果が肯定であり、かつ注目ワードが形容詞である場合、第1キーワード生成部210は、キーワードテーブルTBLaを参照して、注目ワードに対応付けられている名詞のワードを第1キーワードKW1として生成する。例えば、注目ワードが「美味しい」である場合、第1キーワード生成部210は、第1キーワードKW1として「飲食物」を生成する。 Also, the first keyword generation unit 210 determines whether the focused word is included in the keyword table TBLa. The first keyword generation unit 210 does not generate the first keyword KW1 when the determination result is negative. Therefore, the first keyword KW1 is limited to the keywords included in the keyword table TBLa. On the other hand, when the determination result is affirmative and the focused word is a noun, the first keyword generation unit 210 generates the focused word as the first keyword KW1. When the determination result is affirmative and the focused word is an adjective, the first keyword generation unit 210 refers to the keyword table TBLa and generates a noun word associated with the focused word as the first keyword KW1. To do. For example, when the attention word is “delicious”, the first keyword generation unit 210 generates “food and drink” as the first keyword KW1.
 このように、第1キーワード生成部210は、ユーザUの発言が曖昧な場合であっても、ユーザUの発言に関連する第1キーワードKW1を生成する。 In this way, the first keyword generation unit 210 generates the first keyword KW1 related to the remark of the user U even when the remark of the user U is ambiguous.
 次に、第2キーワード生成部220Aは、画像信号Sgの示す画像から抽出したオブジェクト画像の各々について第2キーワードKW2を生成する。第2キーワード生成部220Aは、抽出部221と変換部222とを有する。 Next, the second keyword generation unit 220A generates the second keyword KW2 for each of the object images extracted from the image indicated by the image signal Sg. The second keyword generation unit 220A has an extraction unit 221 and a conversion unit 222.
 抽出部221は、画像信号Sgの示す画像から複数のオブジェクト画像を抽出する。1画面の画像には、多数のオブジェクト画像が存在する。 The extraction unit 221 extracts a plurality of object images from the image indicated by the image signal Sg. There are many object images in one screen image.
 画像信号Sgの示す画像が、図5に示される画像である場合、抽出部221が抽出する画像は、例えば、オブジェクト画像OB1~OB5である。 When the image indicated by the image signal Sg is the image shown in FIG. 5, the images extracted by the extraction unit 221 are, for example, the object images OB1 to OB5.
 変換部222は、抽出部221によって抽出された複数のオブジェクト画像OB1~OB5の各々を第2キーワードKW2に変換する。変換部222は、例えば、機械学習により学習された画像認識モデルを用いて、各オブジェクト画像OBを第2キーワードKW2に変換する。但し、第2キーワードKW2は、キーワードテーブルTBLaに記憶されているキーワードに含まれる。例えば、図5にされるオブジェクト画像OB1は「ワイン」、オブジェクト画像OB2は「ワイングラス」、オブジェクト画像OB3は「時計」、オブジェクト画像OB4は「キャンドル」、オブジェクト画像OB5は「洋食」に変換される。 The conversion unit 222 converts each of the plurality of object images OB1 to OB5 extracted by the extraction unit 221 into the second keyword KW2. The conversion unit 222 converts each object image OB into the second keyword KW2 using, for example, an image recognition model learned by machine learning. However, the second keyword KW2 is included in the keywords stored in the keyword table TBLa. For example, the object image OB1 shown in FIG. 5 is converted into “wine”, the object image OB2 is converted into “wine glass”, the object image OB3 is converted into “clock”, the object image OB4 is converted into “candle”, and the object image OB5 is converted into “western food”. It
 特定部230Aは第2キーワードKW2と第1キーワードKW1との関連性の程度を示す関連度に基づいて、第2キーワード生成部220Aによって生成された複数の第2キーワードKW2の中から対象キーワードWxを特定する。より具体的には、特定部230Aは、キーワードテーブルTBLaを参照して、第2キーワードKW2と第1キーワードKW1との関連性の程度を示す関連度を第2キーワードKW2と第1キーワードKW1との組ごとに取得する。特定部230Aは、最も関連度が大きい第2キーワードKW2と第1キーワードKW1との組みに含まれる第2キーワードKW2を対象キーワードWxとして特定する。後述のように、コメント生成部240は特定部230Aが特定した対象キーワードWxに関連するコメントを生成する。 The identifying unit 230A selects the target keyword Wx from the plurality of second keywords KW2 generated by the second keyword generating unit 220A based on the degree of association indicating the degree of association between the second keyword KW2 and the first keyword KW1. Identify. More specifically, the identifying unit 230A refers to the keyword table TBLa to determine the degree of association between the second keyword KW2 and the first keyword KW1 that indicates the degree of association between the second keyword KW2 and the first keyword KW1. Acquire for each group. The identifying unit 230A identifies the second keyword KW2 included in the set of the second keyword KW2 and the first keyword KW1 having the highest degree of association as the target keyword Wx. As described below, the comment generating unit 240 generates a comment related to the target keyword Wx specified by the specifying unit 230A.
 例えば、ユーザUが図5に示す画像を見て「美味しそうだな」と発言したとする。また、第1キーワードKW1として「飲食物」が生成され、第2キーワードKW2として、「ワイン」、「ワイングラス」、「時計」、「キャンドル」、及び「洋食」が生成されることを想定する。この場合、特定部230Aは、「飲食物」と「ワイン」との関連度、「飲食物」と「ワイングラス」との関連度、「飲食物」と「時計」との関連度、「飲食物」と「キャンドル」との関連度、「飲食物」と「洋食」との関連度を、キーワードテーブルTBLaを参照して取得する。特定部230Aは、取得した複数の関連度を比較して、関連度が最も高い第2キーワードKW2を対象キーワードWxとして特定する。 For example, suppose that the user U looks at the image shown in FIG. 5 and says “It looks delicious”. Further, it is assumed that “food and drink” is generated as the first keyword KW1 and “wine”, “wine glass”, “clock”, “candle”, and “western food” are generated as the second keyword KW2. .. In this case, the specifying unit 230A determines the degree of association between “food and drink” and “wine”, the degree of association between “food and drink” and “wine glass”, the degree of association between “food and drink” and “clock”, and “food and drink”. The degree of association between “thing” and “candle” and the degree of association between “food and drink” and “western food” are acquired by referring to the keyword table TBLa. The identifying unit 230A identifies the second keyword KW2 having the highest degree of association as the target keyword Wx by comparing the obtained plurality of degrees of association.
 コメント生成部240は、対象キーワードWxに関連するコメントを生成する。コメントとは、対象キーワードWxについての説明(explanation)又は解説(exposition)を意味する。また、コメントはレコメンド(recommendation)を含む概念である。このため、対象キーワードWxに関連してユーザUに購入を勧める商品及び当該商品を取り扱う店舗に関する情報がコメントに含まれる。コメント生成部240は、対象キーワードWxに対応付けられて記憶されたコメントをコメントテーブルTBLbから読み出すことによってコメントを生成する。また、コメント生成部240は、ネットワークNWに接続される検索サイトにアクセスして当該検索サイトから対象キーワードWxに関連する情報を取得し、取得した情報をコメントとして生成してもよい。例えば、対象キーワードWxが「ラーメン」である場合、コメント生成部240は、GPS装置27などで生成される位置情報の近くのラーメン屋を検索し、検索結果をコメントとして出力してもよい。 The comment generator 240 generates a comment related to the target keyword Wx. The comment means an explanation or an explanation of the target keyword Wx. A comment is a concept that includes recommendations. Therefore, the comment includes information about the product recommended to the user U and the store handling the product in relation to the target keyword Wx. The comment generation unit 240 generates a comment by reading from the comment table TBLb the comment stored in association with the target keyword Wx. Further, the comment generation unit 240 may access a search site connected to the network NW, acquire information related to the target keyword Wx from the search site, and generate the acquired information as a comment. For example, when the target keyword Wx is “ramen”, the comment generation unit 240 may search for a ramen restaurant near the position information generated by the GPS device 27 or the like and output the search result as a comment.
[1.4.ユーザ装置20の動作]
 次に、ユーザ装置20の動作について説明する。図6は、ユーザ装置20の動作を示すフローチャートである。
[1.4. Operation of User Device 20]
Next, the operation of the user device 20 will be described. FIG. 6 is a flowchart showing the operation of the user device 20.
 まず、処理装置21は、ユーザUの音声に基づいて注目ワードを特定する(ステップS1)。処理装置21は、ユーザUの音声をテキストに変換する音声認識処理と、変換したテキストから名詞及び形容詞を特定する特定処理を実行することによって、注目ワードを抽出する。抽出ワードは、名詞、又は、名詞が特定されない場合には形容詞である。 First, the processing device 21 identifies the attention word based on the voice of the user U (step S1). The processing device 21 extracts the attention word by performing a voice recognition process for converting the voice of the user U into text and a specifying process for specifying a noun and an adjective from the converted text. The extraction word is a noun or an adjective if the noun is not specified.
 次に、処理装置21は、注目ワードがキーワードテーブルTBLaに含まれているか否かを判定する(ステップS2)。注目ワードがキーワードテーブルTBLaに含まれていない場合、処理装置21は、処理をステップS1に戻し、キーワードテーブルTBLaに含まれる注目ワードが特定されるまで(すなわち、ステップS2の判定結果が肯定となるまで)、ステップS1及びS2の処理を繰り返す。 Next, the processing device 21 determines whether or not the focused word is included in the keyword table TBLa (step S2). When the word of interest is not included in the keyword table TBLa, the processing device 21 returns the process to step S1 until the word of interest included in the keyword table TBLa is specified (that is, the determination result of step S2 is affirmative). Up to), the processes of steps S1 and S2 are repeated.
 ステップS2の判定結果が肯定の場合、処理装置21は注目ワードが名詞であるか否かを判定する(ステップS3)。注目ワードが名詞である場合、処理装置21は注目ワードを第1キーワードKW1として生成する。ステップS1の処理において、処理装置21は名詞又は形容詞を注目ワードとして抽出しているので、ステップS3の判定結果が否定の場合、注目ワードは形容詞となる。この場合、処理装置21はキーワードテーブルTBLaを参照して、注目ワードに対応付けられている名詞のワードを第1キーワードKW1として生成する(ステップS5)。 If the determination result in step S2 is affirmative, the processing device 21 determines whether or not the focused word is a noun (step S3). When the attention word is a noun, the processing device 21 generates the attention word as the first keyword KW1. In the process of step S1, the processing device 21 extracts a noun or an adjective as the attention word. Therefore, when the determination result of step S3 is negative, the attention word is the adjective. In this case, the processing device 21 refers to the keyword table TBLa and generates the word of the noun associated with the focused word as the first keyword KW1 (step S5).
 次に、処理装置21は、画像信号Sgの示す画像からオブジェクト画像を抽出する(ステップS6)。1フレームの画像には、通常、複数のオブジェクト画像が存在する。このため、処理装置21は、ステップS6の処理において複数のオブジェクト画像を抽出する。この後、処理装置21は、抽出された複数のオブジェクト画像の各々を第2キーワードKW2に変換する(ステップS7)。 Next, the processing device 21 extracts an object image from the image indicated by the image signal Sg (step S6). A plurality of object images usually exist in one frame image. Therefore, the processing device 21 extracts a plurality of object images in the process of step S6. After that, the processing device 21 converts each of the plurality of extracted object images into the second keyword KW2 (step S7).
 次に、処理装置21は、第2キーワードKW2と第1キーワードKW1との関連度に基づいて、ステップS7で生成された複数の第2キーワードKW2の中から対象キーワードWxを特定する。 Next, the processing device 21 identifies the target keyword Wx from the plurality of second keywords KW2 generated in step S7 based on the degree of association between the second keyword KW2 and the first keyword KW1.
 次に、処理装置21は、対象キーワードWxに関連するコメントを生成する(ステップS9)。ステップS9の処理において、処理装置21は対象キーワードWxに対応付けられて記憶されたコメントをコメントテーブルTBLbから読み出すことによってコメントを生成する。処理装置21は、生成したコメントを以下のいずれかの方法で出力する。(a)処理装置21は、生成したコメントの画像がスーパーインポーズされた動画データが表わす動画を、表示装置241に表示させる。(b) 処理装置21は、近距離無線通信装置26を用いて、生成したコメントの画像がスーパーインポーズされた動画データをテレビジョン受像機30に送信する。(c)処理装置21は、生成したコメントを音データに変換し、コメントを表わす音データを動画の音データと合成して合成結果をスピーカ242から放音させる。(d) 処理装置21は、近距離無線通信装置26を用いて、コメントを表わす音データと動画の音データとの合成結果をテレビジョン受像機30に送信する。(a)から(d)の方法を任意に組み合わせても良い。 Next, the processing device 21 generates a comment related to the target keyword Wx (step S9). In the process of step S9, the processing device 21 generates a comment by reading the comment stored in association with the target keyword Wx from the comment table TBLb. The processing device 21 outputs the generated comment by one of the following methods. (a) The processing device 21 causes the display device 241 to display the moving image represented by the moving image data in which the generated comment image is superimposed. (b) The processing device 21 uses the short-range wireless communication device 26 to transmit the moving image data in which the generated comment image is superimposed to the television receiver 30. (c) The processing device 21 converts the generated comment into sound data, combines the sound data representing the comment with the sound data of the moving image, and causes the speaker 242 to emit the combined result. (d) The processing device 21 uses the short-range wireless communication device 26 to transmit the synthesis result of the sound data representing the comment and the sound data of the moving image to the television receiver 30. The methods (a) to (d) may be combined arbitrarily.
 また、処理装置21は、ステップS1からステップS5までの処理において第1キーワード生成部210として機能し、ステップS6の処理において抽出部221として機能し、ステップS7の処理において変換部222として機能する。さらに、処理装置21は、ステップS8の処理において特定部230Aとして機能し、ステップS9の処理おいてコメント生成部240として機能する。 Further, the processing device 21 functions as the first keyword generation unit 210 in the processing of steps S1 to S5, functions as the extraction unit 221 in the processing of step S6, and functions as the conversion unit 222 in the processing of step S7. Further, the processing device 21 functions as the identifying unit 230A in the process of step S8, and functions as the comment generating unit 240 in the process of step S9.
 以上、説明したようにユーザ装置20の一例である情報処理装置は、ユーザUの音声に基づいて第1キーワードKW1を生成する第1キーワード生成部210と、画像信号Sgの示す画像から抽出した複数のオブジェクト画像の各々について第2キーワードKW2を生成する第2キーワード生成部220Aと、各第2キーワードKW2と第1キーワードKW1との関連性の程度に基づいて、第2キーワード生成部220Aによって生成された複数の第2キーワードKW2の中からコメントの対象となる対象キーワードWxを特定する特定部230Aと、対象キーワードWxに関連するコメントを生成するコメント生成部240と、を備える。 As described above, the information processing apparatus, which is an example of the user apparatus 20, includes the first keyword generation unit 210 that generates the first keyword KW1 based on the voice of the user U, and the plurality of extracted from the image indicated by the image signal Sg. Generated by the second keyword generation unit 220A based on the degree of association between the second keyword KW2 and the first keyword KW1 and the second keyword generation unit 220A that generates the second keyword KW2 for each of the object images. Also, a specifying unit 230A that specifies a target keyword Wx to be a comment target from the plurality of second keywords KW2 and a comment generating unit 240 that generates a comment related to the target keyword Wx are provided.
 この態様によれば、ユーザUがコメントの対象となるオブジェクト画像を指定することなく、ユーザUの曖昧な発言に応答して、コメントを生成することができる。 According to this aspect, the comment can be generated in response to the vague statement of the user U without the user U designating the object image to be the comment target.
 また、一の第2キーワードKW2が第1キーワードKW1に不一致である場合の第2キーワードKW2と第1キーワードKW1との関連度と比較して、他の第2キーワードKW2が第1キーワードKW1に一致する場合の関連度が高くなる。従って、複数の第2キーワードKW2のいずれかが第1キーワードKW1と一致する場合、特定部230Aは第1キーワードKW1と一致する第2キーワードKW2を対象キーワードWxとして特定する。この場合、特定部230Aは、複数の第2キーワードKW2の各々が第1キーワードKW1に一致するかを判定し、複数の第2キーワードKW2のいずれかに関する判定結果が肯定の場合は、第1キーワードKW1に一致する第2キーワードKW2を対象キーワードWxとして特定することができる。このため、キーワードテーブルTBLaを参照して関連度を取得する必要が無く、処理負荷を軽減することができる。 Further, as compared with the degree of association between the second keyword KW2 and the first keyword KW1 when the one second keyword KW2 does not match the first keyword KW1, another second keyword KW2 matches the first keyword KW1. The degree of relevance when doing is high. Therefore, when any of the plurality of second keywords KW2 matches the first keyword KW1, the specifying unit 230A specifies the second keyword KW2 that matches the first keyword KW1 as the target keyword Wx. In this case, the identifying unit 230A determines whether each of the plurality of second keywords KW2 matches the first keyword KW1, and when the determination result regarding any of the plurality of second keywords KW2 is affirmative, the first keyword The second keyword KW2 that matches KW1 can be specified as the target keyword Wx. Therefore, it is not necessary to refer to the keyword table TBLa to acquire the degree of association, and the processing load can be reduced.
[2.第2実施形態]
 第2実施形態のサービスシステム1は、ユーザ装置20における処理装置21の機能を除いて、第1実施形態のサービスシステム1と同一である。図7は第2実施形態の処理装置21の機能を示す機能ブロック図である。第2実施形態の処理装置21は、第2キーワード生成部220Aの替わりに第2キーワード生成部220Bを備える点で、第1実施形態の処理装置21と相違する。
[2. Second Embodiment]
The service system 1 of the second embodiment is the same as the service system 1 of the first embodiment, except for the function of the processing device 21 in the user device 20. FIG. 7 is a functional block diagram showing the functions of the processing device 21 of the second embodiment. The processing device 21 of the second embodiment differs from the processing device 21 of the first embodiment in that a second keyword generation unit 220B is provided instead of the second keyword generation unit 220A.
 図7に示されるように第2キーワード生成部220Bは、抽出部221、変換部222、及び解析部223を備える。解析部223には画像信号Sgが供給される。解析部223は、動画の画像信号Sgを解析して解析結果を抽出部221に出力する。 As shown in FIG. 7, the second keyword generation unit 220B includes an extraction unit 221, a conversion unit 222, and an analysis unit 223. The image signal Sg is supplied to the analysis unit 223. The analysis unit 223 analyzes the image signal Sg of the moving image and outputs the analysis result to the extraction unit 221.
 解析部223は、例えば、画像信号Sgの任意のフレームに含まれる複数のオブジェクト画像の各々を、第1の評価項目から第4の評価項目を用いて評価し、評価値の合計を解析結果として抽出部221に出力する。この場合、抽出部221は評価値の合計が所定値を超えるオブジェクト画像を抽出する。 The analysis unit 223, for example, evaluates each of the plurality of object images included in an arbitrary frame of the image signal Sg using the first to fourth evaluation items, and the total of the evaluation values as the analysis result. It is output to the extraction unit 221. In this case, the extraction unit 221 extracts an object image whose total evaluation value exceeds a predetermined value.
 第1の評価項目は、1画面の面積に対するオブジェクト画像の面積の割合であり、オブジェクト画像の割合が大きいほど当該オブジェクト画像の評価値が高い。第2の評価項目は、ユーザUから見たオブジェクト画像の遠近であり、オブジェクト画像が手前に位置するほど当該オブジェクト画像の評価値が高い。第3の評価項目は、オブジェクト画像の明るさであり、オブジェクト画像の明るさが明るいほど当該オブジェクト画像の評価値が高い。第4の評価項目は、オブジェクト画像の位置であり、オブジェクト画像の位置が画面の中心に近いほど当該オブジェクト画像の評価値が高い。第1から第4の評価項目は、いずれも1画面の画像中でユーザUの関心を引く要素である。複数の評価項目を用いてオブジェクト画像を評価することによって、ユーザUの関心が高いオブジェクト画像を抽出することができる。 The first evaluation item is the ratio of the area of the object image to the area of one screen, and the larger the ratio of the object image, the higher the evaluation value of the object image. The second evaluation item is the perspective of the object image viewed from the user U, and the evaluation value of the object image is higher as the object image is closer to the front. The third evaluation item is the brightness of the object image, and the brighter the brightness of the object image, the higher the evaluation value of the object image. The fourth evaluation item is the position of the object image, and the closer the position of the object image is to the center of the screen, the higher the evaluation value of the object image. Each of the first to fourth evaluation items is an element that attracts the interest of the user U in the image of one screen. By evaluating the object image using a plurality of evaluation items, the object image in which the user U is highly interested can be extracted.
 次に、第2実施形態におけるユーザ装置20の動作を説明する。図8は、第2実施形態に係るユーザ装置20の動作を示すフローチャートである。同図に示すフローチャートは、ステップS6の替わりにステップS6_1及びS6_2を実行する点を除いて、図6に示す第1実施形態のフローチャートと同一である。以下、相違点について説明する。 Next, the operation of the user device 20 in the second embodiment will be described. FIG. 8 is a flowchart showing the operation of the user device 20 according to the second embodiment. The flowchart shown in the figure is the same as the flowchart of the first embodiment shown in FIG. 6 except that steps S6_1 and S6_2 are executed instead of step S6. The differences will be described below.
 ステップS6_1において、処理装置21は解析部223として機能し、あるフレームに含まれる複数のオブジェクト画像の各々を評価項目ごとに評価した複数の評価値を取得し、これら評価値の合計を算出する。例えば、図5に示すオブジェクト画像OB1~OB5の解析結果は、図9に示すものとなる。この例では、オブジェクト画像OB1~OB5の各々について評価値の合計は「11」~「16」の範囲にある。 In step S6_1, the processing device 21 functions as the analysis unit 223, acquires a plurality of evaluation values obtained by evaluating each of a plurality of object images included in a frame for each evaluation item, and calculates the sum of these evaluation values. For example, the analysis results of the object images OB1 to OB5 shown in FIG. 5 are as shown in FIG. In this example, the sum of the evaluation values for each of the object images OB1 to OB5 is in the range of “11” to “16”.
 ステップS6_2において、処理装置21は抽出部221として機能し、評価値の合計が所定値を超えるオブジェクト画像を抽出する。例えば、所定値が「13」であり、かつ、各オブジェクト画像について図9に示す評価値の合計が得られた場合を想定する。この場合、処理装置21はオブジェクト画像OB2及びOB5を抽出する。なお、ステップS7以降の処理は、図6を参照して第1実施形態で説明した処理と同一であるので、説明を省略する。 In step S6_2, the processing device 21 functions as the extraction unit 221 and extracts an object image whose total evaluation value exceeds a predetermined value. For example, it is assumed that the predetermined value is “13” and that the sum of the evaluation values shown in FIG. 9 is obtained for each object image. In this case, the processing device 21 extracts the object images OB2 and OB5. Note that the processing from step S7 onward is the same as the processing described in the first embodiment with reference to FIG. 6, so description will be omitted.
 以上、説明したように第2実施形態によれば、第2キーワード生成部220Bは、画像信号Sgを解析する解析部223と、解析部223の解析結果に基づいて、画像信号Sgの示す画像から複数のオブジェクト画像を抽出する抽出部221と、複数のオブジェクト画像の各々を第2キーワードKW2に変換する変換部222と、を備える。 As described above, according to the second embodiment, the second keyword generating unit 220B analyzes the image signal Sg from the image indicated by the image signal Sg based on the analysis unit 223 that analyzes the image signal Sg and the analysis result of the analyzing unit 223. An extraction unit 221 that extracts a plurality of object images, and a conversion unit 222 that converts each of the plurality of object images into a second keyword KW2 are provided.
 この態様によれば、画像信号Sgの解析結果に基づいて複数のオブジェクト画像を抽出するので、解析結果を用いることなくオブジェクト画像を抽出する場合と比較して、抽出するオブジェクト画像の数を減らすことができる。従って、変換部222の処理負荷を軽減できる。 According to this aspect, since a plurality of object images are extracted based on the analysis result of the image signal Sg, the number of object images to be extracted can be reduced as compared with the case where the object images are extracted without using the analysis result. You can Therefore, the processing load of the conversion unit 222 can be reduced.
 なお、解析部223は、複数のフレームに渡る画像信号Sgを解析して、解析結果を生成してもよい。この場合、解析部223は、第1の評価項目から第4の評価項目に加え、オブジェクト画像の動きに関する第5の評価項目を採用してもよい。第5の評価項目の一例は、動きのあるオブジェクト画像が画面内に存在する時間長に相当するフレーム数であり、この評価項目によれば、フレーム数が大きくなるほど(オブジェクト画像が画面内に存在する時間が長くなるほど)当該オブジェクト画像の評価値が高い。例えば、映画の主人公が動く場合、主人公の動きに追従するように映画が撮影されることが多い。このため、画像信号Sgが表わす動画が映画である場合、主人公を表わすオブジェクト画像の評価値及び主人公が所持する所持品を表わすオブジェクト画像の評価値を高くする。この結果、ユーザUが着目するオブジェクト画像を抽出部221が抽出する可能性を高めることができる。逆に、ユーザUが着目しないオブジェクト画像を抽出部221が抽出する可能性を低減できる。 The analysis unit 223 may analyze the image signal Sg over a plurality of frames and generate an analysis result. In this case, the analysis unit 223 may adopt the fifth evaluation item regarding the movement of the object image in addition to the first evaluation item to the fourth evaluation item. An example of the fifth evaluation item is the number of frames corresponding to the length of time that a moving object image exists in the screen. According to this evaluation item, the larger the number of frames is (the object image exists in the screen The evaluation value of the object image is higher (the longer the time is spent). For example, when the hero of a movie moves, the movie is often shot so as to follow the movement of the hero. Therefore, when the moving image represented by the image signal Sg is a movie, the evaluation value of the object image representing the protagonist and the evaluation value of the object image representing the belongings possessed by the protagonist are increased. As a result, it is possible to increase the possibility that the extraction unit 221 extracts the object image focused on by the user U. On the contrary, it is possible to reduce the possibility that the extraction unit 221 extracts the object image that the user U does not pay attention to.
[3.第3実施形態]
 第3実施形態のサービスシステム1は、ユーザ装置20における処理装置21の機能及び記憶装置22の記憶内容を除いて、第1実施形態のサービスシステム1と同一である。図10は第3実施形態の処理装置21の機能を示す機能ブロック図である。第3実施形態の処理装置21は、特定部230Aの替わりに特定部230Bを備える点で、第1実施形態の処理装置21と相違する。
[3. Third Embodiment]
The service system 1 of the third embodiment is the same as the service system 1 of the first embodiment, except for the function of the processing device 21 in the user device 20 and the stored contents of the storage device 22. FIG. 10 is a functional block diagram showing the functions of the processing device 21 of the third embodiment. The processing device 21 of the third embodiment is different from the processing device 21 of the first embodiment in that it includes a specifying unit 230B instead of the specifying unit 230A.
 第3実施形態のユーザ装置20の記憶装置22は、行動履歴テーブルTBLcを記憶する。行動履歴テーブルTBLcにはユーザUの行動履歴が記憶される。行動履歴には、ユーザUのインターネット検索履歴、商品及びサービスの購入履歴、SNS(Social Networking Service)のアクティビティ、及びWebブラウザのブックマークなどが含まれる。 The storage device 22 of the user device 20 of the third embodiment stores the action history table TBLc. The behavior history table TBLc stores the behavior history of the user U. The action history includes the Internet search history of the user U, the purchase history of products and services, the activity of SNS (Social Networking Service), the bookmark of the Web browser, and the like.
 特定部230Bは、各第2キーワードKW2と第1キーワードKW1との関連性の程度を示す関連度及びユーザUの行動履歴に基づいて、第2キーワード生成部220Aで生成された複数の第2キーワードKW2の中から対象キーワードWxを特定する。 The identifying unit 230B generates a plurality of second keywords generated by the second keyword generating unit 220A, based on the degree of association between each second keyword KW2 and the first keyword KW1 and the behavior history of the user U. The target keyword Wx is specified from KW2.
 まず、特定部230Bは、第2キーワード生成部220Aによって生成された複数の第2キーワードKW2のうち、第1キーワードKW1との関連度が所定値以上となる第2キーワードKW2を選択する。選択された第2キーワードKW2は、対象キーワードWxの候補となる。次に、特定部230Bは、行動履歴テーブルTBLcに記憶された行動履歴を参照して、選択された第2キーワードKW2の中から対象キーワードWxを特定する。例えば、関連度に基づいて選択された第2キーワードKW2が、「ワイン」及び「洋食」であったとする。また、行動履歴テーブルTBLcにワインの購入履歴が記録されているとする。この場合、特定部230Bは、行動履歴テーブルTBLcを参照して、ユーザUにワインの購入履歴があることを検知すると、「ワイン」と「洋食」のうち、「ワイン」を対象キーワードWxとして特定する。この結果、コメント生成部240は、「ワイン」に関するコメントを生成することができる。 First, the identifying unit 230B selects, from the plurality of second keywords KW2 generated by the second keyword generating unit 220A, the second keyword KW2 having a degree of association with the first keyword KW1 that is equal to or greater than a predetermined value. The selected second keyword KW2 becomes a candidate for the target keyword Wx. Next, the identifying unit 230B identifies the target keyword Wx from the selected second keywords KW2 with reference to the action history stored in the action history table TBLc. For example, assume that the second keyword KW2 selected based on the degree of association is “wine” and “western food”. Further, it is assumed that wine purchase history is recorded in the action history table TBLc. In this case, when the specifying unit 230B detects that the user U has a purchase history of wine by referring to the action history table TBLc, it specifies “wine” as the target keyword Wx of “wine” and “Western food”. To do. As a result, the comment generator 240 can generate a comment regarding “wine”.
 次に、第3実施形態におけるユーザ装置20の動作を説明する。図11は、第3実施形態に係るユーザ装置20の動作を示すフローチャートである。同図に示すフローチャートは、ステップS8の替わりにステップS8_1及びS8_2を実行する点を除いて、図6に示す第1実施形態のフローチャートと同一である。以下、相違点について説明する。 Next, the operation of the user device 20 in the third embodiment will be described. FIG. 11 is a flowchart showing the operation of the user device 20 according to the third embodiment. The flowchart shown in the figure is the same as the flowchart of the first embodiment shown in FIG. 6 except that steps S8_1 and S8_2 are executed instead of step S8. The differences will be described below.
 ステップS8_1において、処理装置21は特定部230Bとして機能し、ステップS7で生成された複数の第2キーワードKW2のうち、第1キーワードKW1との関連度が所定値以上となる第2キーワードKW2を選択する。 In step S8_1, the processing device 21 functions as the identifying unit 230B, and selects the second keyword KW2 having a degree of association with the first keyword KW1 that is equal to or greater than a predetermined value from the plurality of second keywords KW2 generated in step S7. To do.
 ステップS8_2において、処理装置21は特定部230Bとして機能し、行動履歴に基づいて、ステップS8_1の処理で選択された第2キーワードKW2のうち、行動履歴に関連する第2キーワードKW2を対象キーワードWxとして特定する。 In step S8_2, the processing device 21 functions as the identifying unit 230B, and the second keyword KW2 related to the action history among the second keywords KW2 selected in the process of step S8_1 is set as the target keyword Wx based on the action history. Identify.
 第3実施形態によれば、特定部230Bは、関連性の程度を示す関連度及びユーザUの行動履歴に基づいて、複数の第2キーワードKW2の中から対象キーワードWxを特定する。この態様によれば、ユーザUの行動履歴を考慮して対象キーワードWxを特定するため、ユーザUの行動履歴を考慮しない場合と比較して、ユーザUの関心の高いコメントを提供することができる。 According to the third embodiment, the identifying unit 230B identifies the target keyword Wx from the plurality of second keywords KW2 based on the degree of association indicating the degree of association and the behavior history of the user U. According to this aspect, since the target keyword Wx is specified in consideration of the behavior history of the user U, it is possible to provide a comment of high interest to the user U, as compared with the case where the behavior history of the user U is not considered. ..
 なお、図11を参照して説明したユーザ装置20の動作では、特定部230Bは、関連度を用いて第2キーワード生成部220Aが生成した複数の第2キーワードKW2のうち、対象キーワードWxの候補となる第2キーワードKW2を選択し(ステップS8_1)、その後、行動履歴に基づいて対象キーワードWxを特定する(ステップS8_2)が、順序を逆転させてもよい。即ち、特定部230Bは、行動履歴に基づいて第2キーワード生成部220Aで生成した複数の第2キーワードKW2のうち対象キーワードWxの候補となる第2キーワードKW2を選択し、その後、関連度を用いて対象キーワードWxを特定してもよい。加えて、特定部230Bは、行動履歴及び関連度を同時に用いて、複数の第2キーワードKW2の中から、対象キーワードWxを特定してもよい。特定部230Bは、例えば、行動履歴に関連する第2キーワードKW2については関連度に所定値を加算し、所定値が加算された関連度を複数の第2キーワードKW2同士で比較して対象キーワードWxを特定してもよい。 Note that in the operation of the user device 20 described with reference to FIG. 11, the identifying unit 230B is a candidate for the target keyword Wx among the plurality of second keywords KW2 generated by the second keyword generating unit 220A using the degree of association. Then, the second keyword KW2 is selected (step S8_1), and then the target keyword Wx is specified based on the action history (step S8_2), but the order may be reversed. That is, the identification unit 230B selects the second keyword KW2 that is a candidate for the target keyword Wx from the plurality of second keywords KW2 generated by the second keyword generation unit 220A based on the action history, and then uses the degree of association. Alternatively, the target keyword Wx may be specified. In addition, the identifying unit 230B may identify the target keyword Wx from the plurality of second keywords KW2 by using the action history and the degree of association at the same time. The specifying unit 230B adds, for example, a predetermined value to the degree of association for the second keyword KW2 related to the action history, and compares the degree of association to which the predetermined value is added between the plurality of second keywords KW2 to determine the target keyword Wx. May be specified.
[4.第4実施形態]
 第4実施形態のサービスシステム1は、ユーザ装置20における処理装置21の機能及び記憶装置22の記憶内容を除いて、第1実施形態のサービスシステム1と同一である。図12は第4実施形態の処理装置21の機能を示す機能ブロック図である。第4実施形態の処理装置21は、特定部230Aの替わりに特定部230Cを備える点で、第1実施形態の処理装置21と相違する。
[4. Fourth Embodiment]
The service system 1 of the fourth embodiment is the same as the service system 1 of the first embodiment, except for the function of the processing device 21 in the user device 20 and the stored contents of the storage device 22. FIG. 12 is a functional block diagram showing the functions of the processing device 21 of the fourth embodiment. The processing device 21 of the fourth embodiment is different from the processing device 21 of the first embodiment in that a specifying unit 230C is provided instead of the specifying unit 230A.
 第4実施形態のユーザ装置20の記憶装置22は、プロファイルデータDPと評価テーブルTBLdとを記憶する。プロファイルデータDPはユーザUのプロファイルを示す。プロファイルとは、ユーザUの属性の意味であり、年齢、性別などの項目が含まれる。 The storage device 22 of the user device 20 of the fourth embodiment stores the profile data DP and the evaluation table TBLd. The profile data DP indicates the profile of the user U. The profile means the attribute of the user U, and includes items such as age and sex.
 評価テーブルTBLdには、プロファイルの項目ごとの評価値がキーワードと対応付けて記憶される。評価値は、キーワードに対するユーザUの関心の程度を示す値である。図13は評価テーブルTBLdの記憶内容の一例を示す。例えば、キーワード「車」について、性別「男」の評価値は「7」であるのに対し、性別「女」の評価値は「4」である。これは、男性が女性より車に関心が高いことを示している。 In the evaluation table TBLd, the evaluation value for each profile item is stored in association with the keyword. The evaluation value is a value indicating the degree of interest of the user U with respect to the keyword. FIG. 13 shows an example of the stored contents of the evaluation table TBLd. For example, for the keyword “car”, the evaluation value for gender “male” is “7”, whereas the evaluation value for gender “female” is “4”. This shows that men are more interested in cars than women.
 特定部230Cは、各第2キーワードKW2と第1キーワードKW1との関連性の程度を示す関連度及びユーザUのプロファイルに基づいて、第2キーワード生成部220Aで生成された複数の第2キーワードKW2の中から対象キーワードWxを特定する。 The identifying unit 230C uses the plurality of second keywords KW2 generated by the second keyword generating unit 220A based on the degree of association between each second keyword KW2 and the first keyword KW1 and the profile of the user U. The target keyword Wx is specified from among these.
 まず、特定部230Cは、第2キーワード生成部220Aによって生成された複数の第2キーワードKW2のうち、第1キーワードKW1との関連度が所定値以上となる第2キーワードKW2を選択する。次に、特定部230Cは、プロファイルデータDPと評価テーブルTBLdとを用いて、選択された第2キーワードKW2の中から対象キーワードWxを特定する。具体的には、選択された第2キーワードKW2の各々について、ユーザUのプロファイルの複数の項目にそれぞれ対応する評価値を合計した合計評価値を算出し、合計評価値が最も高い第2キーワードKW2を対象キーワードWxとして特定する。 First, the identifying unit 230C selects the second keyword KW2 having a degree of association with the first keyword KW1 that is equal to or more than a predetermined value, from among the plurality of second keywords KW2 generated by the second keyword generating unit 220A. Next, the specifying unit 230C specifies the target keyword Wx from the selected second keywords KW2 using the profile data DP and the evaluation table TBLd. Specifically, for each of the selected second keywords KW2, a total evaluation value obtained by summing evaluation values corresponding to a plurality of items of the profile of the user U is calculated, and the second keyword KW2 having the highest total evaluation value is calculated. Is specified as the target keyword Wx.
 次に、第4実施形態におけるユーザ装置20の動作を説明する。図14は、第4実施形態に係るユーザ装置20の動作を示すフローチャートである。同図に示すフローチャートは、ステップS8_2の替わりにステップS8_3を実行する点を除いて、図12に示す第3実施形態のフローチャートと同一である。以下、相違点について説明する。 Next, the operation of the user device 20 in the fourth embodiment will be described. FIG. 14 is a flowchart showing the operation of the user device 20 according to the fourth embodiment. The flowchart shown in the figure is the same as the flowchart of the third embodiment shown in FIG. 12 except that step S8_3 is executed instead of step S8_2. The differences will be described below.
 ステップS8_3において、処理装置21は特定部230Cとして機能し、プロファイルに基づいて、ステップS8_1の処理で選択された第2キーワードKW2のうち、ユーザUのプロファイルの合計評価値が最も高い第2キーワードKW2を対象キーワードWxとして特定する。 In step S8_3, the processing device 21 functions as the specifying unit 230C, and the second keyword KW2 having the highest total evaluation value of the profile of the user U among the second keywords KW2 selected in the process of step S8_1 based on the profile. Is specified as the target keyword Wx.
 第4実施形態によれば、特定部230Cは、関連性の程度を示す関連度及びユーザUのプロファイルに基づいて、複数の第2キーワードKW2の中から対象キーワードWxを特定する。この態様によれば、ユーザUのプロファイルを考慮して対象キーワードWxを特定するため、ユーザUのプロファイルを考慮しない場合と比較して、ユーザUの関心の高いコメントを提供することができる。 According to the fourth embodiment, the identifying unit 230C identifies the target keyword Wx from the plurality of second keywords KW2 based on the degree of association indicating the degree of association and the profile of the user U. According to this aspect, since the target keyword Wx is specified in consideration of the profile of the user U, it is possible to provide a comment of high interest to the user U, as compared with the case where the profile of the user U is not considered.
 なお、図14を参照して説明したユーザ装置20の動作では、特定部230Bは、関連度を用いて第2キーワード生成部220Aで生成した複数の第2キーワードKW2のうち対象キーワードWxの候補となる第2キーワードKW2を選択し(ステップS8_1)、その後、プロファイルに基づいて対象キーワードWxを特定する(ステップS8_3)が、順序を逆転させてもよい。即ち、特定部230Cは、プロファイルに基づいて第2キーワード生成部220Aで生成した複数の第2キーワードKW2のうち対象キーワードWxの候補となる第2キーワードKW2を選択し、その後、関連度を用いて対象キーワードWxを特定してもよい。加えて、特定部230Cは、プロファイル及び関連度を同時に用いて、複数の第2キーワードKW2の中から、対象キーワードWxを特定してもよい。特定部230Cは、例えば、プロファイルに基づく合計評価値を関連度に加算し、複数の第2キーワードKW2についてのそれぞれの加算結果を比較して対象キーワードWxを特定してもよい。 In the operation of the user device 20 described with reference to FIG. 14, the identifying unit 230B determines that the target keyword Wx is a candidate of the target keyword Wx among the plurality of second keywords KW2 generated by the second keyword generating unit 220A using the degree of association. The second keyword KW2 is selected (step S8_1), and then the target keyword Wx is specified based on the profile (step S8_3), but the order may be reversed. That is, the specifying unit 230C selects the second keyword KW2 that is a candidate for the target keyword Wx from the plurality of second keywords KW2 generated by the second keyword generating unit 220A based on the profile, and then uses the degree of association. The target keyword Wx may be specified. In addition, the specifying unit 230C may specify the target keyword Wx from the plurality of second keywords KW2 by using the profile and the degree of association at the same time. The specifying unit 230C may add the total evaluation value based on the profile to the degree of association, and compare the addition results of the plurality of second keywords KW2 to specify the target keyword Wx.
[5.変形例]
 本発明は、以上に例示した各実施形態に限定されない。具体的な変形の態様を以下に例示する。以下の例示から任意に選択された2以上の態様を併合してもよい。
[5. Modification]
The present invention is not limited to the embodiments illustrated above. Specific modes of modification will be exemplified below. Two or more aspects arbitrarily selected from the following examples may be merged.
(1)上述した各実施形態において、抽出部221が画像信号Sgの画像からオブジェクト画像を抽出するフレームは以下のフレームであってもよい。
 第1に、抽出部221は、視聴率の高いフレームでオブジェクト画像を抽出してもよい。この場合、抽出部221は、視聴率を外部装置からリアルタイムで取得すればよい。具体的には、抽出部221は、取得した視聴率が所定の視聴率を超えたフレームでオブジェクト画像の抽出を実行する。視聴率が高いフレームは、他のフレームと比較してユーザUの関心が他の高いと推定される。従って、ユーザUの関心が高いフレームの画像からオブジェクト画像が抽出されるので、ユーザUに有益なコメントを生成できる。
 第2に、抽出部221は、ユーザUの音声信号Saに基づいて、ユーザUが歓声をあげたフレームでオブジェクト画像を抽出してもよい。
 第3に、抽出部221は、番組情報に基づいて番組の主題となるフレームでオブジェク画像を抽出してもよい。例えば、抽出部221は、第2実施形態で説明した解析部223を用いて、画像信号Sgを解析し、番組の主題となるフレームを特定してもよい。この場合、解析部223は、ネットワークNWを介して外部装置から番組情報を取得すればよい。
 また、上述した各実施形態では、画像信号Sgは動画を示す信号として説明したが、画像信号Sgは静止画を示す信号であってもよい。
(1) In each of the above-described embodiments, the frame in which the extraction unit 221 extracts the object image from the image of the image signal Sg may be the following frame.
First, the extraction unit 221 may extract the object image in a frame having a high audience rating. In this case, the extraction unit 221 may acquire the audience rating from the external device in real time. Specifically, the extraction unit 221 extracts the object image in a frame in which the acquired audience rating exceeds a predetermined audience rating. It is estimated that the user U has a higher interest in a frame having a higher audience rating than other frames. Therefore, since the object image is extracted from the image of the frame in which the user U has a high interest, a comment useful for the user U can be generated.
Secondly, the extraction unit 221 may extract the object image in a frame in which the user U cheers based on the audio signal Sa of the user U.
Thirdly, the extraction unit 221 may extract the object image in a frame that is the subject of the program based on the program information. For example, the extraction unit 221 may use the analysis unit 223 described in the second embodiment to analyze the image signal Sg and identify the frame that is the subject of the program. In this case, the analysis unit 223 may acquire the program information from the external device via the network NW.
Further, in each of the above-described embodiments, the image signal Sg has been described as a signal indicating a moving image, but the image signal Sg may be a signal indicating a still image.
(2)上述した各実施形態において、第2キーワードKW2と第1キーワードKW1との関連性の程度を示す関連度は、キーワードテーブルTBLaに記憶されていたが、これに限定されない。
 例えば、特定部230A、230B、及び230Cは、複数の単語が意味によって階層化された木構造を有するキーワードテーブルTBLa(キーワードデータの一例)から特定されるノード数に応じた関連度を取得してもよい。具体的には、第1キーワード生成部210は、キーワードテーブルTBLaに含まれる単語を第1キーワードKW1として生成する。また、第2キーワード生成部220A及び220Bは、キーワードテーブルTBLaに含まれる単語を第2キーワードKW2として生成する。特定部230A、230B、及び230Cは、キーワードテーブルTBLaの木構造において、第1キーワードKW1から第2キーワードKW2までの経路におけるノード数を関連度として取得する。
 さらに具体的には、キーワードテーブルTBLaのデータ構造が図3に示される木構造である場合を想定する。例えば、第1キーワードKW1が「酒」であり、第2キーワードKW2が「フライドポテト」である場合、「酒」から「フライドポテト」に至る経路は、ノード「酒」→ノード「飲み物」→ノード「飲食物」→ノード「食べ物」→ノード「洋食」→ノード「フライドポテト」となる。従って、第1キーワードKW1「酒」から第2キーワードKW2「フライドポテト」に至る経路のノード数は、「5」となる。また、第1キーワードKW1が「飲食物」であり、第2キーワードKW2が「フライドポテト」である場合、「飲食物」から「フライドポテト」に至る経路は、ノード「飲食物」→ノード「食べ物」→ノード「洋食」→ノード「フライドポテト」となる。従って、第1キーワードKW1「飲食物」から第2キーワードKW2「フライドポテト」に至る経路のノード数は、「3」となる。第1キーワードKW1と第2キーワードKW2とを結ぶ経路のノード数が少ないほど関連度が高いから、上記の例においては、第1キーワードKW1「飲食物」と第2キーワードKW2「フライドポテト」との関連度は、第1キーワードKW1「酒」と第2キーワードKW2「フライドポテト」との関連度よりも高い。
 ノード数に応じて関連度を特定することによって、ユーザ装置20において必要となるキーワードテーブルTBLaの記憶容量を削減することができる。
(2) In each of the above-described embodiments, the degree of association indicating the degree of association between the second keyword KW2 and the first keyword KW1 is stored in the keyword table TBLa, but the present invention is not limited to this.
For example, the identifying units 230A, 230B, and 230C obtain the degree of association according to the number of nodes identified from the keyword table TBLa (an example of keyword data) having a tree structure in which a plurality of words are layered according to meaning. Good. Specifically, the first keyword generation unit 210 generates a word included in the keyword table TBLa as the first keyword KW1. Also, the second keyword generation units 220A and 220B generate words included in the keyword table TBLa as the second keyword KW2. The specifying units 230A, 230B, and 230C acquire the number of nodes in the route from the first keyword KW1 to the second keyword KW2 in the tree structure of the keyword table TBLa as the degree of association.
More specifically, it is assumed that the data structure of the keyword table TBLa is the tree structure shown in FIG. For example, when the first keyword KW1 is “liquor” and the second keyword KW2 is “french potato”, the route from “liquor” to “french potato” is node “liquor” → node “drink” → node “Food and drink” → node “food” → node “Western food” → node “fried potato”. Therefore, the number of nodes on the route from the first keyword KW1 "liquor" to the second keyword KW2 "fried potatoes" is "5". Further, when the first keyword KW1 is “food and drink” and the second keyword KW2 is “fries”, the route from “food and drink” to “fries” is node “food and drink” → node “food” -> Node "Western food"-> node "fries". Therefore, the number of nodes on the route from the first keyword KW1 "food and drink" to the second keyword KW2 "fried potatoes" is "3". The smaller the number of nodes on the route connecting the first keyword KW1 and the second keyword KW2, the higher the degree of association. The degree of association is higher than the degree of association between the first keyword KW1 “sake” and the second keyword KW2 “fried potato”.
By specifying the degree of association according to the number of nodes, it is possible to reduce the storage capacity of the keyword table TBLa required in the user device 20.
(3)上述した各実施形態において、抽出部221は、ユーザUの行動履歴を考慮せずにオブジェクト画像を抽出したが、画像信号Sgの示す画像から行動履歴に基づいてオブジェクト画像を抽出してもよい。この場合、抽出部221は、第3実施形態で説明した行動履歴テーブルTBLcを参照して、商品の購入履歴等から例えばユーザUの好みの色を特定し、特定した色のオブジェクト画像を抽出してもよい。この変形例によれば、オブジェクト画像を絞り込むことができるので、変換部222の処理負荷を軽減できる。 (3) In each of the above-described embodiments, the extraction unit 221 extracts the object image without considering the action history of the user U, but extracts the object image from the image indicated by the image signal Sg based on the action history. Good. In this case, the extraction unit 221 refers to the action history table TBLc described in the third embodiment, identifies a favorite color of the user U, for example, from the purchase history of the product, and extracts the object image of the identified color. May be. According to this modification, the object images can be narrowed down, so that the processing load of the conversion unit 222 can be reduced.
(4)上述した実施形態は適宜組み合わせることが可能である。例えば、第2実施形態の第2キーワード生成部220Bを第3実施形態及び第4実施形態の第2キーワード生成部220Aの替わりに用いてもよい。 (4) The above-described embodiments can be combined appropriately. For example, the second keyword generation unit 220B of the second embodiment may be used instead of the second keyword generation unit 220A of the third and fourth embodiments.
(5)上述した各実施形態の説明に用いたブロック図は、機能単位のブロックを示している。これらの機能ブロック(構成部)は、ハードウェア及び/又はソフトウェアの任意の組み合わせによって実現される。また、各機能ブロックの実現手段は特に限定されない。すなわち、各機能ブロックは、物理的及び/又は論理的に結合した1つの装置により実現されてもよいし、物理的及び/又は論理的に分離した2つ以上の装置を直接的及び/又は間接的に(例えば、有線及び/又は無線)で接続し、これら複数の装置により実現されてもよい。例えば、変換部222の機能はネットワークNWを介して接続されるサーバ装置から提供されてもよい。同様に、キーワードテーブルTBLaもサーバ装置に設けられてもよい。
 また、上述した各実施形態の説明に用いた「装置」という文言は、回路、デバイス又はユニット等の他の用語に読替えてもよい。
(5) The block diagram used in the description of each of the above-described embodiments shows blocks of functional units. These functional blocks (components) are realized by an arbitrary combination of hardware and / or software. Further, the means for realizing each functional block is not particularly limited. That is, each functional block may be realized by one device that is physically and / or logically coupled, or may be directly and / or indirectly connected to two or more devices that are physically and / or logically separated. (For example, wired and / or wireless), and may be realized by a plurality of these devices. For example, the function of the conversion unit 222 may be provided from a server device connected via the network NW. Similarly, the keyword table TBLa may be provided in the server device.
Further, the wording “apparatus” used in the description of each of the above-described embodiments may be replaced with another term such as a circuit, a device, or a unit.
(6)上述した各実施形態における処理手順、シーケンス、フローチャートなどは、矛盾の無い限り、順序を入れ替えてもよい。例えば、本明細書で説明した方法については、例示的な順序で様々なステップの要素を提示しており、提示した特定の順序に限定されない。 (6) The order of the processing procedures, sequences, flowcharts, etc. in each of the above-described embodiments may be changed as long as there is no contradiction. For example, the methods described herein present elements of the various steps in a sample order, and are not limited to the specific order presented.
(7)上述した各実施形態において、入出力された情報等は特定の場所(例えば、メモリ)に保存されてもよいし、管理テーブルで管理してもよい。入出力される情報等は、上書き、更新、又は追記され得る。出力された情報等は削除されてもよい。入力された情報等は他の装置へ送信されてもよい。 (7) In each of the above-described embodiments, the input / output information and the like may be stored in a specific place (for example, a memory) or may be managed by a management table. Information that is input / output may be overwritten, updated, or added. The output information and the like may be deleted. The input information and the like may be transmitted to another device.
(8)上述した各実施形態において、判定は、1ビットで表される値(0か1か)によって行われてもよいし、真偽値(Boolean:true又はfalse)によって行われてもよいし、数値の比較(例えば、所定の値との比較)によって行われてもよい。 (8) In each of the above-described embodiments, the determination may be performed by a value (0 or 1) represented by 1 bit, or by a true / false value (Boolean: true or false). However, the comparison may be performed by comparing numerical values (for example, comparison with a predetermined value).
(9)上述した各実施形態では、記憶装置22は、処理装置21が読取可能な記録媒体であり、ROM及びRAMなどを例示したが、フレキシブルディスク、光磁気ディスク(例えば、コンパクトディスク、デジタル多用途ディスク、Blu-ray(登録商標)ディスク)、スマートカード、フラッシュメモリデバイス(例えば、カード、スティック、キードライブ)、CD-ROM(Compact Disc-ROM)、レジスタ、リムーバブルディスク、ハードディスク、フロッピー(登録商標)ディスク、磁気ストリップ、データベース、サーバその他の適切な記憶媒体である。また、プログラムは、ネットワークNWから送信されても良い。また、プログラムは、電気通信回線を介して通信網から送信されても良い。 (9) In each of the above-described embodiments, the storage device 22 is a recording medium that can be read by the processing device 21, and is exemplified by a ROM and a RAM, but a flexible disk, a magneto-optical disk (for example, a compact disk, a digital multi-disk). Purpose disk, Blu-ray (registered trademark) disk), smart card, flash memory device (for example, card, stick, key drive), CD-ROM (Compact Disc-ROM), register, removable disk, hard disk, floppy (registered Trademark disks, magnetic strips, databases, servers and other suitable storage media. Further, the program may be transmitted from the network NW. Further, the program may be transmitted from a communication network via an electric communication line.
(10)上述した各実施形態は、LTE(Long Term Evolution)、LTE-A(LTE-Advanced)、SUPER 3G、IMT-Advanced、4G、5G、FRA(Future Radio Access)、W-CDMA(登録商標)、GSM(登録商標)、CDMA2000、UMB(Ultra Mobile Broadband)、IEEE 802.11(Wi-Fi)、IEEE 802.16(WiMAX)、IEEE 802.20、UWB(Ultra-WideBand)、Bluetooth(登録商標)、その他の適切なシステムを利用するシステム及び/又はこれらに基づいて拡張された次世代システムに適用されてもよい。 (10) Each of the above-described embodiments includes LTE (Long Term Evolution), LTE-A (LTE-Advanced), SUPER 3G, IMT-Advanced, 4G, 5G, FRA (Future Radio Access), W-CDMA (registered trademark) ), GSM (registered trademark), CDMA2000, UMB (Ultra Mobile Broadband), IEEE 802.11 (Wi-Fi), IEEE 802.11 (WiMAX), IEEE 802.20, UWB (Ultra-WideBand), Bluetooth (registered). (Trademark), other systems utilizing appropriate systems, and / or next-generation systems extended based on these.
(11)上述した各実施形態において、説明した情報及び信号などは、様々な異なる技術のいずれかを使用して表されてもよい。例えば、上述の説明全体に渡って言及され得るデータ、命令、コマンド、情報、信号、ビット、シンボル、チップなどは、電圧、電流、電磁波、磁界若しくは磁性粒子、光場若しくは光子、又はこれらの任意の組み合わせによって表されてもよい。
 なお、本明細書で説明した用語及び/又は本明細書の理解に必要な用語については、同一の又は類似する意味を有する用語と置き換えてもよい。
(11) In each of the embodiments described above, the information, signals, etc. described may be represented using any of a variety of different technologies. For example, data, instructions, commands, information, signals, bits, symbols, chips, etc., that may be mentioned throughout the above description are voltage, current, electromagnetic waves, magnetic fields or magnetic particles, optical fields or photons, or any of these. May be represented by a combination of
Note that the terms described in this specification and / or the terms necessary for understanding this specification may be replaced with terms having the same or similar meanings.
(12)図4、図7、図10、及び図12に例示された各機能は、ハードウェア及びソフトウェアの任意の組合せによって実現される。また、各機能は、単体の装置によって実現されてもよいし、相互に別体で構成された2個以上の装置によって実現されてもよい。 (12) Each function illustrated in FIGS. 4, 7, 10, and 12 is realized by an arbitrary combination of hardware and software. Further, each function may be realized by a single device, or may be realized by two or more devices that are configured separately from each other.
(13)上述した各実施形態で例示したプログラムは、ソフトウェア、ファームウェア、ミドルウェア、マイクロコード又はハードウェア記述言語と呼ばれるか、他の名称によって呼ばれるかを問わず、命令、命令セット、コード、コードセグメント、プログラムコード、サブプログラム、ソフトウェアモジュール、アプリケーション、ソフトウェアアプリケーション、ソフトウェアパッケージ、ルーチン、サブルーチン、オブジェクト、実行可能ファイル、実行スレッド、手順又は機能等を意味するよう広く解釈されるべきである。
 また、ソフトウェア、命令などは、伝送媒体を介して送受信されてもよい。例えば、ソフトウェアが、同軸ケーブル、光ファイバケーブル、ツイストペア及びデジタル加入者回線(DSL)などの有線技術及び/又は赤外線、無線及びマイクロ波などの無線技術を使用してウェブサイト、サーバ、又は他のリモートソースから送信される場合、これらの有線技術及び/又は無線技術は、伝送媒体の定義内に含まれる。
(13) The program illustrated in each of the above-described embodiments is an instruction, an instruction set, a code, or a code segment regardless of whether it is called software, firmware, middleware, microcode, a hardware description language, or another name. , Program code, subprograms, software modules, applications, software applications, software packages, routines, subroutines, objects, executable files, threads of execution, procedures or functions, etc., should be construed broadly.
Further, software, instructions, etc. may be transmitted and received via a transmission medium. For example, the software may use a wired technology such as coaxial cable, fiber optic cable, twisted pair and digital subscriber line (DSL) and / or wireless technology such as infrared, wireless and microwave to websites, servers, or other When transmitted from a remote source, these wireline and / or wireless technologies are included within the definition of transmission medium.
(14)上述した各実施形態において、「システム」及び「ネットワーク」という用語は、互換的に使用される。 (14) In the above-described embodiments, the terms "system" and "network" are used interchangeably.
(15)上述した各実施形態において、情報、パラメータなどは、絶対値で表されてもよいし、所定の値からの相対値で表されてもよいし、対応する別の情報で表されてもよい。 (15) In each of the above-described embodiments, information, parameters, and the like may be represented by an absolute value, a relative value from a predetermined value, or another corresponding information. Good.
(16)上述した各実施形態において、ユーザ装置20は、移動局である場合が含まれる。移動局は、当業者によって、加入者局、モバイルユニット、加入者ユニット、ワイヤレスユニット、リモートユニット、モバイルデバイス、ワイヤレスデバイス、ワイヤレス通信デバイス、リモートデバイス、モバイル加入者局、アクセス端末、モバイル端末、ワイヤレス端末、リモート端末、ハンドセット、ユーザエージェント、モバイルクライアント、クライアント、又はいくつかの他の適切な用語で呼ばれる場合もある。 (16) In each of the above-described embodiments, the case where the user device 20 is a mobile station is included. A mobile station can be a subscriber station, mobile unit, subscriber unit, wireless unit, remote unit, mobile device, wireless device, wireless communication device, remote device, mobile subscriber station, access terminal, mobile terminal, wireless, by a person skilled in the art. It may also be referred to as a terminal, remote terminal, handset, user agent, mobile client, client, or some other suitable term.
(17)上述した各実施形態において、「接続された(connected)」という用語、又はこれらのあらゆる変形は、2又はそれ以上の要素間の直接的又は間接的なあらゆる接続又は結合を意味し、互いに「接続」された2つの要素間に1又はそれ以上の中間要素が存在することを含むことができる。要素間の接続は、物理的なものであっても、論理的なものであっても、或いはこれらの組み合わせであってもよい。本明細書で使用する場合、2つの要素は、1又はそれ以上の電線、ケーブル及び/又はプリント電気接続を使用することにより、並びにいくつかの非限定的かつ非包括的な例として、無線周波数領域、マイクロ波領域及び光(可視及び不可視の両方)領域の波長を有する電磁エネルギーなどの電磁エネルギーを使用することにより、互いに「接続」されると考えることができる。 (17) In each of the embodiments described above, the term "connected" or any variation thereof means any direct or indirect connection or coupling between two or more elements, It can include the presence of one or more intermediate elements between two elements that are “connected” to each other. The connections between the elements may be physical, logical, or a combination thereof. As used herein, two elements are radio frequency by using one or more wires, cables and / or printed electrical connections, and as some non-limiting and non-exhaustive examples. By using electromagnetic energy, such as electromagnetic energy having wavelengths in the region, the microwave region and the light (both visible and invisible) region, it can be considered to be “connected” to each other.
(18)上述した各実施形態において、「に基づいて」という記載は、別段に明記されていない限り、「のみに基づいて」を意味しない。言い換えれば、「に基づいて」という記載は、「のみに基づいて」と「に少なくとも基づいて」の両方を意味する。 (18) In each of the above-described embodiments, the description “based on” does not mean “based only on” unless otherwise specified. In other words, the phrase "based on" means both "based only on" and "based at least on."
(19)本明細書で使用する「第1」、「第2」などの呼称を使用した要素へのいかなる参照も、それらの要素の量又は順序を全般的に限定するものではない。これらの呼称は、2つ以上の要素間を区別する便利な方法として本明細書で使用され得る。従って、第1及び第2の要素への参照は、2つの要素のみがそこで採用され得ること、又は何らかの形で第1の要素が第2の要素に先行しなければならないことを意味しない。 (19) As used herein, any reference to elements using the designations "first", "second", etc. does not generally limit the amount or order of those elements. These designations may be used herein as a convenient way to distinguish between two or more elements. Thus, references to the first and second elements do not imply that only two elements may be employed therein, or that the first element must precede the second element in any way.
(20)上述した各実施形態において「含む(including)」、「含んでいる(comprising)」、及びそれらの変形が、本明細書あるいは特許請求の範囲で使用されている限り、これら用語は、用語「備える」と同様に、包括的であることが意図される。さらに、本明細書あるいは特許請求の範囲において使用されている用語「又は(or)」は、排他的論理和ではないことが意図される。 (20) As long as the terms "including", "comprising", and variations thereof in each of the above-described embodiments are used in the present specification or claims, these terms are: Like the term “comprising” is intended to be inclusive. Furthermore, the term "or" as used in the specification or claims is not intended to be an exclusive OR.
(21)本願の全体において、例えば、英語におけるa、an及びtheのように、翻訳によって冠詞が追加された場合、これらの冠詞は、文脈から明らかにそうではないことが示されていなければ、複数を含む。 (21) Throughout this application, where translations add articles, such as a, an, and the in English, unless these articles clearly indicate otherwise, Including multiple.
(22)本発明が本明細書中に説明した実施形態に限定されないことは当業者にとって明白である。本発明は、特許請求の範囲の記載に基づいて定まる本発明の趣旨及び範囲を逸脱することなく修正及び変更態様として実施できる。従って、本明細書の記載は、例示的な説明を目的とし、本発明に対して何ら制限的な意味を有さない。また、本明細書に例示した態様から選択された複数の態様を組合せてもよい。 (22) It is obvious to those skilled in the art that the present invention is not limited to the embodiments described herein. The present invention can be implemented as modified and changed modes without departing from the spirit and scope of the present invention defined based on the description of the claims. Therefore, the description of the present specification is for the purpose of exemplification, and has no restrictive meaning to the present invention. In addition, a plurality of modes selected from the modes exemplified in this specification may be combined.
1…サービスシステム、10…動画配信サーバ、11…処理装置、20…ユーザ装置、21…処理装置、22…記憶装置、210…第1キーワード生成部、220A,220B…第2キーワード生成部、220B…第2キーワード生成部、221…抽出部、222…変換部、223…解析部、230A,230B,230C…特定部、240…コメント生成部、KW1…第1キーワード、KW2…第2キーワード、TBLa…キーワードテーブル、TBLb…コメントテーブル、TBLc…行動履歴テーブル、Wx…対象キーワード。 DESCRIPTION OF SYMBOLS 1 ... Service system, 10 ... Video distribution server, 11 ... Processing device, 20 ... User device, 21 ... Processing device, 22 ... Storage device, 210 ... 1st keyword generation part, 220A, 220B ... 2nd keyword generation part, 220B ... second keyword generating unit, 221, ... extracting unit, 222 ... converting unit, 223 ... analyzing unit, 230A, 230B, 230C ... specifying unit, 240 ... comment generating unit, KW1 ... first keyword, KW2 ... second keyword, TBLa ... keyword table, TBLb ... comment table, TBLc ... action history table, Wx ... target keyword.

Claims (5)

  1.  ユーザの音声に基づいて第1キーワードを生成する第1キーワード生成部と、
     画像信号の示す画像から抽出した複数のオブジェクト画像に1対1で対応する複数の第2キーワードを生成する第2キーワード生成部と、
     前記複数の第2キーワードの各々と前記第1キーワードとの関連性の程度に基づいて、前記複数の第2キーワードの中からコメントの対象となる対象キーワードを特定する特定部と、
     前記対象キーワードに関連するコメントを生成するコメント生成部と、
     を備える情報処理装置。
    A first keyword generation unit that generates a first keyword based on a user's voice;
    A second keyword generation unit that generates a plurality of second keywords corresponding to the plurality of object images extracted from the image indicated by the image signal on a one-to-one basis;
    A specifying unit that specifies a target keyword to be a comment target from the plurality of second keywords based on the degree of association between each of the plurality of second keywords and the first keyword;
    A comment generation unit that generates a comment related to the target keyword,
    An information processing apparatus including.
  2.  前記複数の第2キーワードのうち、1の第2キーワードが前記第1キーワードに不一致である場合の関連度の程度と比較して、他の第2キーワードが前記第1キーワードに一致する場合の関連性の程度は高く、
     前記特定部は、前記複数の第2キーワードのうち、前記第1キーワードと一致する第2キーワードを前記対象キーワードとして特定する、請求項1に記載の情報処理装置。
    Of the plurality of second keywords, the degree of association when one second keyword does not match the first keyword, and the degree of association when another second keyword matches the first keyword The degree of sex is high,
    The information processing device according to claim 1, wherein the specifying unit specifies, as the target keyword, a second keyword that matches the first keyword among the plurality of second keywords.
  3.  前記第2キーワード生成部は、
     前記画像信号を解析する解析部と、
     前記解析部の解析結果に基づいて、前記画像信号の示す画像から前記複数のオブジェクト画像を抽出する抽出部と、
     前記複数のオブジェクト画像の各々を、前記複数の第2キーワードのうち、対応する第2キーワードに変換する変換部と、を備える。
     請求項1又は2に記載の情報処理装置。
    The second keyword generation unit,
    An analysis unit for analyzing the image signal,
    An extraction unit that extracts the plurality of object images from the image indicated by the image signal based on the analysis result of the analysis unit;
    A conversion unit that converts each of the plurality of object images into a corresponding second keyword of the plurality of second keywords.
    The information processing apparatus according to claim 1.
  4.  前記特定部は、各第2キーワードと前記第1キーワードとの前記関連性の程度及び前記ユーザの行動履歴に基づいて、前記複数の第2キーワードの中から前記対象キーワードを特定する請求項1から3までのいずれか1項に記載の情報処理装置。 2. The identifying unit identifies the target keyword from the plurality of second keywords based on the degree of association between each second keyword and the first keyword and the action history of the user. The information processing apparatus according to any one of 3 to 3.
  5.  前記特定部は、各第2キーワードと前記第1キーワードとの前記関連性の程度及び前記ユーザのプロファイルに基づいて、前記複数の第2キーワードの中から前記対象キーワードを特定する請求項1から3までのいずれか1項に記載の情報処理装置。 4. The identifying unit identifies the target keyword from the plurality of second keywords based on the degree of association between each second keyword and the first keyword and the profile of the user. The information processing apparatus according to any one of items 1 to 7.
PCT/JP2019/042300 2018-10-30 2019-10-29 Information processing device WO2020090790A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2020553924A JPWO2020090790A1 (en) 2018-10-30 2019-10-29 Information processing equipment

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2018204132 2018-10-30
JP2018-204132 2018-10-30

Publications (1)

Publication Number Publication Date
WO2020090790A1 true WO2020090790A1 (en) 2020-05-07

Family

ID=70463248

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2019/042300 WO2020090790A1 (en) 2018-10-30 2019-10-29 Information processing device

Country Status (2)

Country Link
JP (1) JPWO2020090790A1 (en)
WO (1) WO2020090790A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2010176479A (en) * 2009-01-30 2010-08-12 Fujifilm Corp Image keyword appending apparatus, image search device and method of controlling them
JP2010181461A (en) * 2009-02-03 2010-08-19 Olympus Corp Digital photograph frame, information processing system, program, and information storage medium
JP6396568B1 (en) * 2017-09-19 2018-09-26 ヤフー株式会社 PROVIDING PROGRAM, PROVIDING DEVICE, PROVIDING METHOD, TERMINAL DEVICE, AND INFORMATION PROVIDING DEVICE

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005173739A (en) * 2003-12-08 2005-06-30 Nippon Telegr & Teleph Corp <Ntt> Program for object arrangement

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2010176479A (en) * 2009-01-30 2010-08-12 Fujifilm Corp Image keyword appending apparatus, image search device and method of controlling them
JP2010181461A (en) * 2009-02-03 2010-08-19 Olympus Corp Digital photograph frame, information processing system, program, and information storage medium
JP6396568B1 (en) * 2017-09-19 2018-09-26 ヤフー株式会社 PROVIDING PROGRAM, PROVIDING DEVICE, PROVIDING METHOD, TERMINAL DEVICE, AND INFORMATION PROVIDING DEVICE

Also Published As

Publication number Publication date
JPWO2020090790A1 (en) 2021-12-23

Similar Documents

Publication Publication Date Title
JP5878987B2 (en) Channel navigation on connected media devices by keyword selection
CN110431514B (en) System and method for context driven intelligence
US9058616B2 (en) System and method for providing mobile advertisement
US11436266B2 (en) Method and apparatus for processing information of a terminal
US10783885B2 (en) Image display device, method for driving the same, and computer readable recording medium
CN107895056A (en) A kind of information recommendation method, device, electronic equipment and storage medium
JP6640893B2 (en) Method and apparatus for inputting characters
US20140108555A1 (en) Method and apparatus for identifying network functions based on user data
US11748797B2 (en) System and method for providing recommendations to a target user based upon review and ratings data
US10725620B2 (en) Generating interactive menu for contents search based on user inputs
CN113609308B (en) Knowledge graph construction method and device, storage medium and electronic equipment
US20160110372A1 (en) Method and apparatus for providing location-based social search service
US20160094887A1 (en) Tv program-based shopping guide system and tv program-based shopping guide method
WO2020090790A1 (en) Information processing device
CN107111650A (en) Information processing system and information processor
KR20160053413A (en) Apparatus and method for providing information corresponding contents input into conversation windows
CN114239501A (en) Contract generation method, apparatus, device and medium
CN113780827A (en) Article screening method and device, electronic equipment and computer readable medium
CN112529646A (en) Commodity classification method and device
KR20200125413A (en) Method and apparatus for personalizng movie tribute language
US11800198B2 (en) Information processing apparatus, server, user apparatus, and information processing system
US11941666B2 (en) System and method for automatic matching search advertisement based on product preliminary class
EP4195135A1 (en) Information processing device, information processing method, information processing system, and program
CN113806588B (en) Method and device for searching video
KR102554282B1 (en) Device, method and system for recommending object

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19878450

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2020553924

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19878450

Country of ref document: EP

Kind code of ref document: A1