WO2021240644A1 - Programme, dispositif et procédé de sortie d'informations - Google Patents

Programme, dispositif et procédé de sortie d'informations Download PDF

Info

Publication number
WO2021240644A1
WO2021240644A1 PCT/JP2020/020734 JP2020020734W WO2021240644A1 WO 2021240644 A1 WO2021240644 A1 WO 2021240644A1 JP 2020020734 W JP2020020734 W JP 2020020734W WO 2021240644 A1 WO2021240644 A1 WO 2021240644A1
Authority
WO
WIPO (PCT)
Prior art keywords
information
event
moving image
output
commentary
Prior art date
Application number
PCT/JP2020/020734
Other languages
English (en)
Japanese (ja)
Inventor
健二 山本
教数 塩月
Original Assignee
富士通株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 富士通株式会社 filed Critical 富士通株式会社
Priority to PCT/JP2020/020734 priority Critical patent/WO2021240644A1/fr
Priority to PCT/JP2020/039429 priority patent/WO2021240837A1/fr
Priority to JP2022527479A priority patent/JPWO2021240837A1/ja
Publication of WO2021240644A1 publication Critical patent/WO2021240644A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/235Processing of additional data, e.g. scrambling of additional data or processing content descriptors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring

Definitions

  • the disclosed technology relates to an information output program, an information output device, and an information output method.
  • a commentary additional voice generation device that generates a commentary voice by synthesizing a commentary manuscript (text data) related to the content of a video.
  • This device detects a talking section (voiced sound section), which is a voiced sound section, and a pause section, which is a silent or background sound-only voice section, from the video sound which is the sound of the video. Then, this device converts the commentary sound into the speech speed based on the section length of the pause section, and adds the commentary sound converted into the speech speed to the video sound.
  • voiced sound section voiced sound section
  • a pause section which is a silent or background sound-only voice section
  • the disclosed technology outputs information showing the actual condition and explanation at an appropriate timing in live distribution contents such as sports games without the need for a manuscript prepared in advance and the actual condition and commentator. With the goal.
  • the disclosed technique acquires a sports image including sound information and a moving image, and event information related to an event indicated by each section of the moving image, and based on the acquired event information, the disclosed technique is used. A commentary on the event is generated for each section. Further, the disclosed technique adjusts the output timing of the generated live commentary based on the output timing of at least one section of the sound information and the moving image, and adjusts the output timing of the sound information and the moving image at least. Output with one.
  • One aspect is that it is possible to output information showing the actual condition and explanation at an appropriate timing in live distribution contents such as sports games without the need for a manuscript prepared in advance and the actual condition and commentator. Has.
  • the disclosed technology is applied to the live distribution content of a baseball game.
  • the content is distributed by adding the generated live commentary to the sound information collected at the stadium (hereinafter referred to as "ballpark audio"), the moving image taken at the stadium, or the video including the stadium audio. The case of doing so will be described.
  • the information output system 100 includes an information output device 10, a video distribution system 32, a stats input system 34, and a user terminal 36.
  • the video distribution system 32 shoots a baseball game held at a stadium with a camera and outputs the shot video.
  • the video includes a stadium sound and a moving image composed of a plurality of frames.
  • Time information is associated with each sampling point of the stadium sound and each frame of the moving image, and the stadium sound and the moving image are synchronized based on this time information.
  • the time information is the date and time when the video was shot, the elapsed time from the start of the game, and the like.
  • the stats input system 34 is a system for a person in charge to input stats information about a match while acquiring a video output from the video distribution system 32 and watching the video.
  • the content of the event is input for each event corresponding to one play such as pitching, hitting, running, and defense.
  • a time stamp is added to each event together with the input of the event, for example, manually by the person in charge.
  • the user terminal 36 is a terminal used by a user who uses the service provided by the information output system 100.
  • the user terminal 36 has a function of receiving content distributed from the information output device 10 and a function of outputting at least one of audio and moving images.
  • the user terminal 36 is, for example, a personal computer, a smartphone, a tablet terminal, a mobile phone, a television, a radio, or the like.
  • the information output device 10 generates a commentary on the video and outputs the content with the commentary added to the video acquired from the video distribution system 32.
  • the information output device 10 includes a video acquisition unit 11, an analysis unit 12, a stats acquisition unit 13, a synchronization unit 14, a generation unit 15, and a synthesis unit 16. .. Further, a scene information DB (Database) 21, a stats DB 22, an event DB 23, and a template DB 24 are stored in a predetermined storage area of the information output device 10.
  • the video acquisition unit 11, the analysis unit 12, the stats acquisition unit 13, and the synchronization unit 14 are examples of the disclosure technology acquisition unit. Further, the synthesis unit 16 is an example of an output unit of the disclosed technology.
  • the video acquisition unit 11 acquires the video output from the video distribution system 32, and divides the video into a stadium audio and a moving image.
  • the video acquisition unit 11 passes the divided moving image to the analysis unit 12, and also delivers the acquired video to the synthesis unit 16.
  • the analysis unit 12 acquires scene information for each section corresponding to each event in the moving image by image analysis of the moving image delivered from the video acquisition unit 11. Specifically, the analysis unit 12 detects the switching point of the cut of the camera from the difference in the pixel value between each frame of the moving image, and detects the switching point between the switching points as one section. Further, the analysis unit 12 recognizes the scene shown by the moving image of each section by using the recognition model.
  • the scene can be, for example, a scene that captures a defensive body shape, a scene that captures a batter standing in a turn at bat, a scene that captures the state of a bench, a scene of a running base, a scene of a pick-off ball, a scene of sliding, and the like.
  • the recognition model is machine-learned in advance about the correspondence between the moving image for each section and the label indicating the type of the correct scene shown by the moving image.
  • the analysis unit 12 acquires information such as a ball count, a strike count, an out count (hereinafter referred to as “BSO”), a score, an inning, and a runner's situation from the telop part of the frame image included in each section. This information can be obtained by comparison with a predetermined format, character recognition processing, or the like.
  • BSO will be referred to as "ball count (B) -strike count (S) -out count (O) (for example, 0-0-0)".
  • the analysis unit 12 stores the information acquired for each section in the scene information DB 21 in association with the time information associated with the start frame of the section and the time information associated with the end frame.
  • FIG. 3 shows an example of the scene information DB 21.
  • each row corresponds to the scene information of one section.
  • "Sequence No.” is assigned to each scene information in the order of time information.
  • the "start time” is the time information associated with the start frame of the section
  • the "end time” is the time information associated with the end frame of the section.
  • the "scene” is information indicating the type of the scene recognized by using the recognition model.
  • "inning” and "pre-event BSO” are information acquired from the telop in the frame image of the section. The information included in the scene information is not limited to the above example.
  • the stats acquisition unit 13 acquires stats information for each event input in the stats input system 34 and stores it in the stats DB 22.
  • FIG. 4 shows an example of the stats DB22.
  • each row corresponds to stats information for one event.
  • "Sequence No.” is assigned to each stats information in the order of time information.
  • the stats information includes a "start time” and a "end time” which are times entered by the person in charge as the start time and end time of the event indicated by each stats information.
  • stats information includes "inning" at the time of the event, "batter team” which is the name of the team to which the batter belongs, “batter” which is the name of the batter, and “pitcher team” which is the name of the team to which the pitcher belongs. And the information of "pitcher” which is the name of the pitcher is included.
  • stats information includes "number of pitches in the turn at bat”, “event content”, “direction of hitting” when the event is a hit, and “event result” indicating the number of pitches in the turn at bat for the batter at the time of the event. Contains information.
  • stats information includes "pre-event BSO” which is the BSO before the event and "post-event BSO” based on the event result. The information included in the stats information is not limited to the above example.
  • the synchronization unit 14 generates event information in which the scene information and the stats information are synchronized by associating each of the stats information with the scene information of each section based on the order of the stats information.
  • the scene information is the information acquired by analyzing the moving image
  • the scenes that can be acquired are limited.
  • the time information (start time and end time) of the scene information is the time information associated with each frame of the moving image, it represents the accurate time information of the scene information acquired from each section and also. It is also synchronized with the time information of the stadium voice.
  • the stats information can also acquire detailed information that cannot be acquired by the analysis of moving images.
  • the time information of the stats information is input by the person in charge, there may be cases of inaccuracies, input omissions, etc., and the grain size is coarse, and synchronization with the time information of the stadium voice is guaranteed. It has not been.
  • the time information of the stats information is corrected based on the time information of the scene information, and the information becomes accurate. This makes it possible to generate event information that is more detailed than the scene information and more accurate in the time information than the stats information.
  • the synchronization unit 14 has a sequence number of each of the scene information and the stats information.
  • Event information is generated by associating the stats information with the scene information in which the information of the items common to the scene information and the stats information match, while guaranteeing the context. Common items are, for example, "scene” and “event content”, “pre-event BSO”, and the like.
  • the synchronization unit 14 stores the generated event information in the event DB 23.
  • FIG. 5 shows an example of the event DB 23.
  • each row corresponds to one event information.
  • the event information has an item in which the item of the scene information and the item of the stats information are integrated.
  • the synchronization unit 14 may refer to, for example, sequence No. 14 for scene information in which the corresponding stats information does not exist.
  • Event information is generated by associating the same information as the immediately preceding stats information in order.
  • each of the scene information and the event information is an example of the event information of the disclosed technology
  • the stats information is an example of the external information of the disclosed technology.
  • the generation unit 15 Based on the event information stored in the event DB 23, the generation unit 15 generates a sentence (hereinafter, referred to as “actual commentary”) that is a commentary or a commentary regarding the event indicated by each event information. Specifically, the generation unit 15 selects a template corresponding to each event information from a plurality of live commentary templates stored in the template DB 24, and combines the selected template with the event information to provide a live commentary. To generate.
  • actual commentary a sentence that is a commentary or a commentary regarding the event indicated by each event information.
  • FIG. 6 shows an example of the template DB 24.
  • each row (each record) is information about one template, and by arranging a plurality of templates, a template group corresponding to one event information is formed.
  • the same "template group ID” is assigned as identification information to the templates included in the same template group, and "sequence No. in the template group” is assigned in the order of output.
  • the template DB 24 includes information on a "speaker type" indicating whether each template is intended to be a commentator or a commentator as a speaker.
  • the "template” is a format in which parameters are inserted in a part of a sentence that is a commentary or a commentary.
  • the part of ⁇ > is the part where the parameter is inserted, and the numbers 1, 2, ... Are assigned in ⁇ > in the order of appearance in each template.
  • the item name of the event information is specified as the "parameter type”.
  • "before (or after) B (or S)” defined as the parameter type represents only the corresponding count of BSO.
  • the template DB 24 includes a reproduction time (hereinafter, referred to as “voice time”) when the live commentary generated based on each template is used as voice data.
  • voice time reproduction time
  • the template DB 24 has items of "template”, "voice time”, and “parameter type” corresponding to each of a plurality of languages. You may remember.
  • the generation unit 15 selects the template group corresponding to the event information by using the selection model for selecting the template group suitable for the event information.
  • the selection model is a model in which the correspondence between the event information and the optimum template group for the event indicated by the event information is machine-learned in advance, and the degree of conformity between the target event information and each of the template groups is output. It is a model to do.
  • the generation unit 15 has a “voice time” that is the sum of the voice times of each template included in the predetermined template groups in descending order of the goodness of fit output from the selection model. Is calculated. Then, the generation unit 15 has the highest degree of conformity among the template groups whose voice time is shorter than the time from the "start time” to the "end time” of the event information (hereinafter referred to as "event time”). Select. For example, when the goodness of fit with each template group as shown in FIG. 7 is obtained for the event information having an event time of 20 seconds, the generation unit 15 selects the template group whose template group ID is 3.
  • the generation unit 15 inserts the value of the event information item indicated by the "parameter type" into the ⁇ > part of each template included in the selected template group, and generates a live commentary.
  • the sequence No. of the event DB 23 shown in FIG. Regarding the event information of 5 it is assumed that the template group whose template group ID is 1 as shown in FIG. 6 is selected. Of these, the sequence No. in the template group. Take the template of 6 as an example.
  • the generation unit 15 inserts the event result "foul" in ⁇ 1> of the template, inserts "1" which is B after the event in ⁇ 2>, and S after the event in ⁇ 3>. Insert "1".
  • the generation unit 15 generates a commentary sentence "Foul. This is one ball and one strike.”
  • the generation unit 15 passes the generated commentary to the synthesis unit 16.
  • the synthesizing unit 16 is a commentary on the actual situation delivered from the generation unit 15 based on the output timing of at least one of the stadium audio and the moving image of the video delivered from the video acquisition unit 11 for each section corresponding to each event. Adjust the output timing of the statement. Then, the synthesis unit 16 generates and outputs the content with the live commentary so that the live commentary sentence whose output timing is adjusted is output together with at least one of the stadium sound and the moving image.
  • the synthesizing unit 16 generates a content obtained by synthesizing the stadium audio and the audio data indicating the commentary text as the content with the commentary (see A in FIG. 8).
  • This content is applicable to radio broadcasting and the like.
  • the synthesizing unit 16 generates a content obtained by synthesizing the original video (with stadium sound) or moving image (without stadium sound) and the audio data showing the live commentary as the content with the live commentary.
  • This content can be applied to television broadcasting, Internet video distribution, and the like.
  • the synthesizing unit 16 synthesizes the original video (with stadium sound) or moving image (without stadium sound) and the image data (subtitle) that visualizes the text showing the live commentary as the content with the live commentary. (See B in FIG. 8). This content can also be applied to television broadcasting, Internet video distribution, and the like.
  • the synthesizing unit 16 When synthesizing at least one of the stadium sound and the moving image with the live commentary, the synthesizing unit 16 corresponds to the time information of the moving image or the time information of the stadium sound synchronized with the time information of the moving image and the live commentary. Synchronize with the time information of the event information. Since the time information of the event information matches the time information of the moving image, it is possible to easily synchronize the two.
  • the information output device 10 can be realized by, for example, the computer 40 shown in FIG.
  • the computer 40 includes a CPU (Central Processing Unit) 41, a memory 42 as a temporary storage area, and a non-volatile storage unit 43. Further, the computer 40 includes an input / output device 44 such as an input unit and a display unit, and an R / W (Read / Write) unit 45 that controls reading and writing of data to the storage medium 49. Further, the computer 40 includes a communication I / F (Interface) 46 connected to a network such as the Internet.
  • the CPU 41, the memory 42, the storage unit 43, the input / output device 44, the R / W unit 45, and the communication I / F 46 are connected to each other via the bus 47.
  • the storage unit 43 can be realized by an HDD (Hard Disk Drive), an SSD (Solid State Drive), a flash memory, or the like.
  • the storage unit 43 as a storage medium stores an information output program 50 for causing the computer 40 to function as the information output device 10.
  • the information output program 50 includes a video acquisition process 51, an analysis process 52, a stats acquisition process 53, a synchronization process 54, a generation process 55, and a synthesis process 56.
  • the storage unit 43 has an information storage area 60 in which information constituting each of the scene information DB 21, the stats DB 22, the event DB 23, and the template DB 24 is stored.
  • the CPU 41 reads the information output program 50 from the storage unit 43, expands it into the memory 42, and sequentially executes the processes included in the information output program 50.
  • the CPU 41 operates as the video acquisition unit 11 shown in FIG. 2 by executing the video acquisition process 51. Further, the CPU 41 operates as the analysis unit 12 shown in FIG. 2 by executing the analysis process 52. Further, the CPU 41 operates as the stats acquisition unit 13 shown in FIG. 2 by executing the stats acquisition process 53. Further, the CPU 41 operates as the synchronization unit 14 shown in FIG. 2 by executing the synchronization process 54. Further, the CPU 41 operates as the generation unit 15 shown in FIG. 2 by executing the generation process 55. Further, the CPU 41 operates as the synthesis unit 16 shown in FIG.
  • the CPU 41 reads information from the information storage area 60, and expands each of the scene information DB 21, the stats DB 22, the event DB 23, and the template DB 24 into the memory 42.
  • the computer 40 that has executed the information output program 50 functions as the information output device 10.
  • the CPU 41 that executes the program is hardware.
  • the function realized by the information output program 50 can also be realized by, for example, a semiconductor integrated circuit, more specifically, an ASIC (Application Specific Integrated Circuit) or the like.
  • a semiconductor integrated circuit more specifically, an ASIC (Application Specific Integrated Circuit) or the like.
  • the video distribution system 32 shoots a baseball game held at the stadium with a camera and starts outputting the shot video.
  • the stats input system 34 acquires the video output from the video distribution system 32, and the person in charge inputs the stats information regarding the match while watching the video. Further, in the information output device 10, the information output process shown in FIG. 10 is executed.
  • the information output process is an example of the information output method of the disclosed technology.
  • step S12 the video acquisition unit 11 acquires the video for a predetermined time output from the video distribution system 32, and divides the video into a stadium audio and a moving image. Then, the video acquisition unit 11 passes the divided moving image to the analysis unit 12, and also delivers the acquired video to the synthesis unit 16.
  • step S14 the analysis unit 12 acquires scene information for each section corresponding to each event in the moving image by performing image analysis on the moving image delivered from the video acquisition unit 11. Then, the analysis unit 12 stores the information acquired for each section in the scene information DB 21 in association with the time information associated with the start frame of the section and the time information associated with the end frame. ..
  • step S16 the stats acquisition unit 13 acquires the stats information for each event input in the stats input system 34 and stores it in the stats DB 22.
  • step S18 the synchronization unit 14 sets the sequence No. of each of the scene information and the stats information. While guaranteeing the context, the stats information is associated with the scene information in which the information of the items common to the scene information and the stats information match. As a result, the synchronization unit 14 generates event information. The synchronization unit 14 stores the generated event information in the event DB 23.
  • step S20 the generation unit 15 selects a template corresponding to each event information from the templates of the plurality of live commentary sentences stored in the template DB 24, and combines the selected template with the event information to perform the live commentary. Generate a commentary. The generation unit 15 passes the generated commentary to the synthesis unit 16.
  • step S22 the synthesizing unit 16 synthesizes at least one of the stadium audio and the moving image with the audio data or the image data (subtitles) indicating the live commentary.
  • the synthesis unit 16 synchronizes the time information of the stadium sound synchronized with the time information of the moving image or the time information of the moving image with the time information of the event information corresponding to the live commentary.
  • the synthesizing unit 16 generates and outputs content with a live commentary in which at least one of the stadium audio and the moving image and the output timing of the live commentary are synchronized. Then, the information output process ends.
  • the content with live commentary output from the synthesis unit 16 is distributed to the user terminal 36, and the user who uses the user terminal 36 can view the content with live commentary.
  • the information output device acquires a video of a baseball game including a stadium sound and a moving image, and in the moving image, it is used for each event. Acquire the scene information for each corresponding section. Further, the information output device acquires stats information input externally based on the video. The information output device associates the stats information with the scene information to generate event information having accurate time information for the video and detailed information about the event. Then, the information output device generates a live commentary for each event based on each event information and a template.
  • the information output device adjusts the output timing of the generated live commentary based on the output timing of at least one section of the stadium audio and the moving image, and outputs the output together with at least one of the stadium audio and the moving image.
  • the information output device outputs information showing the actual condition and the explanation at an appropriate timing in the live distribution content such as a sports match without requiring a manuscript prepared in advance and the actual condition and the commentator.
  • the commentator and a commentator since there is no need for a commentator and a commentator, labor costs, distribution equipment, and other costs can be reduced.
  • the template according to the attribute information of the user who views the delivered content with live commentary may be selected.
  • a template for generating a commentary that is biased toward one of the teams, a template for generating a commentary that is biased toward the other team, and a template for a neutral position are prepared.
  • the information of the user's favorite team is acquired as the user's attribute information. Information may be acquired in advance, or information input by the user may be acquired before the start of distribution or during distribution. It may also be estimated based on the user's past viewing history.
  • select a template select a template that generates a commentary that is biased toward the user's favorite team.
  • the live commentary can be flexibly changed according to the user's preference, and the distributed content can be diversified.
  • the attributes of the user are not limited to the favorite team, but may be gender, age, proficiency level for rules, and the like.
  • the case where the live comment is mainly generated is explained, but the same applies to the commentary. Specifically, by preparing a template for commentary using event information and a selection model that associates the event information with the template for appropriate commentary for the event information, the commentary according to the event information is prepared. Can generate statements. In the case where two speakers, a commentator and a commentator, are assumed, when the commentary sentence is converted into voice data, the voice sound may be different depending on the speaker.
  • the event information is generated by associating the stats information with the scene information
  • the accuracy of the image analysis may be improved so that detailed information equivalent to the stats information may be acquired as the scene information.
  • external information other than stats information may be acquired and included in the event information. For example, prepare information such as past competition results, pitcher's ball type, and batting average for each batter's course. Then, such information related to the team, pitcher, and batter at each event can be included in the event information. This makes it possible to generate commentary sentences such as "So far, the batting average of BB players against AA pitchers is 30%" and "This batter is strong in raising the outside angle".
  • information collected from teams and players, information collected from the Internet, etc. can be prepared, and such information related to teams, pitchers, and batters at each event can be included in the event information. good. Then, for example, for event information corresponding to a scene in which a bench or an audience seat is shown, a template that uses such information is selected. In addition, this information may be used to generate a live commentary for a portion where time is available between events. This makes it possible to generate a commentary such as "AA player, yesterday's hit by pitch does not seem to have any effect.”
  • the present invention is not limited to this, and the selection may be made on a rule basis. For example, if the event result includes out, the template A may be selected, and if the pre-event BSO is 3-2-2 and there is a runner, the template B may be selected according to a predetermined rule.
  • a template group having a voice time shorter than the event time has been described, but the present invention is not limited to this.
  • the template group having the highest degree of conformity may be selected without considering the audio time.
  • the audio time of the selected template group is longer than the event time, it may be processed to speed up the playback speed of the audio data showing the live commentary, or some templates included in the template group may be deleted. do it.
  • an example of a baseball game has been described as a sports image, but the application example of the disclosed technology is not limited to this, and can be applied to, for example, soccer, basketball, and the like.
  • the application example of the disclosed technology is not limited to this, and can be applied to, for example, soccer, basketball, and the like.
  • soccer by analyzing the moving image, along with the time information corresponding to each frame, the running speed of the player, the current score (telop information), the distance of the pass, the bias of the positioning of all the players, and the attack direction. Is acquired as scene information.
  • stats information player names, positioning, play contents such as sliding and passes, play results, etc. are acquired.
  • the player name, the running speed of the player, the positioning, the result of the shot, the number of points scored when the shot is made, the content of the play such as steel, no-look pass, rebound, etc. are acquired as scene information.
  • stats information the team name, player name, score, etc. are acquired.
  • the event information may be generated by associating the stats information with the scene information, the template corresponding to the event information may be selected, and the live commentary may be generated.
  • the present invention is not limited to this.
  • the program according to the disclosed technology can also be provided in a form stored in a storage medium such as a CD-ROM, a DVD-ROM, or a USB memory.
  • Information output device 11
  • Video acquisition unit 12 Analysis unit 13
  • Stats acquisition unit 14 Synchronization unit 15
  • Generation unit 16 Synthesis unit 21
  • Scene information DB 22 Stats DB 23
  • Event DB 24 Template DB 32
  • Video distribution system 34 Stats input system 36
  • User terminal 40 Computer 41
  • Memory 43 Storage unit 49
  • Information output program 100 Information output system

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Security & Cryptography (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

Ce dispositif de sortie d'informations : acquiert une image vidéo d'une partie de baseball qui comprend des sons dans un stade et une image dynamique ; acquiert des informations de scène de sections respectives correspondant à des événements dans l'image dynamique ; acquiert des informations statistiques qui ont été saisies de manière externe d'après l'image vidéo ; associe les informations statistiques aux informations de scène afin de générer des informations d'événements qui comprennent des informations précises sur l'heure d'horloge par rapport à l'image vidéo, ainsi que des informations détaillées par rapport à un événement ; génère des commentaires détaillés concernant les événements respectifs d'après les informations d'événements et les modèles ; et, pour chacune des sections, ajuste la synchronisation de sortie des commentaire détaillés générés d'après la synchronisation de sortie des sons du stade et/ou de l'image dynamique, ce qui permet de générer les commentaires conjointement avec les sons du stade et/ou l'image dynamique.
PCT/JP2020/020734 2020-05-26 2020-05-26 Programme, dispositif et procédé de sortie d'informations WO2021240644A1 (fr)

Priority Applications (3)

Application Number Priority Date Filing Date Title
PCT/JP2020/020734 WO2021240644A1 (fr) 2020-05-26 2020-05-26 Programme, dispositif et procédé de sortie d'informations
PCT/JP2020/039429 WO2021240837A1 (fr) 2020-05-26 2020-10-20 Programme, dispositif et procédé de sortie d'informations
JP2022527479A JPWO2021240837A1 (fr) 2020-05-26 2020-10-20

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2020/020734 WO2021240644A1 (fr) 2020-05-26 2020-05-26 Programme, dispositif et procédé de sortie d'informations

Publications (1)

Publication Number Publication Date
WO2021240644A1 true WO2021240644A1 (fr) 2021-12-02

Family

ID=78723236

Family Applications (2)

Application Number Title Priority Date Filing Date
PCT/JP2020/020734 WO2021240644A1 (fr) 2020-05-26 2020-05-26 Programme, dispositif et procédé de sortie d'informations
PCT/JP2020/039429 WO2021240837A1 (fr) 2020-05-26 2020-10-20 Programme, dispositif et procédé de sortie d'informations

Family Applications After (1)

Application Number Title Priority Date Filing Date
PCT/JP2020/039429 WO2021240837A1 (fr) 2020-05-26 2020-10-20 Programme, dispositif et procédé de sortie d'informations

Country Status (2)

Country Link
JP (1) JPWO2021240837A1 (fr)
WO (2) WO2021240644A1 (fr)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2024011105A (ja) * 2022-07-14 2024-01-25 株式会社電通 実況音声リアルタイム生成システム

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005025413A (ja) * 2003-06-30 2005-01-27 Nec Corp コンテンツ処理装置、コンテンツ処理方法及びプログラム
JP2005236541A (ja) * 2004-02-18 2005-09-02 Nippon Telegr & Teleph Corp <Ntt> 野球映像対応付け支援方法及び装置及びプログラム
JP2007184740A (ja) * 2006-01-06 2007-07-19 Nippon Hoso Kyokai <Nhk> コンテンツ送信装置およびコンテンツ出力装置
JP2012039280A (ja) * 2010-08-05 2012-02-23 Nippon Hoso Kyokai <Nhk> 解説放送文作成支援装置及びプログラム
JP2012129980A (ja) * 2010-11-24 2012-07-05 Jvc Kenwood Corp チャプター作成装置、チャプター作成方法、及びチャプター作成プログラム
JP2017151864A (ja) * 2016-02-26 2017-08-31 国立大学法人東京工業大学 データ生成装置
JP2017203827A (ja) * 2016-05-10 2017-11-16 日本放送協会 解説音声再生装置及びそのプログラム
WO2018216729A1 (fr) * 2017-05-24 2018-11-29 日本放送協会 Dispositif de production d'assistance audio, procédé de production d'assistance audio et système de diffusion
JP6472912B1 (ja) * 2018-02-20 2019-02-20 ヤフー株式会社 表示処理プログラム、表示処理装置、および表示処理方法

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005025413A (ja) * 2003-06-30 2005-01-27 Nec Corp コンテンツ処理装置、コンテンツ処理方法及びプログラム
JP2005236541A (ja) * 2004-02-18 2005-09-02 Nippon Telegr & Teleph Corp <Ntt> 野球映像対応付け支援方法及び装置及びプログラム
JP2007184740A (ja) * 2006-01-06 2007-07-19 Nippon Hoso Kyokai <Nhk> コンテンツ送信装置およびコンテンツ出力装置
JP2012039280A (ja) * 2010-08-05 2012-02-23 Nippon Hoso Kyokai <Nhk> 解説放送文作成支援装置及びプログラム
JP2012129980A (ja) * 2010-11-24 2012-07-05 Jvc Kenwood Corp チャプター作成装置、チャプター作成方法、及びチャプター作成プログラム
JP2017151864A (ja) * 2016-02-26 2017-08-31 国立大学法人東京工業大学 データ生成装置
JP2017203827A (ja) * 2016-05-10 2017-11-16 日本放送協会 解説音声再生装置及びそのプログラム
WO2018216729A1 (fr) * 2017-05-24 2018-11-29 日本放送協会 Dispositif de production d'assistance audio, procédé de production d'assistance audio et système de diffusion
JP6472912B1 (ja) * 2018-02-20 2019-02-20 ヤフー株式会社 表示処理プログラム、表示処理装置、および表示処理方法

Also Published As

Publication number Publication date
WO2021240837A1 (fr) 2021-12-02
JPWO2021240837A1 (fr) 2021-12-02

Similar Documents

Publication Publication Date Title
CN107615766B (zh) 用于创建和分配多媒体内容的系统和方法
US10293263B2 (en) Custom content feed based on fantasy sports data
US8121462B2 (en) Video edition device and method
JP5010292B2 (ja) 映像属性情報出力装置、映像要約装置、プログラムおよび映像属性情報出力方法
US7988560B1 (en) Providing highlights of players from a fantasy sports team
US20120087640A1 (en) Information processing apparatus, information processing method, information processing program, and information processing system
JP4621758B2 (ja) コンテンツ情報再生装置、コンテンツ情報再生システム、および情報処理装置
TW201416888A (zh) 場景片段播放系統、方法及其記錄媒體
JP2016107001A (ja) 抽出プログラム、方法、及び装置
JP3923932B2 (ja) 映像要約装置、映像要約方法及びプログラム
US20130222418A1 (en) Providing a Graphic for Video Production
WO2021240644A1 (fr) Programme, dispositif et procédé de sortie d&#39;informations
JP5407708B2 (ja) 撮影映像処理装置、制御方法及びプログラム
US10200764B2 (en) Determination method and device
WO2022249522A1 (fr) Dispositif de traitement d&#39;informations, procédé de traitement d&#39;informations et système de traitement d&#39;informations
JPWO2004012100A1 (ja) コンテンツ要約装置及びコンテンツ要約プログラム
JP2022067478A (ja) 情報処理プログラム、装置、及び方法
CN112233647A (zh) 信息处理设备和方法以及计算机可读存储介质
JP4323937B2 (ja) 映像コメント生成装置及びそのプログラム
JP2016004566A (ja) 提示情報制御装置、方法及びプログラム
US20230179817A1 (en) Information processing apparatus, video distribution system, information processing method, and recording medium
WO2022163023A1 (fr) Dispositif de correction de contenu, serveur de distribution de contenu, procédé de correction de contenu et support d&#39;enregistrement
WO2022074788A1 (fr) Dispositif de traitement d&#39;informations, procédé de traitement d&#39;informations et programme
JP2021087180A (ja) 動画編集装置、動画編集方法及びコンピュータプログラム
JP2014053943A (ja) 撮影映像処理装置、制御方法及びプログラム

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20937573

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20937573

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP