WO2017217612A1 - Procédé de création et de partage de sous-titres d'un contenu vidéo à l'aide d'une fonction à un seul contact tactile - Google Patents

Procédé de création et de partage de sous-titres d'un contenu vidéo à l'aide d'une fonction à un seul contact tactile Download PDF

Info

Publication number
WO2017217612A1
WO2017217612A1 PCT/KR2016/012880 KR2016012880W WO2017217612A1 WO 2017217612 A1 WO2017217612 A1 WO 2017217612A1 KR 2016012880 W KR2016012880 W KR 2016012880W WO 2017217612 A1 WO2017217612 A1 WO 2017217612A1
Authority
WO
WIPO (PCT)
Prior art keywords
video content
time
user
time interval
subtitle
Prior art date
Application number
PCT/KR2016/012880
Other languages
English (en)
Korean (ko)
Inventor
박문수
Original Assignee
주식회사 사이
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 주식회사 사이 filed Critical 주식회사 사이
Publication of WO2017217612A1 publication Critical patent/WO2017217612A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B20/00Signal processing not specific to the method of recording or reproducing; Circuits therefor
    • G11B20/10Digital recording or reproducing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/488Data services, e.g. news ticker
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/278Subtitling

Definitions

  • the present invention relates to a method for generating and sharing captions of video content.
  • Korean Patent Publication No. 10-1419871 discloses a caption editing apparatus and a caption editing method.
  • the caption editing method displays a video preview screen, displays a video clip corresponding to the preview screen, detects a user gesture for the video clip, and enters a caption editing mode for the video clip.
  • the subtitle editing method may detect a user's gesture for inputting a subtitle, deleting a subtitle, moving a subtitle, copying a subtitle, and performing a subtitle editing operation in response to the detected gesture.
  • An object of the present invention is to provide a method for generating and sharing captions of video content.
  • a method of generating and sharing captions of video content including: playing back video content and at least one time interval according to a user's touch operation. setting a time duration, receiving a text corresponding to the at least one time interval from the user, and combining the at least one time interval and text corresponding to the at least one time interval Generating a subtitle of the video content, wherein setting at least one time interval according to a touch operation of the user includes: generating a subtitle of the video content during a playback time of the video content according to a touch input of the user.
  • the touch input time point of the user is set as a start time of the time interval, According to a user's touch release, the user's touch release time is set as an end time of the time interval during the playback time of the video content.
  • the method may further include displaying an object that visualizes an audio signal of the video content adjacent to the played video content, wherein a reference corresponding to a current view is displayed on the object that visualizes the audio signal. It further includes the step of displaying.
  • the method comprises: transmitting subtitles of the video content to a server, receiving at least one subtitle of the video content generated by another user from the server, and the other user And providing the at least one subtitle of the video content generated by the at least one time interval of the subtitle in an editable state.
  • a method of generating and sharing captions of video contents including: playing back video contents and at least one time interval according to a user's touch operation.
  • Setting a subtitle receiving a text corresponding to the at least one time interval from the user, and combining the at least one time interval and the text corresponding to the at least one time interval to subtitle the video content.
  • generating at least one time interval according to the user's touch operation wherein the user's touch input time point of the playback time of the video content is determined according to the user's touch input.
  • the image of the playback time of the video content A predetermined unit time period before the time point from the user's touch input point, and sets the start time of the time interval.
  • the method may further include displaying an object that visualizes an audio signal of the video content adjacent to the played video content, wherein a reference corresponding to a current view is displayed on the object that visualizes the audio signal. It further includes the step of displaying.
  • the method further comprises displaying a unit time manipulation window providing a plurality of predetermined unit times adjacent to the played video content, at least in accordance with the touch manipulation of the user.
  • the setting of one time interval may include using the predetermined unit time selected by the user among the plurality of predetermined unit times, and the predetermined unit time from the touch input time point of the user during the playback time of the video content. The previous time point is set as the start time of the time interval.
  • the setting of the at least one time interval according to the touch operation of the user may include playing the video content by using a predetermined unit time selected by the user's touch input among the plurality of predetermined unit times.
  • a time point before the predetermined unit time from the touch input time point of the user is set as a start time of the time interval.
  • the setting of at least one time interval according to the touch manipulation of the user may be automatically determined based on an audio signal of the video content or at least one time interval set by another user.
  • a time point before the predetermined unit time from the touch input time point of the user among the playback time of the video content is set as the start time of the time interval.
  • the method comprises: transmitting subtitles of the video content to a server, receiving at least one subtitle of the video content generated by another user from the server, and the other user And providing the at least one subtitle of the video content generated by the at least one time interval of the subtitle in an editable state.
  • a user can generate subtitles of video contents in real time and easily through touch operations, and can share and modify subtitles with other users in real time, and through collective intelligence, Subtitles with high reliability can be distributed.
  • FIG. 1 is a schematic block diagram illustrating a configuration of a caption generation and sharing system of video content.
  • FIG. 2 is a schematic block diagram for describing the configuration of the server of FIG. 1.
  • FIG. 3 is a schematic block diagram illustrating the configuration of the client of FIG. 1.
  • FIG. 4 is a flowchart schematically illustrating a method of generating and sharing captions of video content according to an exemplary embodiment of the present invention.
  • FIG. 5 is a flowchart schematically illustrating a method of playing video content by sharing captions of video content according to an exemplary embodiment of the present invention.
  • FIG. 6 is a flowchart schematically illustrating a method of generating a new subtitle of the video content by sharing the subtitles of the video content according to an embodiment of the present invention.
  • FIG. 7 is a schematic flowchart illustrating a subtitle generation interface according to an embodiment of the present invention.
  • FIG. 8 is a diagram schematically illustrating a time section setting screen of a caption generating interface according to an embodiment of the present invention.
  • FIG. 9 is a schematic diagram illustrating a time section setting method of a caption generating interface according to an embodiment of the present invention.
  • FIG. 10 is a schematic flowchart illustrating a subtitle generation interface according to another embodiment of the present invention.
  • FIG. 11 is a diagram schematically illustrating a time section setting screen of a caption generating interface according to another embodiment of the present invention.
  • FIG. 12 is a schematic diagram for explaining a method of setting a time interval of a caption generation interface according to another embodiment of the present invention.
  • FIG. 13 is a schematic diagram illustrating a subtitle generation interface according to another embodiment of the present invention.
  • FIGS. 14 to 15 are schematic diagrams for describing a time interval setting method of a caption generation interface according to another embodiment of the present invention.
  • 16 is a diagram schematically illustrating a time section selection screen of a subtitle generation interface according to embodiments of the present invention.
  • 17 to 18 are schematic diagrams illustrating a time section modification screen of a caption generation interface according to embodiments of the present invention.
  • 19 is a diagram schematically illustrating a text input screen of a caption generating interface according to embodiments of the present invention.
  • FIG. 20 is a diagram schematically illustrating a caption selection screen of a caption generating interface according to embodiments of the present invention.
  • FIG. 21 is a diagram schematically illustrating a time interval sharing screen of a caption generating interface according to embodiments of the present invention.
  • FIG. 22 is a diagram schematically illustrating a time section and a text sharing screen of a caption generating interface according to embodiments of the present invention.
  • spatially relative terms below “, “ beneath “, “ lower”, “ above “, “ upper” It can be used to easily describe a component's correlation with other components. Spatially relative terms are to be understood as including terms in different directions of components in use or operation in addition to the directions shown in the figures. For example, when flipping a component shown in the drawing, a component described as “below” or “beneath” of another component may be placed “above” the other component. Can be. Thus, the exemplary term “below” can encompass both an orientation of above and below. Components may be oriented in other directions as well, so spatially relative terms may be interpreted according to orientation.
  • Subtitle refers to a character displayed on the screen together with the video content when the video content is played back. Subtitles include any text for explanation to the viewer regarding the title, cast, place, time, dialogue, production, etc. of the video content.
  • the caption is configured to include a time duration for displaying the caption in the reproduction time of the video content and text displayed on the screen as the caption corresponding to the time period.
  • FIG. 1 is a schematic block diagram illustrating a configuration of a caption generation and sharing system of video content.
  • a subtitle generation and sharing system of video content includes a server 100 and a plurality of clients 200.
  • the server 100 and the plurality of clients 200 communicate data and / or information with each other via a network.
  • the network is provided in a wired and / or wireless network.
  • the network can have any protocol, scale, or topology.
  • the server 100 stores at least one video content and provides the at least one video content to the client 200 in response to a request of the client 200.
  • the server 100 receives and stores at least one subtitle of at least one video content from the client 200.
  • the server 100 provides the client 200 with at least one subtitle of the at least one video content in response to a request of the client 200.
  • the server 100 provides the client 200 with at least one video content, at least one video content, subtitles of the video content, or subtitles of the video content at the request of the client 200.
  • FIG. 2 is a schematic block diagram for describing the configuration of the server of FIG. 1.
  • the server 100 includes a communication unit 110, a user management unit 120, a content providing unit 130, a subtitle providing unit 140, a user database 150, a content database 160, and a subtitle database. And 170.
  • the communication unit 110 performs wired and / or wireless communication with the client 200.
  • the user manager 120 registers a user and creates a user account.
  • the user manager 120 performs addition, deletion, modification, search, and the like of user account information.
  • the content provider 130 searches for the video content requested by the client 200 from at least one video content stored in the content database 160, and provides the searched video content to the client 200 through the communication unit 110.
  • the caption provider 140 searches for captions of the video content requested by the client 200 from among at least one caption of the at least one video content stored in the caption database 170, and transmits the caption of the searched video content to the communication unit 110.
  • the user database 150 stores user information such as user accounts, user profiles, user logs, and the like.
  • the content database 160 stores at least one video content.
  • at least one video content stored in the content database 160 is divided by country or type.
  • video content such as entertainment, drama, movies, documentaries, courses, and the like may be stored in the content database 160, but is not limited thereto.
  • the caption database 170 stores captions of at least one video content.
  • subtitles stored in subtitle database 170 are classified according to the associated video content.
  • subtitle related information such as the creator of the subtitle, the date of creation, a subtitle language or a description of the subtitle, and the like are stored together in the subtitle database 170.
  • server 100 further includes components not shown in FIG. 2, or does not include some components shown in FIG. 2. May be deformed.
  • Client 200 represents a computer device used by a user.
  • the client 200 may be provided as a mobile device such as a smart phone, a tablet, a personal digital assistant (PDA), but is not limited thereto.
  • Client 200 may be provided to any non-exemplified computer device capable of communicating data and / or information over a network.
  • the client 200 receives only predetermined video content from the server 100, plays the received video content, and generates captions of the video content according to a user's manipulation.
  • the client 200 transmits the caption of the video content generated by the user to the server 100.
  • the client 200 receives a caption of the video content generated by a user different from the predetermined video content from the server 100, and plays the video content by using the caption.
  • the client 200 receives a caption of predetermined video content generated by another user from the server 100, and generates a new caption of the video content according to a user's operation based on the caption.
  • the client 200 transmits a new subtitle of the video content generated by the user to the server 100.
  • the client 200 performs a client program for generating and sharing captions of video content.
  • the client program may be provided in the form of a web browser, a desktop application, a mobile application, and the like, but is not limited thereto.
  • FIG. 3 is a schematic block diagram illustrating the configuration of the client of FIG. 1.
  • the client 200 includes a wireless communication unit 210, an A / V input unit 220, a user input unit 230, a sensing unit 240, an output unit 250, a storage unit 260, and an interface.
  • the unit 270, a controller 280, and a power supply unit 290 are included.
  • the wireless communication unit 210 communicates wirelessly with an external device such as the server 100.
  • the wireless communication unit 210 wirelessly communicates using a wireless communication scheme such as mobile communication, WiBro, Bluetooth, Wi-Fi, Zigbee, ultrasound, infrared, RF, and the like.
  • a wireless communication scheme such as mobile communication, WiBro, Bluetooth, Wi-Fi, Zigbee, ultrasound, infrared, RF, and the like.
  • the wireless communication scheme of the client 200 is not limited to the specific embodiment.
  • the wireless communication unit 210 transmits data and / or information received from the external device to the controller 280, and transmits data and / or information transmitted from the controller 280 to the external device.
  • the wireless communication unit 210 may include a mobile communication module 211 and a short-range communication module 212.
  • the wireless communication unit 210 includes the location information module 213 to obtain location information of the client 200.
  • Location information of the client 200 may be provided from, for example, a GPS positioning system, a WiFi positioning system, a cellular positioning system, or a beacon positioning system, but the present invention is not limited thereto. Location information may be provided from the positioning systems.
  • the wireless communication unit 210 transmits the location information received from the positioning system to the control unit 280.
  • the A / V input unit 220 is for inputting a video or audio signal, and may include a camera module 221 and a microphone module 222.
  • the user input unit 230 receives various information from the user.
  • the user input unit 230 includes input means such as a keypad, a button, a switch, a touch pad, and a jog wheel.
  • input means such as a keypad, a button, a switch, a touch pad, and a jog wheel.
  • a touch screen may be configured.
  • the sensing unit 240 detects the state of the client 200 or the state of the user.
  • the sensing unit 240 may include sensing means such as a touch sensor, a proximity sensor, a pressure sensor, a vibration sensor, a geomagnetic sensor, a gyro sensor, a speed sensor, an acceleration sensor, and a biometric sensor. In some embodiments, the sensing unit 240 is used for user input.
  • the output unit 250 notifies the user of various kinds of information.
  • the output unit 250 outputs information in the form of text, video or audio.
  • the output unit 250 may include a display module 251 and a speaker module 252.
  • the display module 251 is a plasma display panel (PDP), liquid crystal display (LCD), thin film transistor (TFT) LCD, organic light emitting diode (OLED), flexible display, three-dimensional display, electronic ink display, or the present invention. It may be provided in any form well known in the art.
  • the output unit 250 may further include any form of output means well known in the art.
  • the storage unit 260 stores various data and commands.
  • the storage unit 260 stores system software and various applications for the operation of the client 200.
  • the storage unit 260 may be a random access memory (RAM), a read only memory (ROM), an erasable-programmable ROM (EPROM), an electrically EPROM (EEPROM), a flash memory, a hard disk, a removable disk, or a technical field to which the present invention belongs.
  • the interface unit 270 serves as a path to an external device connected to the client 200.
  • the interface unit 270 receives data and / or information from an external device or receives power and transmits the data and / or information to components inside the client 200, or transmits data and / or information inside the client 200 to an external device. Or supply internal power.
  • the interface unit 270 may include, for example, a wired / wireless headset port, a charging port, a wired / wireless data port, a memory card port, a universal serial bus (USB) port, and an identification module. Port may be connected to a connected device, an audio input / output (I / O) port, a video input / output (I / O) port, or the like.
  • the controller 280 controls other components to control the overall operation of the client 200.
  • the controller 280 performs system software and various applications stored in the storage 260.
  • the power supply unit 290 may include a wireless communication unit 210, an A / V input unit 220, a user input unit 230, a sensing unit 240, an output unit 250, a storage unit 260, an interface unit 270, Supply power for the operation of the control unit 280.
  • the power supply unit 290 may include an internal battery.
  • the client 200 further includes components not shown in FIG. 3 or does not include some components shown in FIG. 3. May be deformed.
  • server 100 may be modified to be provided in plural as necessary.
  • a user directly generates captions of predetermined video content or shares captions of the video content generated by another user, and uses the captions to make the video content. You can watch it. Alternatively, the user may modify the subtitles of the video content generated by another user to be more complete. Within the subtitle generation and sharing system of the video content of FIG. 1, at least some subtitles generated by the user may be traded to another user for a fee.
  • FIG. 4 is a flowchart schematically illustrating a method of generating and sharing captions of video content according to an exemplary embodiment of the present invention.
  • the client 200 receives predetermined video content from the server 100.
  • step S320 the client 200 generates a caption of the video content according to a user's manipulation.
  • a method of generating subtitles of specific video content will be described in detail with reference to FIGS. 7 to 15.
  • step S330 the client 200 transmits the caption of the video content generated by the user to the server 100.
  • FIG. 5 is a flowchart schematically illustrating a method of playing video content by sharing captions of video content according to an exemplary embodiment of the present invention.
  • the client 200 receives predetermined video content from the server 100.
  • step S420 the client 200 receives at least one subtitle of the video content generated by another user from the server 100.
  • step S430 the client 200 plays the video content by using the caption selected by the user among the at least one caption received from the server 100.
  • FIG. 6 is a flowchart schematically illustrating a method of generating a new subtitle of the video content by sharing the subtitles of the video content according to an embodiment of the present invention.
  • the client 200 receives predetermined video content from the server 100.
  • step S520 the client 200 receives at least one subtitle of the video content generated by another user from the server 100.
  • the client 200 provides a subtitle selected by a user among the at least one subtitle generated by another user or a time interval of the subtitle in an editable state.
  • the client 200 generates a new caption of the video content according to the user's operation based on the caption or the time interval of the caption.
  • step S550 the client 200 transmits the caption of the video content generated by the user to the server 100.
  • FIG. 7 is a schematic flowchart illustrating a subtitle generation interface according to an embodiment of the present invention.
  • the client 200 plays predetermined video content received from the server 100.
  • the client 200 displays an audio signal object of the video content.
  • An audio signal object visualizes the audio signal of the video content.
  • a reference that corresponds to the current time is displayed on the audio signal object. The user may recognize the audio signal of the current time point through the reference and may refer to it for setting a time section to be described later.
  • the client 200 sets at least one time interval according to the user's touch input and touch release.
  • the client 200 sets the user's touch input time point as a start time of the time interval among the playing time of the video content according to a user's touch input, and releases the user's touch.
  • the touch release time point of the user is set as an end time of the time interval during the playback time of the video content.
  • step S640 the client 200 receives a text corresponding to the at least one time interval from the user.
  • the client 200 In operation S650, the client 200 generates the caption of the video content by combining the at least one time interval with the text. Each time interval has its own text mapped to it.
  • FIG. 8 is a diagram schematically illustrating a time section setting screen of a caption generating interface according to an embodiment of the present invention.
  • the video content 10 is played in the video playback window.
  • the timeline 11 of the video content 10 is displayed adjacent to the video content 10.
  • the timeline 11 represents the total playback time and the current time point of the video content 10.
  • the timeline 11 is disposed inside or outside the video playback window. In some embodiments, when the timeline 11 is placed inside the video playback window, the timeline 11 is displayed overlapped on the video content 10 being played.
  • the audio signal object 12 of the video content 10 is displayed in the audio signal window adjacent to the timeline 11. Below the audio signal object 12, a list of at least one time interval 15 of the video content 10 is displayed in the caption editing window.
  • a reference 13 corresponding to the current time is displayed on the audio signal object 12 .
  • the region 14 corresponding to the time interval 15 set by the user among the audio signals of the video content 10 on the audio signal object 12 is displayed to be distinguished from other regions.
  • the region 14 is distinguished from other regions by using a bounding box as shown in FIG. 8.
  • the area 14 is displayed differently from other areas of different sizes or brightness.
  • the display method of the region 14 is not limited thereto.
  • the user sets the time period 15 via a touch 30 to the audio signal window.
  • the user sets the time period 15 via a touch 30 to the audio signal window and the subtitle editing window. The user may set the time interval 15 through the touch 30 for an arbitrary region.
  • FIG. 9 is a schematic diagram illustrating a time section setting method of a caption generating interface according to an embodiment of the present invention.
  • the first time point is displayed.
  • the time between t1 and the second time point t2 is set as a time interval for displaying subtitles. That is, the first time point t1 is set as the start time of the time section, and the second time point t2 is set as the end time of the time section.
  • FIG. 10 is a schematic flowchart illustrating a subtitle generation interface according to another embodiment of the present invention.
  • the client 200 plays predetermined video content received from the server 100.
  • the client 200 displays an audio signal object of the video content.
  • An audio signal object visualizes the audio signal of the video content.
  • the reference corresponding to the current time point is displayed on the audio signal object.
  • step S730 the client 200 displays the unit time operation window.
  • the unit time operation window is for selecting the unit time of the user.
  • the unit time operation window provides a plurality of predetermined unit times.
  • the client 200 sets at least one time interval according to the user's touch input and unit time.
  • the client 200 sets the user's touch input time point as an end time of the time interval during the playback time of the video content according to the user's touch input, and during the playback time of the video content.
  • a time point before a predetermined unit time from the touch input time point of the user is set as a start time of the time interval.
  • step S750 the client 200 receives a text corresponding to the at least one time interval from the user.
  • the client 200 In operation S760, the client 200 generates the caption of the video content by combining the at least one time interval with the text.
  • FIG. 11 is a diagram schematically illustrating a time section setting screen of a caption generating interface according to another embodiment of the present invention.
  • the video content 10 is played in the video playback window.
  • the timeline 11 of the video content 10 is displayed adjacent to the video content 10.
  • the timeline 11 represents the total playback time and the current time point of the video content 10.
  • the audio signal object 12 of the video content 10 is displayed in the audio signal window adjacent to the timeline 11.
  • the unit time operation window 16 is displayed adjacent to the audio signal object 12.
  • the unit time operation window 16 provides a plurality of predetermined unit times.
  • the user may set a plurality of predetermined unit times provided by the unit time manipulation window 16.
  • the unit time operation window 16 may provide unit times such as 0.3 seconds, 1 second, 2 seconds, 4 seconds, AUTO, and the like, but is not limited thereto.
  • a list of at least one time interval 15 of the video content 10 is displayed in the subtitle editing window at the bottom of the unit time operation window 16.
  • a reference 13 corresponding to the current time is displayed on the audio signal object 12 .
  • the region 14 corresponding to the time interval 15 set by the user among the audio signals of the video content 10 on the audio signal object 12 is displayed to be distinguished from other regions.
  • the user selects the unit time for setting the time interval through the touch 30 on the unit time operation window 16.
  • the client 200 sets the time interval 15 using the unit time selected by the user among a plurality of predetermined unit times on the unit time operation window 16.
  • the user sets the time period 15 via a touch 30 to the audio signal window. In some embodiments, the user sets the time period 15 via a touch 30 to the audio signal window and the subtitle editing window. The user may set the time interval 15 through the touch 30 for an arbitrary region. The user selects a specific unit time by inputting the touch 30 to the unit time operation window 16 before setting the time section 15.
  • the user sets the time interval 15 via a touch 30 to the unit time manipulation window 16.
  • the touch input is not only for selecting a specific unit time but also for setting an end time of a time interval for displaying a subtitle.
  • the client 200 when the user selects AUTO, the client 200 automatically determines the unit time. In some embodiments, the client 200 automatically determines the unit time based on the audio signal of the video content 10. In some embodiments, the client 200 automatically determines the unit time based on subtitles generated by other users (at least one time interval set by another user). In some embodiments, the client 200 analyzes the unit time frequently used by the user, and automatically determines an appropriate unit time according to the analysis result.
  • FIG. 12 is a schematic diagram for explaining a method of setting a time interval of a caption generation interface according to another embodiment of the present invention.
  • the time between the first time point t1 and the second time point t2 displays a caption. It is set to a time interval to.
  • the first time point t1 is determined as a time point that is before a predetermined unit time from the second time point t2.
  • the second time point t2 is set as the end time of the time period, and then the first time point t1 is set as the start time of the time period. 12 illustrates a case where 4 seconds is selected as a unit time, for example.
  • FIG. 13 is a schematic diagram illustrating a subtitle generation interface according to another embodiment of the present invention.
  • the subtitle providing unit 140 of the server 100 includes a machine learning module 141.
  • the machine learning module 141 learns the audio signal of the video content.
  • machine learning module 141 learns subtitles (at least one time interval set by another user) generated by another user.
  • the machine learning module 141 may learn a plurality of subtitles related to one video content or learn a plurality of subtitles having different target video contents.
  • Reference numeral 161 denotes predetermined video content
  • reference numeral 171 denotes a plurality of subtitles related to the video content.
  • the machine learning module 141 learns at least one time interval setting pattern of the user. According to the learning result, the machine learning module 141 predicts in real time an optimal time interval for displaying captions of predetermined video content. Similarly, the machine learning module 141 may predict in real time an optimal unit time for setting the start time of the time interval.
  • FIGS. 14 to 15 are schematic diagrams for describing a time interval setting method of a caption generation interface according to another embodiment of the present invention.
  • the client 200 receives information on the optimal time interval predicted by using machine learning from the server 100 and provides the information to the user for reference in the process of setting the time interval.
  • an area 17 corresponding to an optimal time interval predicted using machine learning is displayed on the audio signal object 12 before a user's touch input for setting the time interval is displayed. do.
  • the user may set a start time and an end time of a time interval for displaying a subtitle with reference to the area 17. Naturally, the user may set the time interval differently from the optimal time interval predicted using machine learning.
  • the client 200 receives information about the optimal unit time predicted by using machine learning from the server 100 and provides the information to the user for reference in the process of setting the time interval. .
  • an optimal unit time is provided in the unit time operation window 18.
  • the unit time manipulation window 18 provides one or a plurality of optimal unit times.
  • the unit time manipulation window 18 simultaneously provides the unit time set by the user and the optimal unit time predicted using machine learning. The user may select the optimal unit time with reference to the unit time operation window 18. Naturally, without using the optimal unit time, the user may select the unit time set by the user.
  • an area 17 corresponding to the optimal time interval predicted by using machine learning may be displayed on the audio signal object 12.
  • 16 is a diagram schematically illustrating a time section selection screen of a subtitle generation interface according to embodiments of the present invention.
  • any one time interval 15 of the list of at least one time interval 15 of the video content 10 is selected by the user, the time of the video content 10 in the video playback window.
  • the image corresponding to the start time of the section 15 is displayed.
  • an area 14 corresponding to the time interval 15 of the audio signal of the video content 10 is displayed on the audio signal object 12.
  • 17 to 18 are schematic diagrams illustrating a time section modification screen of a caption generation interface according to embodiments of the present invention.
  • the user adjusts the start time or end time of the time interval 15 by touching the area 14 corresponding to the time interval 15 on the audio signal object 12.
  • the user may adjust a start time or end time of the time interval 15 by inputting a predetermined gesture (eg, drag) after a touch input to the area 14, but is not limited thereto. .
  • a predetermined gesture eg, drag
  • the time adjustment object 19 is displayed in the audio signal window adjacent to the audio signal object 12.
  • the time adjustment object 19 may be disposed adjacent to the left and right of the audio signal object 12, but is not limited thereto.
  • the user may adjust the start time or the end time of the time interval 15 through a touch on the time adjustment object 19.
  • 19 is a diagram schematically illustrating a text input screen of a caption generating interface according to embodiments of the present invention.
  • a user's primary touch 30 is provided for any one time interval 15 of a list of at least one time interval 15 of the video content 10.
  • a video corresponding to the start time of the time section 15 of the video content 10 is displayed in the video playback window, and the user's second touch 30 is provided for the time section 15, the user
  • the text 20 corresponding to the time interval 15 may be input.
  • the text input window is overlapped and disposed on the video playback window.
  • a text input window is disposed adjacent to the time interval 15.
  • FIG. 20 is a diagram schematically illustrating a caption selection screen of a caption generating interface according to embodiments of the present invention.
  • the menu window 21 is displayed adjacent to the video playing window.
  • the menu window 21 may provide a plurality of menus for selecting subtitles, generating subtitles, and the like, but is not limited thereto.
  • the caption selection is selected, a list of at least one caption 22 of the video content 10 is displayed in the caption selection window at the bottom of the menu window 21.
  • subtitle related information such as the creator, date of creation, subtitle language or description of the subtitle, etc., of each subtitle 22 is displayed together within the subtitle selection window.
  • the selected subtitles are overlapped and displayed on the video playback window.
  • FIG. 21 is a diagram schematically illustrating a time interval sharing screen of a caption generating interface according to embodiments of the present invention.
  • the user selects editing of any one of the subtitles 22 from the list of at least one subtitles 22 of the given video content 10.
  • the caption generation screen described with reference to FIG. 8 is displayed, and the caption 22 selected by the user is provided in an editable state.
  • the user may share only at least one time interval 15 of the subtitle 22.
  • an area 14 corresponding to at least one time interval 15 set by another user is displayed, and set by another user in the subtitle editing window at the bottom of the audio signal object 12.
  • a list of at least one time period 15 is displayed.
  • the user may adjust the start time or end time of the time interval 15 set by another user.
  • the user may delete the time interval 15 set by another user.
  • the user may additionally set a time period 15 not set by another user.
  • FIG. 22 is a diagram schematically illustrating a time section and a text sharing screen of a caption generating interface according to embodiments of the present invention.
  • the user selects editing of any one of the subtitles 22 from the list of at least one subtitles 22 of the given video content 10.
  • the caption generation screen described with reference to FIG. 8 is displayed, and the caption 22 selected by the user is provided in an editable state.
  • the user may share both the at least one time interval 15 of the subtitle 22 and the corresponding text 23.
  • a list of at least one time interval 15 and corresponding text 23 set by another user in the subtitle editing window is displayed.
  • the user may adjust the start time or end time of the time interval 15 set by another user.
  • the user may delete the time interval 15 set by another user.
  • the user may additionally set a time period 15 not set by another user.
  • the user can modify the text 23 input by another user.
  • the steps of a method or algorithm described in connection with an embodiment of the invention may be implemented directly in a hardware module, in a software module executed by hardware, or by a combination thereof.
  • Software modules may include random access memory (RAM), read only memory (ROM), erasable programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), flash memory, hard disk, removable disk, CD-ROM, or It may reside in any form of computer readable recording medium well known in the art.

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

La présente invention concerne un procédé de création et de partage de sous-titres d'un contenu vidéo à l'aide d'une fonction à un seul contact tactile. Le procédé, qui est un procédé mis en œuvre par un ordinateur, comprend : une étape de lecture d'un contenu vidéo ; une étape de réglage d'au moins une durée temporelle selon une opération tactile effectuée par un utilisateur ; une étape de réception, en provenance de l'utilisateur, d'un texte correspondant à la ou aux durées temporelles ; et une étape de création de sous-titres du contenu vidéo par combinaison de la ou des secondes durées temporelles et du texte correspondant à celle(s)-ci.
PCT/KR2016/012880 2016-06-17 2016-11-10 Procédé de création et de partage de sous-titres d'un contenu vidéo à l'aide d'une fonction à un seul contact tactile WO2017217612A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR10-2016-0075799 2016-06-17
KR20160075799 2016-06-17

Publications (1)

Publication Number Publication Date
WO2017217612A1 true WO2017217612A1 (fr) 2017-12-21

Family

ID=60663103

Family Applications (2)

Application Number Title Priority Date Filing Date
PCT/KR2016/012890 WO2017217613A1 (fr) 2016-06-17 2016-11-10 Procédé de création et de partage de sous-titres d'un contenu vidéo à l'aide d'un apprentissage automatique
PCT/KR2016/012880 WO2017217612A1 (fr) 2016-06-17 2016-11-10 Procédé de création et de partage de sous-titres d'un contenu vidéo à l'aide d'une fonction à un seul contact tactile

Family Applications Before (1)

Application Number Title Priority Date Filing Date
PCT/KR2016/012890 WO2017217613A1 (fr) 2016-06-17 2016-11-10 Procédé de création et de partage de sous-titres d'un contenu vidéo à l'aide d'un apprentissage automatique

Country Status (1)

Country Link
WO (2) WO2017217613A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110475146A (zh) * 2019-09-05 2019-11-19 珠海市杰理科技股份有限公司 字幕矫正方法、装置及智能音箱

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113593567B (zh) * 2021-06-23 2022-09-09 荣耀终端有限公司 视频声音转文本的方法及相关设备

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004336668A (ja) * 2003-05-12 2004-11-25 National Institute Of Information & Communication Technology 字幕作成用管理サーバおよび分散型字幕番組制作システム
CN104079838A (zh) * 2014-07-08 2014-10-01 丽水桉阳生物科技有限公司 一种具有财经数据字幕制播功能的字幕机
WO2015088196A1 (fr) * 2013-12-09 2015-06-18 넥스트리밍(주) Appareil et procédé d'édition de sous-titres
KR20160024002A (ko) * 2014-08-21 2016-03-04 삼성전자주식회사 비쥬얼 사운드 이미지를 제공하는 방법 및 이를 구현하는 전자 장치
JP2016509408A (ja) * 2013-01-15 2016-03-24 ヴィキ, インク.Viki, Inc. メディアにキャプションを付けるシステム及び方法

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20000049475A (ko) * 2000-03-27 2000-08-05 김기찬 인터넷 멀티미디어형 자막방송 솔루션 장치 개발 및 서비스
KR20090124240A (ko) * 2008-05-29 2009-12-03 주식회사 케이티테크 자막 편집 장치 및 그 방법

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2004336668A (ja) * 2003-05-12 2004-11-25 National Institute Of Information & Communication Technology 字幕作成用管理サーバおよび分散型字幕番組制作システム
JP2016509408A (ja) * 2013-01-15 2016-03-24 ヴィキ, インク.Viki, Inc. メディアにキャプションを付けるシステム及び方法
WO2015088196A1 (fr) * 2013-12-09 2015-06-18 넥스트리밍(주) Appareil et procédé d'édition de sous-titres
CN104079838A (zh) * 2014-07-08 2014-10-01 丽水桉阳生物科技有限公司 一种具有财经数据字幕制播功能的字幕机
KR20160024002A (ko) * 2014-08-21 2016-03-04 삼성전자주식회사 비쥬얼 사운드 이미지를 제공하는 방법 및 이를 구현하는 전자 장치

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110475146A (zh) * 2019-09-05 2019-11-19 珠海市杰理科技股份有限公司 字幕矫正方法、装置及智能音箱

Also Published As

Publication number Publication date
WO2017217613A1 (fr) 2017-12-21

Similar Documents

Publication Publication Date Title
WO2012133982A1 (fr) Dispositif de traitement d'image et procédé de commande du dispositif de traitement d'image
WO2016093506A1 (fr) Terminal mobile et procédé de commande associé
WO2013054957A1 (fr) Appareil de contrôle d'interface d'entrée et procédé associé
WO2011078540A2 (fr) Dispositif mobile et procédé de commande correspondant pour sortie externe dépendant d'une interaction d'utilisateur sur la base d'un module de détection d'image
WO2014017858A1 (fr) Appareil de terminal utilisateur et procédé de commande associé
WO2020162709A1 (fr) Dispositif électronique pour la fourniture de données graphiques basées sur une voix et son procédé de fonctionnement
WO2012133983A1 (fr) Traitement d'image dans un dispositif d'affichage d'image monté sur véhicule
WO2013103275A1 (fr) Procédé et appareil pour mettre en œuvre un système multi-vision par utilisation de multiples terminaux portables
WO2014058250A1 (fr) Terminal utilisateur, serveur fournissant un service de réseau social et procédé de fourniture de contenus
WO2012133981A1 (fr) Dispositif d'affichage d'image et procédé de commande correspondant
WO2019112342A1 (fr) Appareil de reconnaissance vocale et son procédé de fonctionnement
EP3097490A1 (fr) Montre intelligente, dispositif d'affichage et son procédé de commande
WO2015046900A1 (fr) Procédé et dispositif de partage de contenu
WO2016003010A1 (fr) Appareil de traitement d'image numérique et son procédé de commande
WO2017188585A1 (fr) Dispositif d'affichage et son procédé de fonctionnement
WO2016089047A1 (fr) Procédé et dispositif de distribution de contenu
WO2018056617A1 (fr) Dispositif vestimentaire et procédé de fourniture de gadget logiciel correspondant
WO2015072803A1 (fr) Terminal et procédé de commande de terminal
WO2020197012A1 (fr) Appareil d'affichage et procédé de commande de celui-ci
WO2018169374A1 (fr) Dispositif electronique et procédé de commande associé
WO2018093160A2 (fr) Dispositif d'affichage, système et support d'enregistrement
WO2017217612A1 (fr) Procédé de création et de partage de sous-titres d'un contenu vidéo à l'aide d'une fonction à un seul contact tactile
WO2012081787A1 (fr) Appareil de traitement d'images de terminal mobile et procédé associé
WO2014061905A1 (fr) Système permettant d'obtenir un signet basé sur le mouvement et la voix, et procédé s'y rapportant
WO2018056587A1 (fr) Appareil électronique et son procédé de commande

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16905596

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 16905596

Country of ref document: EP

Kind code of ref document: A1