KR20170047547A - Display device and method for controlling the same - Google Patents
Display device and method for controlling the same Download PDFInfo
- Publication number
- KR20170047547A KR20170047547A KR1020150147779A KR20150147779A KR20170047547A KR 20170047547 A KR20170047547 A KR 20170047547A KR 1020150147779 A KR1020150147779 A KR 1020150147779A KR 20150147779 A KR20150147779 A KR 20150147779A KR 20170047547 A KR20170047547 A KR 20170047547A
- Authority
- KR
- South Korea
- Prior art keywords
- output
- information
- subtitle
- time
- voice
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 60
- 230000000007 visual effect Effects 0.000 claims description 69
- 238000001514 detection method Methods 0.000 claims description 15
- 238000013519 translation Methods 0.000 claims description 13
- 238000004891 communication Methods 0.000 description 41
- 238000012545 processing Methods 0.000 description 28
- 230000006870 function Effects 0.000 description 19
- 238000010586 diagram Methods 0.000 description 15
- 230000008569 process Effects 0.000 description 13
- 230000000694 effects Effects 0.000 description 12
- 230000005236 sound signal Effects 0.000 description 11
- 230000008901 benefit Effects 0.000 description 7
- 238000006243 chemical reaction Methods 0.000 description 7
- 238000010295 mobile communication Methods 0.000 description 7
- 230000001360 synchronised effect Effects 0.000 description 7
- 238000007726 management method Methods 0.000 description 6
- 230000005540 biological transmission Effects 0.000 description 5
- 230000008859 change Effects 0.000 description 3
- 239000012141 concentrate Substances 0.000 description 3
- 238000000354 decomposition reaction Methods 0.000 description 3
- 238000011161 development Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 239000004973 liquid crystal related substance Substances 0.000 description 3
- 230000033001 locomotion Effects 0.000 description 3
- 239000000243 solution Substances 0.000 description 3
- 230000003068 static effect Effects 0.000 description 3
- 239000002131 composite material Substances 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 101000591286 Homo sapiens Myocardin-related transcription factor A Proteins 0.000 description 1
- 102100034099 Myocardin-related transcription factor A Human genes 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 230000005684 electric field Effects 0.000 description 1
- 230000005611 electricity Effects 0.000 description 1
- 230000005672 electromagnetic field Effects 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 239000010408 film Substances 0.000 description 1
- 230000020169 heat generation Effects 0.000 description 1
- 230000000977 initiatory effect Effects 0.000 description 1
- 238000002347 injection Methods 0.000 description 1
- 239000007924 injection Substances 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 230000006386 memory function Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003387 muscular Effects 0.000 description 1
- 230000010355 oscillation Effects 0.000 description 1
- 230000001151 other effect Effects 0.000 description 1
- 230000037361 pathway Effects 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 238000007639 printing Methods 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000035807 sensation Effects 0.000 description 1
- 239000004984 smart glass Substances 0.000 description 1
- 238000005507 spraying Methods 0.000 description 1
- 239000010409 thin film Substances 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/488—Data services, e.g. news ticker
- H04N21/4886—Data services, e.g. news ticker for displaying a ticker, e.g. scrolling banner for news, stock exchange, weather data
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/278—Subtitling
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N2201/00—Indexing scheme relating to scanning, transmission or reproduction of documents or the like, and to details thereof
- H04N2201/024—Indexing scheme relating to scanning, transmission or reproduction of documents or the like, and to details thereof deleted
- H04N2201/02406—Arrangements for positioning elements within a head
- H04N2201/02408—Translational positioning
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
Description
The present invention relates to a display device and a control method thereof.
Mobile devices such as a smart phone and a tablet PC have attracted attention in addition to a standing device such as a personal computer (PC) and a television (TV). Fixed devices and mobile devices originally developed in their respective domains, but the area has become obscure due to the recent boom of digital convergence.
In recent years, there have been many cases in which a subtitle is output using a subtitle file stored in a memory when a video stored in a memory is reproduced.
However, the present technology has a disadvantage that the user has to manually synchronize the audio data when the subtitles are not synchronized with the audio data currently being output. Thus, the way to automatically synchronize subtitles remains a matter of debate.
SUMMARY OF THE INVENTION The present invention has been made in order to solve the above-described situations and problems, and it is an object of the present invention to automatically control the synchronization of subtitles.
Another object of the present invention is to recognize, when a plurality of speakers are present in contents output from the display device, whether the subtitle being output is a subtitle related to a certain speaker and give different visual effects to the subtitle associated with each speaker .
The technical problem to be solved by the present invention is not limited to the above-described technical problems and other technical problems which are not mentioned can be clearly understood by those skilled in the art from the following description .
This document discloses various embodiments (s) of digital devices and processing methods in the digital devices.
A display device according to an embodiment of the present invention includes a memory for storing first information on a subtitle associated with a specific content and a reproduction time of the subtitle, a display unit for displaying video data of the specific content, And a controller for controlling the display unit so that the subtitles are output together with the video data using the first information, wherein the control unit controls the display unit so that audio is output from the audio data of the specific content being output, Recognizing second information on the detected time, recognizing third information on a time when the subtitles are output using the first information, determining, based on the comparison result of the second information and the third information, Adjusts the sync of the output subtitle.
According to another aspect of the present invention, there is provided a method of controlling a display device, the method comprising: displaying video data of a specific content; outputting audio data of the specific content; Outputting a subtitle associated with the specific content stored in the memory together with the video data, recognizing second information on a time when audio is detected in the audio data of the specific content being output, Recognizing third information on a time at which the caption is output using the information, and adjusting the sink of the caption being output according to the comparison result of the second information and the third information.
The technical solutions obtained by the present invention are not limited to the above-mentioned solutions, and other solutions not mentioned are clearly described to those skilled in the art from the following description. It can be understood.
Effects of the display device and the control method according to the present invention will be described as follows. According to an embodiment of the present invention, the user can automatically adjust the synchronization of subtitles on the TV without having to manually adjust the sync of the subtitles on the TV, so that the user can concentrate on the content being output.
According to an embodiment of the present invention, when there are a plurality of speakers in the output content, it is possible to recognize which subtitle is related to which speaker and output a different visual effect to each subtitle associated with each speaker, Has the advantage that the content can be easily viewed and understood.
The effects obtained by the present invention are not limited to the above-mentioned effects, and other effects not mentioned can be clearly understood by those skilled in the art from the following description will be.
1 is a schematic diagram illustrating a service system including a digital device according to an exemplary embodiment of the present invention.
2 is a block diagram illustrating a digital device according to an embodiment of the present invention
3 is a block diagram illustrating a digital device according to another embodiment of the present invention
4 is a block diagram illustrating a digital device according to another embodiment of the present invention
5 is a block diagram illustrating a detailed configuration of the control unit of FIGS. 2 to 4 according to an embodiment of the present invention
Figure 6 illustrates an input means coupled to the digital device of Figures 2 through 4 according to one embodiment of the present invention
7 is a block diagram illustrating configuration modules of a display device according to an embodiment of the present invention.
FIG. 8 is a flowchart illustrating an example of a method for adjusting the synchronization of subtitles in a display device according to an exemplary embodiment of the present invention. Referring to FIG.
9 and 10 are diagrams for explaining an example of a method of adjusting the synchronization of the caption when the subtitle output from the display device according to the embodiment of the present invention recognizes that the audio data being output is out of sync with the audio data being output FIG.
11 is a flowchart for explaining an example of a method of analyzing a voice of a recognized voice to give a visual effect to an outputting subtitle when a voice is recognized in audio data being output from a display device according to an embodiment of the present invention .
12 is a view for explaining an example of a method of analyzing a voice of a recognized voice and giving a visual effect to a subtitle being output when a voice is recognized in audio data being outputted from a display device according to an embodiment of the present invention FIG.
13 is a view for explaining an example of a method of mapping a visual effect assigned in accordance with a grammar in a display device according to an embodiment of the present invention to a memory.
FIG. 14 is a flowchart for explaining an example of a method of converting a voice of audio data being outputted from a display device into text and adjusting the synchronization of the caption using the translation result of the converted text according to an embodiment of the present invention.
Hereinafter, various embodiments (s) of a digital device according to the present invention and a method of processing application data in the digital device will be described in detail with reference to the drawings.
The suffix "module "," part ", and the like for components used in the present specification are given only for ease of specification, and both may be used as needed. Also, even when described in ordinal numbers such as " 1st ", "2nd ", and the like, it is not limited to such terms or ordinal numbers.
In addition, although the terms used in the present specification have been selected from the general terms that are widely used in the present invention in consideration of the functions according to the technical idea of the present invention, they are not limited to the intentions or customs of the artisan skilled in the art, It can be different. However, in certain cases, some terms are arbitrarily selected by the applicant, which will be described in the related description section. Accordingly, it should be understood that the term is to be interpreted based not only on its name but on its practical meaning as well as on the contents described throughout this specification.
It is to be noted that the contents of the present specification and / or drawings are not intended to limit the scope of the present invention.
The term "digital device" as used herein refers to a device that transmits, receives, processes, and outputs data, content, service, And includes all devices that perform at least one or more. The digital device can be paired or connected (hereinafter, referred to as 'pairing') with another digital device, an external server, or the like through a wire / wireless network, Can be transmitted / received. At this time, if necessary, the data may be appropriately converted before the transmission / reception. The digital device may be a standing device such as a network TV, a Hybrid Broadcast Broadband TV (HBBTV), a Smart TV, an IPTV (Internet Protocol TV), a PC (Personal Computer) And a mobile device or handheld device such as a PDA (Personal Digital Assistant), a smart phone, a tablet PC, a notebook, and the like. In order to facilitate understanding of the present invention and to facilitate the description of the present invention, FIG. 2, which will be described later, describes a digital TV, and FIG. 3 illustrates and describes a mobile device as an embodiment of a digital device. In addition, the digital device described in this specification may be a configuration having only a panel, a configuration such as a set-top box (STB), a device, a system, etc. and a set configuration .
The term "wired / wireless network" as used herein collectively refers to communication networks that support various communication standards or protocols for pairing and / or data transmission / reception between digital devices or digital devices and external servers. Such a wired / wireless network includes all of the communication networks to be supported by the standard now or in the future, and is capable of supporting one or more communication protocols therefor. Such a wired / wireless network includes, for example, a USB (Universal Serial Bus), a Composite Video Banking Sync (CVBS), a Component, an S-Video (Analog), a DVI (Digital Visual Interface) A communication standard or protocol for a wired connection such as an RGB or a D-SUB, a Bluetooth standard, a radio frequency identification (RFID), an infrared data association (IrDA), an ultra wideband (UWB) (ZigBee), DLNA (Digital Living Network Alliance), WLAN (Wi-Fi), Wibro (Wireless broadband), Wimax (World Interoperability for Microwave Access), HSDPA (High Speed Downlink Packet Access) A long term evolution (LTE-Advanced), and Wi-Fi direct, and a communication standard or protocol for the network.
In addition, when the term is simply referred to as a digital device in this specification, the meaning may mean a fixed device or a mobile device depending on the context, and may be used to mean both, unless specifically stated otherwise.
Meanwhile, a digital device is an intelligent device that supports, for example, a broadcast receiving function, a computer function or a support, at least one external input, and the like. The digital device may be an e-mail, web browsing, Banking, game, application, and so on. In addition, the digital device may include an interface for supporting at least one input or control means (hereinafter, " input means ") such as a handwriting input device, a touch- .
In addition, the digital device can use a standardized general-purpose OS (Operating System), but in particular, the digital device described in this specification uses the Web OS as an embodiment. Therefore, a digital device can handle adding, deleting, amending, and updating various services or applications on a general-purpose OS kernel or a Linux kernel. And through which a more user-friendly environment can be constructed and provided.
Meanwhile, the above-described digital device can receive and process an external input. The external input is connected to an external input device, that is, the digital device, through the wired / wireless network, An input means or a digital device. For example, the external input may be a game device such as a high-definition multimedia interface (HDMI), a playstation or an X-Box, a smart phone, a tablet PC, a pocket photo devices such as digital cameras, printing devices, smart TVs, Blu-ray device devices and the like.
In addition, the term "server" as used herein refers to a digital device or system that supplies data to or receives data from a digital device, that is, a client, and may be referred to as a processor do. The server provides a Web server, a portal server, and advertising data for providing a web page, a web content or a web content or a web service, An advertising server, a content server for providing content, an SNS server for providing a social network service (SNS), a service server provided by a manufacturer, a video on demand (VoD) server, A service server providing a Multichannel Video Programming Distributor (MVPD) for providing a streaming service, a pay service, and the like.
In addition, in the following description for convenience of explanation, only the application is described in the context of the present invention, and the meaning may include not only the application but also the service based on the context and the like. In addition, the application may refer to a web application according to the webOS platform.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS Hereinafter, the present invention will be described in detail with reference to the accompanying drawings.
1 is a schematic diagram illustrating a service system including a digital device according to an exemplary embodiment of the present invention.
1, a service system includes a
The
The
The
The above-described
The
The
Meanwhile, the
In addition, the
In FIG. 1, the
2 is a block diagram illustrating a digital device according to an exemplary embodiment of the present invention.
The digital device described herein corresponds to the
The
The
The TCP /
The
The
The
The audio /
The application manager may include, for example, the
The
The
The
The
The
The
The SI &
The SI &
Meanwhile, the
3 is a block diagram illustrating a digital device according to another embodiment of the present invention.
If the above-described Fig. 2 is described with an example of a digital device as a fixing device, Fig. 3 shows a mobile device as another embodiment of a digital device.
3, the
Hereinafter, each component will be described in detail.
The
The
The broadcast-related information may mean information related to a broadcast channel, a broadcast program, or a broadcast service provider. The broadcast-related information may also be provided through a mobile communication network. In this case, it may be received by the
The broadcast-related information may exist in various forms, for example, in the form of an EPG (Electronic Program Guide) or an ESG (Electronic Service Guide).
The
The broadcast signal and / or broadcast related information received through the
The
The
The short-
The
The A /
The image frame processed by the
The
The
The
The
The
The
Some of these displays may be transparent or light transmissive so that they can be seen through. This can be referred to as a transparent display, and a typical example of the transparent display is TOLED (Transparent OLED) and the like. The rear structure of the
There may be two or
The
The touch sensor may be configured to convert a change in a pressure applied to a specific portion of the
If there is a touch input to the touch sensor, the corresponding signal (s) is sent to the touch controller. The touch controller processes the signal (s) and transmits corresponding data to the
A
Examples of the proximity sensor include a transmission type photoelectric sensor, a direct reflection type photoelectric sensor, a mirror reflection type photoelectric sensor, a high frequency oscillation type proximity sensor, a capacitive proximity sensor, a magnetic proximity sensor, and an infrared proximity sensor. And to detect the proximity of the pointer by the change of the electric field along the proximity of the pointer when the touch screen is electrostatic. In this case, the touch screen (touch sensor) may be classified as a proximity sensor.
Hereinafter, for convenience of explanation, the act of recognizing that the pointer is positioned on the touch screen while the pointer is not in contact with the touch screen is referred to as "proximity touch & The act of actually touching the pointer on the screen is called "contact touch. &Quot; The position where the pointer is proximately touched on the touch screen means a position where the pointer is vertically corresponding to the touch screen when the pointer is touched.
The proximity sensor detects a proximity touch and a proximity touch pattern (e.g., a proximity touch distance, a proximity touch direction, a proximity touch speed, a proximity touch time, a proximity touch position, a proximity touch movement state, and the like). Information corresponding to the detected proximity touch operation and the proximity touch pattern may be output on the touch screen.
The
The
The
The
The
The
The identification module is a chip for storing various information for authenticating the usage right of the
The
The
The
The various embodiments described herein may be implemented in a recording medium readable by a computer or similar device using, for example, software, hardware, or a combination thereof.
According to a hardware implementation, the embodiments described herein may be implemented as application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays May be implemented using at least one of a processor, a controller, micro-controllers, microprocessors, and an electrical unit for performing other functions. In some cases, the implementation described herein Examples may be implemented by the
According to a software implementation, embodiments such as the procedures and functions described herein may be implemented with separate software modules. Each of the software modules may perform one or more of the functions and operations described herein. Software code may be implemented in a software application written in a suitable programming language. Here, the software code is stored in the
4 is a block diagram illustrating a digital device according to another embodiment of the present invention.
Another example of the
The
For example, if the received RF broadcast signal is a digital broadcast signal, the signal is converted into a digital IF signal (DIF). If the received RF broadcast signal is an analog broadcast signal, the signal is converted into an analog baseband image or a voice signal (CVBS / SIF). That is, the
In addition, the
The
The
The stream signal output from the
The external
The external
The A / V input / output unit includes a USB terminal, a CVBS (Composite Video Banking Sync) terminal, a component terminal, an S-video terminal (analog) terminal, A DVI (Digital Visual Interface) terminal, an HDMI (High Definition Multimedia Interface) terminal, an RGB terminal, a D-SUB terminal, and the like.
The wireless communication unit can perform short-range wireless communication with another digital device. The
Also, the external
Meanwhile, the external
The
The
Meanwhile, the
In addition, the
The
The
The
In addition, the
The
4 illustrates an embodiment in which the
The user
For example, the user
In addition, the user
The user
The
The video signal processed by the
The audio signal processed by the
Although not shown in FIG. 4, the
The
The
For example, the
The
On the other hand, the
In addition, the
On the other hand, when entering the application view item, the
The
Although not shown in the drawing, a channel browsing processing unit for generating a channel signal or a thumbnail image corresponding to an external input signal may be further provided.
The channel browsing processing unit receives a stream signal TS output from the
The
The
Meanwhile, the
The
In order to detect the gesture of the user, a sensing unit (not shown) having at least one of a touch sensor, a voice sensor, a position sensor, and an operation sensor may be further included in the
On the other hand, a photographing unit (not shown) for photographing a user may be further provided. The image information photographed by the photographing unit (not shown) may be input to the
The
The
Particularly, it is possible to supply power to a
To this end, the
The
Also, the
The
In addition, the digital device according to the present invention may further include a configuration that omits some of the configuration shown in FIG. On the other hand, unlike the above, the digital device does not have a tuner and a demodulator, and can receive and reproduce the content through the network interface unit or the external device interface unit.
FIG. 5 is a block diagram illustrating a detailed configuration of the control unit of FIGS. 2 to 4 according to an embodiment of the present invention. Referring to FIG.
An example of the control unit includes a
The
The
The video decoder 425 decodes the demultiplexed video signal, and the
The
On the other hand, the video signal decoded by the
The
The
A frame rate conversion unit (FRC) 555 converts a frame rate of an input image. For example, the frame
The
On the other hand, the voice processing unit (not shown) in the control unit can perform the voice processing of the demultiplexed voice signal. Such a voice processing unit (not shown) may support processing various audio formats. For example, even when a voice signal is coded in a format such as MPEG-2, MPEG-4, AAC, HE-AAC, AC-3, or BSAC, a corresponding decoder can be provided.
In addition, the voice processing unit (not shown) in the control unit can process the base, the treble, the volume control, and the like.
A data processing unit (not shown) in the control unit can perform data processing of the demultiplexed data signal. For example, the data processing unit can decode the demultiplexed data signal even when it is coded. Here, the encoded data signal may be EPG information including broadcast information such as a start time and an end time of a broadcast program broadcasted on each channel.
On the other hand, the above-described digital device is an example according to the present invention, and each component can be integrated, added, or omitted according to specifications of a digital device actually implemented. That is, if necessary, two or more components may be combined into one component, or one component may be divided into two or more components. In addition, the functions performed in each block are intended to illustrate the embodiments of the present invention, and the specific operations and devices thereof do not limit the scope of rights of the present invention.
Meanwhile, the digital device may be a video signal processing device that performs signal processing of an image stored in the device or an input image. Other examples of the video signal processing device include a set-top box (STB), a DVD player, a Blu-ray player, a game device, a computer Etc. can be further exemplified.
FIG. 6 is a diagram illustrating input means coupled to the digital device of FIGS. 2 through 4 according to one embodiment of the present invention.
A front panel (not shown) or a control means (input means) provided on the
The control means includes a
The input means may be a communication protocol such as Bluetooth, Radio Frequency Identification (RFID), Infrared Data Association (UDA), Ultra Wideband (UWB), ZigBee, Digital Living Network Alliance (DLNA) At least one can be employed as needed to communicate with the digital device.
The
The
Since the
On the other hand, the control means such as the
The digital device described in this specification uses the Web OS as an OS and / or platform. Hereinafter, the processing such as the configuration or algorithm based on the web OS can be performed in the control unit of the above-described digital device or the like. Here, the control unit includes the control unit in FIGS. 2 to 5 described above and uses it as a broad concept. Accordingly, in order to process services, applications, contents, and the like related to the web OS in the digital device, the hardware and components including related software, firmware, and the like are controlled by a controller Named and explained.
Such a web OS-based platform is intended to enhance development independence and function expandability by integrating services, applications, and the like based on, for example, a luna-service bus, Productivity can be increased. Also, multi-tasking can be supported by efficiently utilizing system resources and the like through a Web OS process and resource management.
Meanwhile, the web OS platform described in this specification can be used not only in fixed devices such as a PC, a TV, and a set-top box (STB) but also in mobile devices such as mobile phones, smart phones, tablet PCs, notebooks, wearable devices .
The structure of the software for digital devices is based on a single process and a closed product based on multi-threading with conventional problem solving and market-dependent monolithic structure, And has been pursuing new platform-based development since then, has pursued cost innovation through chip-set replacement, UI application and external application development efficiency, and developed layering and componentization Layer structure and an add-on structure for add-ons, single-source products, and open applications. More recently, the software architecture provides a modular architecture for functional units, a Web Open API (Application Programming Interface) for echo-systems, and a game engine And a native open API (Native Open API), and thus, a multi-process structure based on a service structure is being created.
Hereinafter, embodiments of the present invention will be described with reference to FIGS. 7 to 14. FIG. In the following, reference will be made to the above description with reference to Figs. 1 to 6 in explaining and understanding the embodiments of the present invention. Also, the above-described display device or digital device may be a display device according to an embodiment of the present invention.
The display device described in the present specification can be applied to various devices such as a TV, a smart TV, a Hybrid Brodacast Broadband Television (HBBTV), a network TV, a Web TV, an Internet Protocol Television (IPTV), a personal computer, a tablet PC, a notebook, A smart phone, a smart watch, a smart glass, and the like.
7 is a block diagram illustrating configuration modules of a display device according to an embodiment of the present invention.
The
The
According to an embodiment of the present invention, the
Meanwhile, according to the embodiment, the specific contents may be a moving picture file in which audio data and video data are encoded.
The
The
The control unit 750 controls operations of the
According to an embodiment of the present invention, the
FIG. 8 is a flowchart illustrating an example of a method for adjusting the synchronization of subtitles in a display device according to an exemplary embodiment of the present invention. Referring to FIG.
According to an embodiment of the present invention, the
The first information on the playback time of the subtitle may include information about a frame in which the subtitle is output. Accordingly, the
The specific content may be content received through an external server or content stored by the user in the
The subtitle may be a subtitle received through an external server or a subtitle stored in a
The
The
According to an embodiment of the present invention, when displaying the caption, the
On the other hand, in step S820, the
The
In step S830, the second information may include information on a time when the detection of the voice is started, information on a time when the detection of the voice is terminated, information on a time interval during which the voice is not detected, And information on a time interval during which the
Specifically, in step S830, the
For example, the
The
In step S840, the third information may include information about a time at which the output of the caption is started, information about a time at which the output of the caption is terminated, information about a time period during which the caption is not output, And information on a time interval during which the
Specifically, in step S840, the
For example, the
The
In step S840, the
In the first method, when the time interval during which the voice is detected and the time period during which the caption is output are different using the second information and the third information, the
For example, when the time at which the voice starts to be detected and the time at which the caption starts to be output are different (for example, when the voice is output but no caption is output), the
In another example, the
As a second method, when the time period during which the voice is not detected and the time period during which the caption is not output are different using the second information and the third information, the
For example, when the time at which the detection of the voice is ended and the time at which the output of the subtitles are terminated are different (for example, when the subtitles are outputted although the voice is not outputted), the
In another example, the
Meanwhile, according to the embodiment, the
According to the present embodiment, there is an advantage that the user can control the sink of the subtitles automatically on the TV without having to manually adjust the synchronization of the subtitles on the TV, so that the user can concentrate on the content being output.
According to an embodiment of the present invention, the
9 and 10 are diagrams for explaining an example of a method of adjusting the synchronization of the caption when the subtitle output from the display device according to the embodiment of the present invention recognizes that the audio data being output is out of sync with the audio data being output FIG.
According to an embodiment of the present invention, the
Referring to FIG. 9A, the
9B, the
The
The
Specifically, the
For example, the
The
Specifically, the
For example, referring to FIG. 10, the
The
The
For example, referring to FIGS. 9A and 9B, the
According to the embodiment, when the time interval during which the voice is not detected and the time period during which the caption is not output are different using the second information and the third information, the
For example, when the time at which the detection of the voice is ended and the time at which the output of the subtitles are terminated are different (for example, when the subtitles are outputted although the voice is not outputted), the
In another example, the
According to an embodiment of the present invention, the
11 is a flowchart for explaining an example of a method of analyzing a voice of a recognized voice to give a visual effect to an outputting subtitle when a voice is recognized in audio data being output from a display device according to an embodiment of the present invention .
According to an embodiment of the present invention, the
The
Specifically, in step S910, the
For example, the
The
In the
The
For example, when the recognized voice corresponds to a first sentence associated with the first person stored in the
In another example, when the recognized voice corresponds to a second sentence associated with the second person stored in the
That is, when the recognized voice corresponds to the stored voice in the
If the recognized voice does not correspond to the stored voice in the memory 710 (S920, No), the
The
For example, if it is assumed that the first visual effect is mapped and stored in the
According to this embodiment, when there are a plurality of speakers in the output content, it is possible to recognize which subtitles the user is outputting and which subtitles are related to which speakers, and give different visual effects to the subtitles associated with each speaker, Is easy to see and easy to understand.
12 is a view for explaining an example of a method of analyzing a voice of a recognized voice and giving a visual effect to a subtitle being output when a voice is recognized in audio data being outputted from a display device according to an embodiment of the present invention FIG.
According to an embodiment of the present invention, the
The
12A, when the recognized voice corresponds to a first sentence associated with the first person A1 stored in the
12 (b), when the recognized voice corresponds to a second sentence associated with the second person A2 stored in the
That is, when the recognized voice corresponds to the stored voice in the
On the other hand, if the recognized voice is not corresponding to the stored voice in the
The
For example, if it is assumed that the first visual effect is mapped and stored in the
According to this embodiment, when there are a plurality of speakers in the output content, it is possible to recognize which subtitles the user is outputting and which subtitles are related to which speakers, and give different visual effects to the subtitles associated with each speaker, Is easy to see and easy to understand.
13 is a view for explaining an example of a method of mapping a visual effect assigned in accordance with a grammar in a display device according to an embodiment of the present invention to a memory.
Referring to FIG. 13, the
For example, the
The
Meanwhile, according to an embodiment of the present invention, the
FIG. 14 is a flowchart for explaining an example of a method of converting a voice of audio data being outputted from a display device into text and adjusting the synchronization of the caption using the translation result of the converted text according to an embodiment of the present invention.
The
Specifically, the
For example, the
The
In one example, an algorithm for converting the recognized speech into text may be stored in
However, according to the embodiment, the
The
However, according to the embodiment, the
The
For example, when "I'm your farther." Is being output through the audio data, the
According to the embodiment, the
Meanwhile, according to the embodiment, the
According to the present embodiment, there is an advantage that the synchronization of subtitles can be adjusted more precisely.
According to at least one embodiment of the present invention, the user can adjust the synchronization of the subtitles automatically on the TV without manually adjusting the subtitles of the user on the TV so that the user can concentrate on the content being output . According to an embodiment of the present invention, when there are a plurality of speakers in the output content, it is possible to recognize which subtitle is related to which speaker and output a different visual effect to each subtitle associated with each speaker, Has the advantage that the content can be easily viewed and understood.
The display device and the control method of a display device disclosed in this specification can be applied to a configuration and a method of the embodiments described above in a limited manner, but the embodiments can be applied to all or a part of each embodiment Or may be selectively combined.
Meanwhile, the operation method of the digital device disclosed in this specification can be implemented as a code readable by a processor in a recording medium readable by a processor included in the digital device. The processor-readable recording medium includes all kinds of recording devices in which data that can be read by the processor is stored. Examples of the recording medium readable by the processor include ROM (Read Only Memory), RAM (Random Access Memory), CD-ROM, magnetic tape, floppy disk, optical data storage device, And may be implemented in the form of a carrier-wave. In addition, the processor-readable recording medium may be distributed over network-connected computer systems so that code readable by the processor in a distributed fashion can be stored and executed.
While the present invention has been particularly shown and described with reference to exemplary embodiments thereof, it will be understood by those of ordinary skill in the art that various changes in form and details may be made therein without departing from the spirit and scope of the invention as defined by the appended claims. Of the right. Further, such modifications are not to be understood individually from the technical idea of the present invention.
700: Display device
710:
720:
730: Memory
740: audio output section
750:
Claims (16)
A display unit for displaying video data of the specific content;
An audio output unit for outputting audio data of the specific content; And
And a control unit for controlling the display unit such that the subtitle is output together with the video data using the first information,
Wherein,
Recognizing second information on a time when audio is detected from the audio data of the specific content being output,
Recognizing third information on a time when the subtitles are output using the first information,
And adjusts the sink of the outputting subtitle according to a result of comparison between the second information and the third information.
Wherein the second information comprises:
Information about a time at which the detection of the voice is started, information about a time at which the detection of the voice is finished, information about a time period during which the voice is not detected, and information about a time interval during which the voice is detected and,
Wherein the third information comprises:
Information on a time at which the output of the subtitles is started, information about a time at which the output of the subtitles is terminated, information about a time period during which the subtitles are not output, and information about time periods during which the subtitles are output / RTI >
Wherein,
And controls the sink of the output subtitle so that a time interval during which the subtitle is output corresponds to a time interval during which the audio is recognized when the time interval during which the audio is detected and the time interval during which the subtitle is output are different.
Wherein,
And adjusts the sink of the output subtitle if the time interval during which the speech is not recognized and the time period during which the subtitle is not output are different.
Wherein,
And adjusts the sink of the output subtitle so that a time period during which the subtitle is not output corresponds to a time period during which the speech is not recognized.
Wherein,
Detecting a range of audio data of the specific content being output,
And recognizes the detected sound range as a time interval in which the sound is sensed when the sensed sound range corresponds to a preset sound range.
The memory,
Storing a first visual effect corresponding to a first voiceprint and a second visual effect corresponding to a second voiceprint,
Wherein,
Recognizes a voice in the output audio data,
Recognizes a voiceprint of the recognized voice,
When the recognized gateways correspond to the first sentence, giving the first visual effect to the outputting subtitle,
And gives the second visual effect to the outputting subtitle if the recognized gates correspond to the second gates.
Wherein,
Recognizes a voice in the output audio data,
Converts the recognized speech into text,
Translating the converted text,
And adjusts the sink of the output subtitle according to a result of translation of the converted text and a result of comparison of the subtitle.
Outputting the audio data of the specific content;
Outputting a subtitle associated with the specific content stored in the memory together with the video data using first information on the reproduction time of the subtitle stored in the memory;
Recognizing second information on a time when audio is detected from audio data of the specific content being output;
Recognizing third information on a time when the subtitles are output using the first information; And
And adjusting a sync of the output subtitle according to a result of comparison between the second information and the third information.
Wherein the second information comprises:
Information about a time at which the detection of the voice is started, information about a time at which the detection of the voice is finished, information about a time period during which the voice is not detected, and information about a time interval during which the voice is detected and,
Wherein the third information comprises:
Information on a time at which the output of the subtitles is started, information about a time at which the output of the subtitles is terminated, information about a time period during which the subtitles are not output, and information about time periods during which the subtitles are output Of the display device.
Adjusting the synchronization of the output subtitle according to the comparison result of the second information and the third information,
Wherein the control unit controls the sink of the output subtitle so that a time interval during which the subtitle is output corresponds to a time interval during which the audio is recognized when the time interval during which the audio is detected and the time interval during which the subtitle is output are different Control method.
Adjusting the synchronization of the output subtitle according to the comparison result of the second information and the third information,
Wherein the control unit controls the sink of the subtitle being output when the time period during which the speech is not recognized and the time period during which the subtitle is not output are different.
Adjusting the synchronization of the output subtitle according to the comparison result of the second information and the third information,
And adjusts the sink of the output subtitle so that a time interval during which the subtitle is not output corresponds to a time interval during which the speech is not recognized.
The step of recognizing the second information on the time when the audio is detected from the audio data of the specific content being output includes:
Sensing a range of audio data of the specific content being output; And
And recognizing the sensed range as a time interval in which a voice is sensed when the sensed range corresponds to a preset range.
The step of outputting the caption associated with the specific content stored in the memory together with the video data using the first information on the reproduction time of the caption stored in the memory,
Recognizing speech in the output audio data;
Recognizing a voiceprint of the recognized voice;
Applying a first visual effect to the outputting subtitle if the recognized gates correspond to a first sentence;
And when the recognized gates correspond to the second sentence, giving a second visual effect to the outputting subtitle.
Recognizing speech in the output audio data;
Converting the recognized speech into text;
Translating the translated text;
And adjusting the synchronization of the output subtitle according to the translation result of the converted text and the comparison result of the subtitle.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020150147779A KR20170047547A (en) | 2015-10-23 | 2015-10-23 | Display device and method for controlling the same |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020150147779A KR20170047547A (en) | 2015-10-23 | 2015-10-23 | Display device and method for controlling the same |
Publications (1)
Publication Number | Publication Date |
---|---|
KR20170047547A true KR20170047547A (en) | 2017-05-08 |
Family
ID=60164342
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020150147779A KR20170047547A (en) | 2015-10-23 | 2015-10-23 | Display device and method for controlling the same |
Country Status (1)
Country | Link |
---|---|
KR (1) | KR20170047547A (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101996551B1 (en) | 2018-01-02 | 2019-07-04 | 중앙대학교 산학협력단 | Apparatus and method for generating subtitles using speech recognition and script |
EP3609189A1 (en) * | 2018-08-10 | 2020-02-12 | Nagravision SA | Testing rendering of screen objects |
KR20200134974A (en) | 2019-05-24 | 2020-12-02 | 연세대학교 산학협력단 | Apparatus and method for controlling image based on user recognition |
CN114640874A (en) * | 2022-03-09 | 2022-06-17 | 湖南国科微电子股份有限公司 | Subtitle synchronization method and device, set top box and computer readable storage medium |
-
2015
- 2015-10-23 KR KR1020150147779A patent/KR20170047547A/en unknown
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101996551B1 (en) | 2018-01-02 | 2019-07-04 | 중앙대학교 산학협력단 | Apparatus and method for generating subtitles using speech recognition and script |
EP3609189A1 (en) * | 2018-08-10 | 2020-02-12 | Nagravision SA | Testing rendering of screen objects |
WO2020030810A1 (en) * | 2018-08-10 | 2020-02-13 | Nagravision Sa | Testing rendering of screen objects |
US11647249B2 (en) | 2018-08-10 | 2023-05-09 | Nagravision S.A. | Testing rendering of screen objects |
KR20200134974A (en) | 2019-05-24 | 2020-12-02 | 연세대학교 산학협력단 | Apparatus and method for controlling image based on user recognition |
CN114640874A (en) * | 2022-03-09 | 2022-06-17 | 湖南国科微电子股份有限公司 | Subtitle synchronization method and device, set top box and computer readable storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR101567832B1 (en) | Digital device and method for controlling the same | |
KR102063075B1 (en) | Service system, digital device and method of processing a service thereof | |
EP3316592A1 (en) | Display device and control method therefor | |
KR102413328B1 (en) | Main speaker, sub speaker and system comprising main speaker and sub speaker | |
KR102348957B1 (en) | Digital device and method for controlling the same | |
KR20160062417A (en) | Multimedia device and method for controlling the same | |
KR20160023089A (en) | Digital device and method for controlling the same | |
KR101632221B1 (en) | Digital device and method for processing service thereof | |
KR20160066268A (en) | Multimedia device and method for controlling the same | |
KR20150127415A (en) | Digital device and method of processing a service thereof | |
KR20170031370A (en) | Mobile terminal and method for controlling the same | |
KR102381142B1 (en) | Mobile terminal and method for controlling the same | |
KR20170028104A (en) | Display device and method for controlling the same | |
KR20160065504A (en) | Multimedia device and method for controlling the same | |
KR20170047547A (en) | Display device and method for controlling the same | |
KR102396035B1 (en) | Digital device and method for processing stt thereof | |
KR20170087307A (en) | Display device and method for controlling the same | |
KR101843537B1 (en) | Method for connecting to wireless receiving apparatus in display apparatus and display apparatus thereof | |
KR20170031898A (en) | Display device and method for controlling the same | |
KR20170138788A (en) | Digital device and controlling method thereof | |
KR20160004562A (en) | Display device and method for controlling the same | |
KR20170002119A (en) | Display device and controlling method thereof | |
KR20170022612A (en) | Display device and method for controlling the same | |
KR20170010484A (en) | Display device and controlling method thereof | |
KR20150101902A (en) | Digital device and method of controlling thereof |