CN112506400A - Page information voice broadcasting method and display device - Google Patents

Page information voice broadcasting method and display device Download PDF

Info

Publication number
CN112506400A
CN112506400A CN202011413371.4A CN202011413371A CN112506400A CN 112506400 A CN112506400 A CN 112506400A CN 202011413371 A CN202011413371 A CN 202011413371A CN 112506400 A CN112506400 A CN 112506400A
Authority
CN
China
Prior art keywords
page
focus
display
text
display device
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011413371.4A
Other languages
Chinese (zh)
Inventor
董杰
贾桂丽
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hisense Visual Technology Co Ltd
Original Assignee
Hisense Visual Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hisense Visual Technology Co Ltd filed Critical Hisense Visual Technology Co Ltd
Priority to CN202011413371.4A priority Critical patent/CN112506400A/en
Publication of CN112506400A publication Critical patent/CN112506400A/en
Priority to PCT/CN2021/081327 priority patent/WO2021238353A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0483Interaction with page-structured environments, e.g. book metaphor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/0485Scrolling or panning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/165Management of the audio stream, e.g. setting of volume, audio stream path
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/08Text analysis or generation of parameters for speech synthesis out of text, e.g. grapheme to phoneme translation, prosody generation or stress or intonation determination
    • G10L13/086Detection of language

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Acoustics & Sound (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The invention discloses a page information voice broadcasting method and display equipment, which respond to the operation of starting a page, display the page, and synchronously broadcast the text information in the page when detecting that the current state of a text broadcasting function is started; controlling the focus to dynamically move in the page along with the broadcast content; the focus is used for marking the position of the current broadcast content in the page. The focus follows that the broadcast content is synchronous, dynamic movement in this application, and voice broadcast arrives which information promptly, and then the focus removes the position at this information place for this information can obtain the focus, and convenience of customers passes through the focus directly perceived, fixes a position fast and reports the position, promotes user experience.

Description

Page information voice broadcasting method and display device
Technical Field
The invention relates to the field of display equipment, in particular to a page information voice broadcasting method and display equipment.
Background
At present, some display devices are additionally provided with a Text To Speech (TTS) function, which belongs To one of Speech synthesis applications and can convert Text information such as help files, electronic specifications, web pages and the like in the display devices into natural Speech for output and broadcast through the Speech. The TTS function can help users with visual disorder to know the text information displayed by the display equipment end, and can also enhance the readability of the text information and deepen the memory of the users to the text.
When the display device displays the application page and simultaneously broadcasts the text information in the application page by voice, if the content in the application page is numerous and the information amount is large, the user can hardly locate the text position currently broadcasted by voice, so that the user can not quickly find and view the information of the specific broadcasting position, and the actual experience of the user is influenced.
Disclosure of Invention
In order to solve the problems in the background art, the invention provides a page information voice broadcasting method and display equipment.
A first aspect provides a display device comprising:
a display for displaying a page;
the sound player is used for broadcasting the text information in the page in a voice way;
a controller configured to perform:
responding to the operation of starting the page, controlling a display to display the page, and controlling a sound player to synchronously broadcast text information in the page when detecting that the current state of a text broadcast function is started;
controlling the focus to dynamically move in the page along with the broadcast content; the focus is used for marking the position of the current broadcast content in the page.
In some embodiments, the controller is further configured to perform: and if the focus is detected to move to a first preset position of a display interface, controlling the page to scroll upwards so as to enable the focus to be at a second preset position of the display interface.
In some embodiments, the first predetermined location is a bottom critical location of a visualization of the display interface.
In some embodiments, the second preset position is a central position of the display interface along the longitudinal direction.
In some embodiments, the focus is marked in units of words, characters, words, sentences, or paragraphs.
In some embodiments, the controller controls the sound player to synchronously play the text information in the page as follows:
acquiring text information loaded in the page and a target language corresponding to the text information;
and controlling a sound player to broadcast the text information according to the target language.
The page information voice broadcasting method provided by the second aspect comprises the following steps:
responding to the operation of starting a page, displaying the page, and synchronously broadcasting text information in the page when detecting that the current state of a text broadcasting function is started;
controlling the focus to dynamically move in the page along with the broadcast content; the focus is used for marking the position of the current broadcast content in the page.
The technical scheme provided by the application is suitable for the display equipment with the TTS function, when a user starts a certain page, the current state of the text broadcast function (namely the TTS function) needs to be detected besides the displayed page, and the state can comprise opened state or unopened state. If the text broadcasting function is not started, only displaying the page without text conversion and voice output; if the text broadcast function is currently in an opened state, the text information in the page is converted into voice information for broadcast while the page is displayed. The key of this application is in having add the focus that is arranged in the current position of reporting content in the page of mark, and this focus can follow and report content synchronization, dynamic movement, and which information is reported to voice promptly, then the focus moves the position at this information place for this information can obtain the focus, and the focus for example adopts the form such as thickening, highlight or colour transform to the information, and convenience of customers passes through the focus directly perceived, fix a position fast and report the position, promotes user experience.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings to be accessed in the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings without creative efforts.
FIG. 1 illustrates a usage scenario of a display device according to some embodiments;
fig. 2 illustrates a hardware configuration block diagram of the control apparatus 100 according to some embodiments;
fig. 3 illustrates a hardware configuration block diagram of the display apparatus 200 according to some embodiments;
FIG. 4 illustrates a software configuration diagram in the display device 200 according to some embodiments;
FIG. 5 is a diagram illustrating a setup interface for a text cast function;
fig. 6 is a view exemplarily showing a state switch setting interface of the text broadcast function;
fig. 7 is a schematic diagram illustrating an interface change after the text broadcast function is enabled;
FIG. 8 is a diagram illustrating the focus of a page display during a voice announcement;
FIG. 9 is a schematic diagram illustrating page scroll adjusting focus position;
fig. 10 is a flowchart illustrating a page information voice broadcasting method.
Detailed Description
To make the purpose and embodiments of the present application clearer, the following will clearly and completely describe the exemplary embodiments of the present application with reference to the attached drawings in the exemplary embodiments of the present application, and it is obvious that the described exemplary embodiments are only a part of the embodiments of the present application, and not all of the embodiments.
It should be noted that the brief descriptions of the terms in the present application are only for the convenience of understanding the embodiments described below, and are not intended to limit the embodiments of the present application. These terms should be understood in their ordinary and customary meaning unless otherwise indicated.
The terms "first," "second," "third," and the like in the description and claims of this application and in the above-described drawings are used for distinguishing between similar or analogous objects or entities and not necessarily for describing a particular sequential or chronological order, unless otherwise indicated. It is to be understood that the terms so used are interchangeable under appropriate circumstances.
The terms "comprises" and "comprising," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a product or apparatus that comprises a list of elements is not necessarily limited to all elements expressly listed, but may include other elements not expressly listed or inherent to such product or apparatus.
The term "module" refers to any known or later developed hardware, software, firmware, artificial intelligence, fuzzy logic, or combination of hardware and/or software code that is capable of performing the functionality associated with that element.
Fig. 1 is a schematic diagram of a usage scenario of a display device according to an embodiment. As shown in fig. 1, the display apparatus 200 is also in data communication with a server 400, and a user can operate the display apparatus 200 through the smart device 300 or the control device 100.
In some embodiments, the control apparatus 100 may be a remote controller, and the communication between the remote controller and the display device includes at least one of an infrared protocol communication or a bluetooth protocol communication, and other short-distance communication methods, and controls the display device 200 in a wireless or wired manner. The user may control the display apparatus 200 by inputting a user instruction through at least one of a key on a remote controller, a voice input, a control panel input, and the like.
In some embodiments, the smart device 300 may include any of a mobile terminal, a tablet, a computer, a laptop, an AR/VR device, and the like.
In some embodiments, the smart device 300 may also be used to control the display device 200. For example, the display device 200 is controlled using an application program running on the smart device.
In some embodiments, the smart device 300 and the display device may also be used for communication of data.
In some embodiments, the display device 200 may also be controlled in a manner other than the control apparatus 100 and the smart device 300, for example, the voice instruction control of the user may be directly received by a module configured inside the display device 200 to obtain a voice instruction, or may be received by a voice control apparatus provided outside the display device 200.
In some embodiments, the display device 200 is also in data communication with a server 400. The display device 200 may be allowed to be communicatively connected through a Local Area Network (LAN), a Wireless Local Area Network (WLAN), and other networks. The server 400 may provide various contents and interactions to the display apparatus 200. The server 400 may be a cluster or a plurality of clusters, and may include one or more types of servers.
In some embodiments, software steps executed by one step execution agent may be migrated on demand to another step execution agent in data communication therewith for execution. Illustratively, software steps performed by the server may be migrated to be performed on a display device in data communication therewith, and vice versa, as desired.
Fig. 2 exemplarily shows a block diagram of a configuration of the control apparatus 100 according to an exemplary embodiment. As shown in fig. 2, the control device 100 includes a controller 110, a communication interface 130, a user input/output interface 140, a memory, and a power supply. The control apparatus 100 may receive an input operation instruction from a user and convert the operation instruction into an instruction recognizable and responsive by the display device 200, serving as an interaction intermediary between the user and the display device 200.
In some embodiments, the communication interface 130 is used for external communication, and includes at least one of a WIFI chip, a bluetooth module, NFC, or an alternative module.
In some embodiments, the user input/output interface 140 includes at least one of a microphone, a touchpad, a sensor, a key, or an alternative module.
Fig. 3 shows a hardware configuration block diagram of the display apparatus 200 according to an exemplary embodiment.
In some embodiments, the display apparatus 200 includes at least one of a tuner demodulator 210, a communicator 220, a detector 230, an external device interface 240, a controller 250, a display 260, an audio output interface 270, a memory, a power supply, a user interface.
In some embodiments the controller comprises a central processor, a video processor, an audio processor, a graphics processor, a RAM, a ROM, a first interface to an nth interface for input/output.
In some embodiments, the display 260 includes a display screen component for displaying pictures, and a driving component for driving image display, a component for receiving image signals from the controller output, displaying video content, image content, and menu manipulation interface, and a user manipulation UI interface, etc.
In some embodiments, the display 260 may be at least one of a liquid crystal display, an OLED display, and a projection display, and may also be a projection device and a projection screen.
In some embodiments, the tuner demodulator 210 receives broadcast television signals via wired or wireless reception, and demodulates audio/video signals, such as EPG data signals, from a plurality of wireless or wired broadcast television signals.
In some embodiments, communicator 220 is a component for communicating with external devices or servers according to various communication protocol types. For example: the communicator may include at least one of a Wifi module, a bluetooth module, a wired ethernet module, and other network communication protocol chips or near field communication protocol chips, and an infrared receiver. The display apparatus 200 may establish transmission and reception of control signals and data signals with the control device 100 or the server 400 through the communicator 220.
In some embodiments, the detector 230 is used to collect signals of the external environment or interaction with the outside. For example, detector 230 includes a light receiver, a sensor for collecting ambient light intensity; alternatively, the detector 230 includes an image collector, such as a camera, which may be used to collect external environment scenes, attributes of the user, or user interaction gestures, or the detector 230 includes a sound collector, such as a microphone, which is used to receive external sounds.
In some embodiments, the external device interface 240 may include, but is not limited to, the following: high Definition Multimedia Interface (HDMI), analog or data high definition component input interface (component), composite video input interface (CVBS), USB input interface (USB), RGB port, and the like. The interface may be a composite input/output interface formed by the plurality of interfaces.
In some embodiments, the controller 250 and the modem 210 may be located in different separate devices, that is, the modem 210 may also be located in an external device of the main device where the controller 250 is located, such as an external set-top box.
In some embodiments, the controller 250 controls the operation of the display device and responds to user operations through various software control programs stored in memory. The controller 250 controls the overall operation of the display apparatus 200. For example: in response to receiving a user command for selecting a UI object to be displayed on the display 260, the controller 250 may perform an operation related to the object selected by the user command.
In some embodiments, the object may be any one of selectable objects, such as a hyperlink, an icon, or other actionable control. The operations related to the selected object are: displaying an operation connected to a hyperlink page, document, image, or the like, or performing an operation of a program corresponding to the icon.
In some embodiments the controller comprises at least one of a Central Processing Unit (CPU), a video processor, an audio processor, a Graphics Processing Unit (GPU), a RAM Random Access Memory (RAM), a ROM (Read-Only Memory), a first to nth interface for input/output, a communication Bus (Bus), and the like.
A CPU processor. For executing operating system and application program instructions stored in the memory, and executing various application programs, data and contents according to various interactive instructions receiving external input, so as to finally display and play various audio-video contents. The CPU processor may include a plurality of processors. E.g. comprising a main processor and one or more sub-processors.
In some embodiments, a graphics processor for generating various graphics objects, such as: at least one of an icon, an operation menu, and a user input instruction display figure. The graphic processor comprises an arithmetic unit, which performs operation by receiving various interactive instructions input by a user and displays various objects according to display attributes; the system also comprises a renderer for rendering various objects obtained based on the arithmetic unit, wherein the rendered objects are used for being displayed on a display.
In some embodiments, the video processor is configured to receive an external video signal, and perform at least one of video processing such as decompression, decoding, scaling, noise reduction, frame rate conversion, resolution conversion, and image synthesis according to a standard codec protocol of the input signal, so as to obtain a signal displayed or played on the direct display device 200.
In some embodiments, the video processor includes at least one of a demultiplexing module, a video decoding module, an image composition module, a frame rate conversion module, a display formatting module, and the like. The demultiplexing module is used for demultiplexing the input audio and video data stream. And the video decoding module is used for processing the video signal after demultiplexing, including decoding, scaling and the like. And the image synthesis module is used for carrying out superposition mixing processing on the GUI signal input by the user or generated by the user and the video image after the zooming processing by the graphic generator so as to generate an image signal for display. And the frame rate conversion module is used for converting the frame rate of the input video. And the display formatting module is used for converting the received video output signal after the frame rate conversion, and changing the signal to be in accordance with the signal of the display format, such as an output RGB data signal.
In some embodiments, the audio processor is configured to receive an external audio signal, decompress and decode the received audio signal according to a standard codec protocol of the input signal, and perform at least one of noise reduction, digital-to-analog conversion, and amplification processing to obtain a sound signal that can be played in the speaker.
In some embodiments, a user may enter user commands on a Graphical User Interface (GUI) displayed on display 260, and the user input interface receives the user input commands through the Graphical User Interface (GUI). Alternatively, the user may input the user command by inputting a specific sound or gesture, and the user input interface receives the user input command by recognizing the sound or gesture through the sensor.
In some embodiments, a "user interface" is a media interface for interaction and information exchange between an application or operating system and a user that enables conversion between an internal form of information and a form that is acceptable to the user. A commonly used presentation form of the User Interface is a Graphical User Interface (GUI), which refers to a User Interface related to computer operations and displayed in a graphical manner. It may be an interface element such as an icon, a window, a control, etc. displayed in the display screen of the electronic device, where the control may include at least one of an icon, a button, a menu, a tab, a text box, a dialog box, a status bar, a navigation bar, a Widget, etc. visual interface elements.
In some embodiments, user interface 280 is an interface that may be used to receive control inputs (e.g., physical buttons on the body of the display device, or the like).
In some embodiments, a system of a display device may include a Kernel (Kernel), a command parser (shell), a file system, and an application program. The kernel, shell, and file system together make up the basic operating system structure that allows users to manage files, run programs, and use the system. After power-on, the kernel is started, kernel space is activated, hardware is abstracted, hardware parameters are initialized, and virtual memory, a scheduler, signals and interprocess communication (IPC) are operated and maintained. And after the kernel is started, loading the Shell and the user application program. The application program is compiled into machine code after being started, and a process is formed.
Referring to fig. 4, in some embodiments, the system is divided into four layers, which are an Application (Applications) layer (abbreviated as "Application layer"), an Application Framework (Application Framework) layer (abbreviated as "Framework layer"), an Android runtime (Android runtime) and system library layer (abbreviated as "system runtime library layer"), and a kernel layer from top to bottom.
In some embodiments, at least one application program runs in the application program layer, and the application programs may be windows (windows) programs carried by an operating system, system setting programs, clock programs or the like; or an application developed by a third party developer. In particular implementations, the application packages in the application layer are not limited to the above examples.
The framework layer provides an Application Programming Interface (API) and a programming framework for the application program of the application layer. The application framework layer includes a number of predefined functions. The application framework layer acts as a processing center that decides to let the applications in the application layer act. The application program can access the resources in the system and obtain the services of the system in execution through the API interface.
As shown in fig. 4, in the embodiment of the present application, the application framework layer includes a manager (Managers), a Content Provider (Content Provider), and the like, where the manager includes at least one of the following modules: an Activity Manager (Activity Manager) is used for interacting with all activities running in the system; the Location Manager (Location Manager) is used for providing the system service or application with the access of the system Location service; a Package Manager (Package Manager) for retrieving various information related to an application Package currently installed on the device; a Notification Manager (Notification Manager) for controlling display and clearing of Notification messages; a Window Manager (Window Manager) is used to manage the icons, windows, toolbars, wallpapers, and desktop components on a user interface.
In some embodiments, the activity manager is used to manage the lifecycle of the various applications as well as general navigational fallback functions, such as controlling exit, opening, fallback, etc. of the applications. The window manager is used for managing all window programs, such as obtaining the size of a display screen, judging whether a status bar exists, locking the screen, intercepting the screen, controlling the change of the display window (for example, reducing the display window, displaying a shake, displaying a distortion deformation, and the like), and the like.
In some embodiments, the system runtime layer provides support for the upper layer, i.e., the framework layer, and when the framework layer is used, the android operating system runs the C/C + + library included in the system runtime layer to implement the functions to be implemented by the framework layer.
In some embodiments, the kernel layer is a layer between hardware and software. As shown in fig. 4, the core layer includes at least one of the following drivers: audio drive, display driver, bluetooth drive, camera drive, WIFI drive, USB drive, HDMI drive, sensor drive (like fingerprint sensor, temperature sensor, pressure sensor etc.) and power drive etc..
The above embodiments describe the hardware/software architecture and functional implementation of the display device. In some application scenarios, the display device may be configured with a TTS function, which converts text information into voice information for playing, so that a user with visual impairment can conveniently know the information content displayed in the current page, and a user without visual impairment can sense the page information from both visual sense and auditory sense, thereby improving the intelligence and user experience of the display device.
In some embodiments, a user may enter a setting page of a text broadcast function (i.e., a TTS function) through a certain operation path, for example, click a "Settings" (Settings) control on a television homepage, enter a setting application, and then select a "no obstacle" option in "device preference setting", so as to display an interface as shown in fig. 5, the user may set an on-off state of the text broadcast function, for example, TalkBack in fig. 5 is an entry for setting the on-off state of the text broadcast function, and at this time, a display below the TalkBack is turned off, which indicates that the text broadcast function is not currently turned on. If the user wants to use the text broadcast function, the user clicks TalkBack to enter the interface shown in fig. 6, and the switching control in fig. 6 displays "enable" in front, which indicates that the current text broadcast function is in an unopened state, and the user can enable the function by clicking the switching control once, and then the interface shows the change shown in fig. 7. In fig. 7, the text display in front of the switch control is changed to "off", that is, it indicates that the current text broadcast function is in an on state, and the user only needs to click the switch control again if the user wants to turn off the text broadcast function. The user can start and stop the text broadcast function according to personal use requirements by operating the switch control through the interface of fig. 6 or fig. 7.
If the user starts the text broadcast function, when the user opens an application, such as a built-in application of the display device or an installed third-party application, and then enters a page in the application, text information in the page is broadcasted synchronously to the user. The language of the text information displayed on the page can be set according to the use environment and the geographical position of the display device, and the language used during voice broadcasting is generally consistent with the character language by default.
The page display contents of different applications are different, if a user needs to locate the specific position of the current voice broadcast content in the page, the voice information and the text information need to be compared, and as the voice information is dynamically and continuously played and the text information amount in the page is possibly numerous, the user is difficult to find and locate, the user cannot quickly find and view the information at the specific broadcast position, and the actual experience of the user is influenced.
In order to solve the technical problem, in the embodiment of the application, after an application page is opened, when text information in the page is broadcasted, a dynamic focus is set on a page UI, and the focus is used for marking a specific position of current broadcast content in the page, that is, the focus needs to be matched with voice broadcast content in real time, and the broadcast content is tracked and moved dynamically.
In some embodiments, the focus may be marked in units of words, characters, words, sentences or paragraphs, or by rows, etc.
In some embodiments, the specific form of focus is, for example, highlighting the marked text; or, a contrast color different from other texts is set for the marked text, if the text marked by the focus is set to be red, the content of other texts is black, and the user can directly position the current broadcast content at the page position through the color difference. It should be noted that the presentation form of the focus is not limited to that described in this embodiment, and other display effects, such as text flashing, text bolding, font difference, font size difference, or positioning identifier, may also be set in practical applications.
As shown in fig. 8, taking an electronic Manual (E-Manual) application built in a display device as an example, when a user opens a page in the electronic Manual, a text broadcast function is executed synchronously, a mode of marking a focus by a line is adopted in fig. 8, when a voice broadcast is performed to "Hisense Mexico …," Hisense Mexico s.de r.l.de c.v. "of the line to which the content belongs may be marked, that is, the focus moves to the position along with the mark, the focus effect in fig. 8 is to set a highlight mark, and the highlight mark may be in different colors, for example, green highlight, so that the focus is presented more intuitively and vividly.
Because the amount of page information is large, and the amount of text information displayed to audiences on the display interface is limited due to the limitation of the screen size, when the focus moves along with the broadcast content, if a user does not operate the scrolling page, the focus may move to the non-visual information position of the user subsequently, that is, the focus cannot be seen by naked eyes on the current display interface. To this end, in some embodiments, in a text voice broadcasting process, the method detects the focus position in real time, when the focus is about to exceed a page visualization range, a user does not need to manually move the focus or slide down the page, but controls the page to automatically scroll upwards according to a visualization window, so that non-visible text information in the previous page is scrolled into the visualization window to be displayed, and meanwhile, a 'text block' marked by the focus is scrolled to a longitudinally central position of the page, so that the focus can be adjusted to an optimal central viewing position from a critical visualization position, and user experience is improved.
As shown in fig. 9, 3 sentences can be currently displayed in the page visualization window, corresponding to Tuner Mode, Auto Scan and DTV Manual Scan, respectively. As in (a) of fig. 9, the current focus moves to Tuner Mode, and the focus marks the whole sentence corresponding to Tuner Mode; then, the focus moves along with the broadcast content conversion statement, referring to (b) in fig. 9, the focus moves to the Auto Scan, and the focus marks the whole statement corresponding to the Auto Scan; referring to (c) in fig. 9, after the report switches the sentence again, the focus moves to the DTV Manual Scan, the focus marks the whole sentence corresponding to the DTV Manual Scan, and at this time, the focus has moved to the bottom critical position of the visualization window, and if the focus continues to move down and the page is not moved, the user cannot see the focus, so in this application, the page is automatically controlled to scroll up, as shown in (d) in fig. 9, and after the page is scrolled up, the focus moves to the ATV Manual Scan at the center of the page.
In some embodiments, as shown in fig. 10, there is provided a page information voice broadcasting method, which is configured to be executed by a controller 250 in a display device, that is, the controller 250 is an execution subject of the method, and the method includes the following program steps:
in response to the operation of the start page, in step S10, it is detected whether the current state of the text-cast function is turned on. If the text broadcasting function is turned on, step S20 is performed, and otherwise, step S60 is performed.
And step S20, displaying the page and synchronously broadcasting the text information in the page.
And step S30, following the broadcast content, controlling the focus to dynamically move in the page.
And step S40, detecting whether the focus moves to a first preset position of a display interface.
If the focus is moved to the first preset position, step S50 is performed, otherwise, steps S30 and S40 are returned to. In some embodiments, the first preset position may be a bottom critical position of the visualization of the display interface, that is, a position in the aforementioned visualization window where the bottom is about to exceed the visualization range, in such a way that the visualization boundary is used as a constraint to control page scrolling and focus position adjustment. The first preset position can be set according to practical application, for example, if the focus is marked according to a line, the first preset position can be a line position corresponding to the Nth line in a voice broadcast mode, and the first preset position is not limited.
And step S50, controlling the page to scroll upwards so that the focus is at a second preset position of the display interface.
In some embodiments, the second preset position may be a middle position of the display interface along the longitudinal direction, and the visual viewing effect of the middle position of the interface is the best, so that after the condition of step S40 is satisfied, the page is controlled to scroll upwards, and since the scrolling speed on the page is greater than the voice broadcast speed, the focus may be driven to move upwards relatively, so as to adjust the focus to the page middle position. The second preset position may be set according to practical applications, and this embodiment is not limited. After the page and the focus are adjusted by step S50, the voice broadcast and the focus follow are continuously performed from step S30.
Step S60, only the page is displayed. When the text broadcasting function is not started, only the page is normally displayed, and text information in the page is not broadcasted.
In some embodiments, the focus is marked in units of characters, words, sentences or paragraphs, or marked according to lines, and the like, specifically based on requirements of practical applications.
In order to enhance the user experience, and preferably maintain relative consistency between the audio and visual characteristics, it is desirable to have the broadcasted voice information in the same language as the page text information. For the display device, a user can select a target language used by a native machine by default in Setting (Setting) according to a use environment, a geographic position and/or a personal language habit, after the target language is set, both a native machine UI and voice can use the target language, and after the page is started, the text information is broadcasted according to the target language.
In other implementation manners, a user can also be supported to set a target language used when the text is converted into the voice, for example, the default language of the local computer is chinese, the user sets the target language to be english, that is, the text information displayed on the page is chinese, and the text information is broadcasted by using english voice.
According to the technical scheme, the technical scheme is suitable for the display equipment with the TTS function, when a user starts a certain page, the current state of the text broadcast function (namely the TTS function) needs to be detected besides the displayed page, and the state can comprise opened state or unopened state. If the text broadcasting function is not started, only displaying the page without text conversion and voice output; if the text broadcast function is currently in an opened state, the text information in the page is converted into voice information for broadcast while the page is displayed. The key point of the method is that a focus for marking the position of the current broadcast content in the page is added, the focus can move synchronously and dynamically along with the broadcast content, namely the information is broadcasted by voice, the focus moves to the position of the information, so that the information can obtain the focus, and the focus adopts the forms of thickening, highlighting or color changing and the like on the information, so that a user can conveniently and visually and quickly position the broadcast position through the focus; in addition, when the focus moves to the limited critical position of the first preset position, in order to guarantee the visibility and the display effect of the focus, the page is automatically controlled to roll upwards, so that the current focus is located at the position between two sides of the page, the user use experience is further promoted, and intelligent and automatic voice broadcasting is realized.
Those skilled in the art will readily appreciate that the techniques of the embodiments of the present invention may be implemented as software plus a required general purpose hardware platform. In a specific implementation, the invention also provides a computer storage medium, which can store a program. When the computer storage medium is located in the display device 200, the program may be executed to include the program steps related to the aforementioned page information voice broadcasting method that the controller 250 is configured to execute. The computer storage medium may be a magnetic disk, an optical disk, a Read-Only Memory (ROM) or a Random Access Memory (RAM).
In this specification, the same and similar parts between the display device embodiment and the method embodiment may be referred to each other, and related contents are not described again.
Finally, it should be noted that: the above embodiments are only used for illustrating the technical solutions of the present application, and not for limiting the same; although the present application has been described in detail with reference to the foregoing embodiments, it should be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some or all of the technical features may be equivalently replaced; and the modifications or the substitutions do not make the essence of the corresponding technical solutions depart from the scope of the technical solutions of the embodiments of the present application.
The foregoing description, for purposes of explanation, has been presented in conjunction with specific embodiments. However, the illustrative discussions above are not intended to be exhaustive or to limit the embodiments to the precise forms disclosed above. Many modifications and variations are possible in light of the above teaching. The embodiments were chosen and described in order to best explain the principles and the practical application, to thereby enable others skilled in the art to best utilize the embodiments and various embodiments with various modifications as are suited to the particular use contemplated.

Claims (10)

1. A display device, comprising:
a display for displaying a page;
the sound player is used for broadcasting the text information in the page in a voice way;
a controller configured to perform:
responding to the operation of starting the page, controlling a display to display the page, and controlling a sound player to synchronously broadcast text information in the page when detecting that the current state of a text broadcast function is started;
controlling the focus to dynamically move in the page along with the broadcast content; the focus is used for marking the position of the current broadcast content in the page.
2. The display device according to claim 1, wherein the controller is further configured to perform:
and if the focus is detected to move to a first preset position of a display interface, controlling the page to scroll upwards so as to enable the focus to be at a second preset position of the display interface.
3. The display device according to claim 2, wherein the first predetermined position is a bottom critical position of a visualization of the display interface.
4. The display device according to claim 2 or 3, wherein the second preset position is a central position of the display interface along the longitudinal direction.
5. The display device according to claim 1, wherein the focus is marked in units of a word, a character, a word, a sentence, or a paragraph.
6. The display device of claim 1, wherein the controller controls the sound player to synchronously play the text information in the page as follows:
acquiring text information loaded in the page and a target language corresponding to the text information;
and controlling a sound player to broadcast the text information according to the target language.
7. A page information voice broadcasting method is characterized by comprising the following steps:
responding to the operation of starting a page, displaying the page, and synchronously broadcasting text information in the page when detecting that the current state of a text broadcasting function is started;
controlling the focus to dynamically move in the page along with the broadcast content; the focus is used for marking the position of the current broadcast content in the page.
8. The method of claim 7, further comprising:
and if the focus is detected to move to a first preset position of a display interface, controlling the page to scroll upwards so as to enable the focus to be at a second preset position of the display interface.
9. The method of claim 8, wherein the first predetermined location is a bottom critical location of a visualization of the display interface.
10. The method according to claim 8 or 9, wherein the second preset position is a central position of the display interface along the longitudinal direction.
CN202011413371.4A 2020-05-29 2020-12-04 Page information voice broadcasting method and display device Pending CN112506400A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202011413371.4A CN112506400A (en) 2020-12-04 2020-12-04 Page information voice broadcasting method and display device
PCT/CN2021/081327 WO2021238353A1 (en) 2020-05-29 2021-03-17 Display method and display device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011413371.4A CN112506400A (en) 2020-12-04 2020-12-04 Page information voice broadcasting method and display device

Publications (1)

Publication Number Publication Date
CN112506400A true CN112506400A (en) 2021-03-16

Family

ID=74970529

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011413371.4A Pending CN112506400A (en) 2020-05-29 2020-12-04 Page information voice broadcasting method and display device

Country Status (1)

Country Link
CN (1) CN112506400A (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113066491A (en) * 2021-03-18 2021-07-02 海信视像科技股份有限公司 Display device and voice interaction method
CN113079400A (en) * 2021-03-25 2021-07-06 海信视像科技股份有限公司 Display device, server and voice interaction method
CN113190697A (en) * 2021-06-02 2021-07-30 口碑(上海)信息技术有限公司 Image information playing method and device
CN113360127A (en) * 2021-05-31 2021-09-07 富途网络科技(深圳)有限公司 Audio playing method and electronic equipment
WO2021238353A1 (en) * 2020-05-29 2021-12-02 海信视像科技股份有限公司 Display method and display device
CN114302021A (en) * 2021-04-30 2022-04-08 海信视像科技股份有限公司 Display device and sound picture synchronization method
CN114564265A (en) * 2021-12-22 2022-05-31 上海小度技术有限公司 Interaction method and device of intelligent equipment with screen and electronic equipment

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102324191A (en) * 2011-09-28 2012-01-18 Tcl集团股份有限公司 Method and system for synchronously displaying audio book word by word
CN103533519A (en) * 2012-07-06 2014-01-22 盛乐信息技术(上海)有限公司 Short message broadcasting method and system
WO2016058847A1 (en) * 2014-10-13 2016-04-21 Thomson Licensing Method for controlling the displaying of text for aiding reading on a display device, and apparatus adapted for carrying out the method, computer program, and computer readable storage medium
CN106648291A (en) * 2016-09-28 2017-05-10 珠海市魅族科技有限公司 Method and device for displaying information and broadcasting information
CN111324330A (en) * 2020-02-07 2020-06-23 掌阅科技股份有限公司 Electronic book playing processing method, computing device and computer storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102324191A (en) * 2011-09-28 2012-01-18 Tcl集团股份有限公司 Method and system for synchronously displaying audio book word by word
CN103533519A (en) * 2012-07-06 2014-01-22 盛乐信息技术(上海)有限公司 Short message broadcasting method and system
WO2016058847A1 (en) * 2014-10-13 2016-04-21 Thomson Licensing Method for controlling the displaying of text for aiding reading on a display device, and apparatus adapted for carrying out the method, computer program, and computer readable storage medium
CN106648291A (en) * 2016-09-28 2017-05-10 珠海市魅族科技有限公司 Method and device for displaying information and broadcasting information
CN111324330A (en) * 2020-02-07 2020-06-23 掌阅科技股份有限公司 Electronic book playing processing method, computing device and computer storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
PONECENTER: "小米5屏幕阅读(TalkBack)功能怎么用", 《百度经验》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021238353A1 (en) * 2020-05-29 2021-12-02 海信视像科技股份有限公司 Display method and display device
CN113066491A (en) * 2021-03-18 2021-07-02 海信视像科技股份有限公司 Display device and voice interaction method
CN113079400A (en) * 2021-03-25 2021-07-06 海信视像科技股份有限公司 Display device, server and voice interaction method
CN114302021A (en) * 2021-04-30 2022-04-08 海信视像科技股份有限公司 Display device and sound picture synchronization method
CN113360127A (en) * 2021-05-31 2021-09-07 富途网络科技(深圳)有限公司 Audio playing method and electronic equipment
CN113190697A (en) * 2021-06-02 2021-07-30 口碑(上海)信息技术有限公司 Image information playing method and device
CN114564265A (en) * 2021-12-22 2022-05-31 上海小度技术有限公司 Interaction method and device of intelligent equipment with screen and electronic equipment
CN114564265B (en) * 2021-12-22 2023-07-25 上海小度技术有限公司 Interaction method and device of intelligent equipment with screen and electronic equipment

Similar Documents

Publication Publication Date Title
CN112506400A (en) Page information voice broadcasting method and display device
CN112672195A (en) Remote controller key setting method and display equipment
CN112580302B (en) Subtitle correction method and display equipment
CN113421532B (en) Backlight adjusting method and display device
CN113014939A (en) Display device and playing method
CN114302021A (en) Display device and sound picture synchronization method
CN112752156A (en) Subtitle adjusting method and display device
CN113395556A (en) Display device and method for displaying detail page
CN112188249A (en) Electronic specification-based playing method and display device
CN113593488A (en) Backlight adjusting method and display device
CN113473241A (en) Display equipment and display control method of image-text style menu
CN113490024A (en) Control device key setting method and display equipment
WO2021109411A1 (en) Text type conversion method and display device
CN113132809B (en) Channel switching method, channel program playing method and display equipment
WO2021238353A1 (en) Display method and display device
CN113064691B (en) Display method and display equipment for starting user interface
CN113434240B (en) Display method and display device of image mode
CN113703705A (en) Display device and list updating method
CN113992960A (en) Subtitle previewing method on display device and display device
CN113286185A (en) Display device and homepage display method
CN113064534A (en) Display method and display equipment of user interface
CN114296664A (en) Auxiliary screen brightness adjusting method and display device
CN114302070A (en) Display device and audio output method
CN112668546A (en) Video thumbnail display method and display equipment
CN113766164B (en) Display equipment and signal source interface display method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20210316