CN112163433B - Key vocabulary matching method and device, electronic equipment and storage medium - Google Patents

Key vocabulary matching method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN112163433B
CN112163433B CN202011052749.2A CN202011052749A CN112163433B CN 112163433 B CN112163433 B CN 112163433B CN 202011052749 A CN202011052749 A CN 202011052749A CN 112163433 B CN112163433 B CN 112163433B
Authority
CN
China
Prior art keywords
target
language type
content
determining
key
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011052749.2A
Other languages
Chinese (zh)
Other versions
CN112163433A (en
Inventor
陈可蓉
钱程
杨晶生
熊梦园
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Zitiao Network Technology Co Ltd
Original Assignee
Beijing Zitiao Network Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Zitiao Network Technology Co Ltd filed Critical Beijing Zitiao Network Technology Co Ltd
Priority to CN202011052749.2A priority Critical patent/CN112163433B/en
Publication of CN112163433A publication Critical patent/CN112163433A/en
Application granted granted Critical
Publication of CN112163433B publication Critical patent/CN112163433B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/40Processing or translation of natural language
    • G06F40/58Use of machine translation, e.g. for multi-lingual retrieval, for server-side translation for client devices or for real-time translation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/284Lexical analysis, e.g. tokenisation or collocates
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel

Abstract

The embodiment of the disclosure discloses a matching method and device of key words, electronic equipment and a storage medium, wherein the method comprises the following steps: responding to a trigger operation of a target key word presented on a target page, and determining a target original language type of the target key word; determining an original text with the same type as the target original language in subtitle information presented on a target page; determining contents to be matched with the target key words from the original text; and determining target content corresponding to the content to be matched in the subtitle information, and displaying the target content in a target page in a distinguishing manner. According to the technical scheme of the embodiment of the disclosure, when the key words are triggered, the technical effect of the target key words can be quickly and accurately searched from the generated subtitle information.

Description

Key vocabulary matching method and device, electronic equipment and storage medium
Technical Field
The embodiment of the disclosure relates to the technical field of computers, in particular to a matching method and device of key words, electronic equipment and a storage medium.
Background
At present, when a certain language type of text is translated into another language type of text, the contexts of some words are different, which results in different forms of translated words, for example, when a chinese sentence is translated into english, a time state corresponding to a translated chinese word may be passive, current, and the like, that is, one chinese word corresponds to a plurality of english words.
If the target content different from the language type of the text needs to be searched from the translated text, the corresponding target content cannot be matched, so that the technical problems of low searching efficiency and poor user experience exist.
Disclosure of Invention
The embodiment of the disclosure provides a matching method and device of key words, electronic equipment and a storage medium, so as to achieve the technical effect of finding the accuracy of target key words from subtitle information.
In a first aspect, an embodiment of the present invention provides a method for matching a key vocabulary, where the method includes:
responding to a trigger operation of a target key word presented on a target page, and determining a target original language type of the target key word;
determining an original text which is the same as the target original language type in the subtitle information presented on the target page;
determining contents to be matched with the target key words from the original text;
and determining target content corresponding to the content to be matched in the subtitle information, and displaying the target content in the target page in a distinguishing manner.
In a second aspect, an embodiment of the present invention further provides an apparatus for matching a key vocabulary, where the apparatus includes:
the target original language type determining module is used for responding to triggering operation of target key words presented on a target page and determining the target original language type of the target key words;
an original text determining module, configured to determine an original text that is the same as the target original language type in the subtitle information presented on the target page;
the content to be matched determining module is used for determining the content to be matched with the target key vocabulary from the original text;
and the target content determining module is used for determining the target content corresponding to the content to be matched in the subtitle information and displaying the target content in the target page in a distinguishing manner.
In a third aspect, an embodiment of the present disclosure further provides an electronic device, where the electronic device includes:
one or more processors;
a storage device for storing one or more programs,
when executed by the one or more processors, cause the one or more processors to implement a method for matching key words as described in any of the embodiments of the present disclosure.
In a fourth aspect, the embodiments of the present disclosure further provide a storage medium containing computer-executable instructions, which when executed by a computer processor, are used for executing the matching method of the key vocabulary according to any one of the embodiments of the present disclosure.
According to the technical scheme of the embodiment of the disclosure, when the target key vocabulary is triggered, the original language type of the obtained key vocabulary can be determined, the original text with the same language type as the original language type in the subtitle information is obtained, the content to be matched with the target key vocabulary with the same type as the original text is further obtained, the target content with the same type as the content to be matched is searched from the subtitle information, and the accuracy of the target content determined from the subtitle information is improved.
Drawings
The above and other features, advantages and aspects of various embodiments of the present disclosure will become more apparent by referring to the following detailed description when taken in conjunction with the accompanying drawings. Throughout the drawings, the same or similar reference numbers refer to the same or similar elements. It should be understood that the drawings are schematic and that elements and features are not necessarily drawn to scale.
Fig. 1 is a schematic flowchart of a key word matching method according to an embodiment of the present disclosure;
fig. 2 is a schematic flowchart of a key word matching method according to a second embodiment of the present disclosure;
fig. 3 is a schematic flowchart of a key word matching method according to a third embodiment of the present disclosure;
fig. 4 is a schematic diagram of a display of target content and a mark on a time axis according to a third embodiment of the disclosure;
fig. 5 is a schematic diagram illustrating a corresponding mark on a time axis being highlighted after the target content is triggered according to a third embodiment of the present disclosure;
fig. 6 is a schematic structural diagram of a matching apparatus for key words according to a fourth embodiment of the present disclosure;
fig. 7 is a schematic structural diagram of an electronic device according to a fifth embodiment of the disclosure.
Detailed Description
Embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While certain embodiments of the present disclosure are shown in the drawings, it is to be understood that the present disclosure may be embodied in various forms and should not be construed as limited to the embodiments set forth herein, but rather are provided for a more thorough and complete understanding of the present disclosure. It should be understood that the drawings and embodiments of the disclosure are for illustration purposes only and are not intended to limit the scope of the disclosure.
It should be understood that the various steps recited in the method embodiments of the present disclosure may be performed in a different order, and/or performed in parallel. Moreover, method embodiments may include additional steps and/or omit performing the illustrated steps. The scope of the present disclosure is not limited in this respect.
The term "include" and variations thereof as used herein are open-ended, i.e., "including but not limited to". The term "based on" is "based, at least in part, on". The term "one embodiment" means "at least one embodiment"; the term "another embodiment" means "at least one additional embodiment"; the term "some embodiments" means "at least some embodiments". Relevant definitions for other terms will be given in the following description.
It should be noted that the terms "first", "second", and the like in the present disclosure are only used for distinguishing different devices, modules or units, and are not used for limiting the order or interdependence relationship of the functions performed by the devices, modules or units.
It is noted that references to "a", "an", and "the" modifications in this disclosure are intended to be illustrative rather than limiting, and that those skilled in the art will recognize that "one or more" may be used unless the context clearly dictates otherwise.
Example one
Fig. 1 is a flowchart illustrating a method for matching a keyword according to an embodiment of the present disclosure, where the embodiment of the present disclosure is applied to a situation where content matched with a keyword is screened from subtitle information, and the method may be performed by a keyword matching apparatus, which may be implemented in software and/or hardware.
As shown in fig. 1, the method of the present embodiment includes:
s110, responding to the trigger operation of the target key words presented on the target page, and determining the target original language type of the target key words.
Wherein the subtitle information is generated based on the multimedia data stream. The multimedia data stream may be an audio frame or a video frame, and accordingly, the subtitle information may be generated based on the audio frame and the video frame corresponding to the multimedia data stream. Optionally, the subtitle information is obtained by performing speech recognition on the audio frame. The target page includes subtitle information generated based on the audio frame. In the process of generating the caption information, corresponding core words can be extracted according to the voice information to be used as key words and displayed on a target page. That is, the target page is a page displaying the caption information and the key vocabulary. The core vocabulary can be used as the key vocabulary of the caption information, and correspondingly, the key vocabulary currently triggered by the user is used as the target key vocabulary. The target original language type refers to an initial language type corresponding to the target key vocabulary. For example, the target key vocabulary is determined based on phonetic information having a language type of Chinese, and the target original language type may be Chinese.
Specifically, when it is detected that the user triggers one of the key words on the target page, the triggered key word may be used as the target key word. And meanwhile, acquiring the original language type of the target key vocabulary as the target original language type.
Illustratively, the key words on the target page include key word A, key word B … and key word N. When the fact that a user triggers the key vocabulary A is detected, the key vocabulary A is a target key vocabulary, and the target original language type of the target key vocabulary A is called.
In this embodiment, in response to a trigger operation on a target key word presented on a target page, determining a target original language type of the target key word may be: when detecting a key word to be triggered displayed on a trigger target page, determining the target key word; determining the original language type of the target key vocabulary according to the mapping relation and the target key vocabulary; and generating the target original language type based on the original language type.
Specifically, when the user triggers the key vocabulary to be triggered displayed on the target page, the triggered key vocabulary to be triggered is used as the target key vocabulary. According to the pre-established corresponding relation between the key vocabulary to be triggered and the original language type, the original language type corresponding to the target key vocabulary is determined, and the original language type is used as the target original language type, so that the technical effects of convenience and high efficiency in determining the target original language type of the target key vocabulary are improved.
Illustratively, the key words to be triggered on the target page include key word a, key word B … and key word N. The mapping relation between the key words and the original language can be key words A-Chinese, and key words B-English … key words N-French. When the user triggers the key vocabulary A to be triggered, the key vocabulary A to be triggered is a target key vocabulary. According to the mapping relation between the key vocabulary and the original language, the target original language type of the target key vocabulary A can be determined to be Chinese.
And S120, determining an original text with the same type as the target original language in the subtitle information presented on the target page.
It should be noted that, the target page includes a language type selection control, and the user may trigger the language type selection control to set the language type of the subtitle information and the key vocabulary displayed on the target page.
Illustratively, the multimedia data stream includes voice information of a plurality of users, and the original language type corresponding to all the voice information may be chinese, japanese, english, french, or the like. If the user does not trigger the language type selection control, the subtitle information corresponding to the voice information can be displayed on the target page in the original language type, for example, the original language type is Chinese, the subtitle information corresponding to the voice information is also Chinese, the original language type is Japanese, the subtitle information corresponding to the voice information is Japanese, and correspondingly, the key vocabulary is displayed on the target page in the original language type. If the user triggers the language type selection control and the selected language type is Chinese, all the voice information and the key words can be translated into Chinese and displayed on the target page.
Therefore, on the basis, after the target key vocabulary is triggered and the target original language type corresponding to the target key vocabulary is determined, the original language type of each voice message of the caption information can be obtained, the text corresponding to the target original language type is called, and the text obtained at the moment can be used as the original text. I.e., the original text is determined based on the speech information of the target original language type.
Illustratively, the language type corresponding to the voice message is Chinese, Japanese, English. If the user does not trigger the language type selection control on the target page, the subtitle information and the key words displayed on the target page can be Chinese, Japanese, English and the like. If the user triggers the language type selection control on the target page and the triggered language type is French, the voice information corresponding to Chinese, Japanese and English needs to be translated into translation data corresponding to French and displayed on the target page as caption information, and meanwhile, key words also need to be translated into French. After a user triggers the key vocabulary, determining that the original language type of the key vocabulary is Chinese, and acquiring an original text generated by the original language type being Chinese.
That is, the original language type corresponding to the subtitle information is obtained according to the target original language type, and the text which is the same as the target original language type is called as the original text.
And S130, determining the content to be matched with the target key vocabulary from the original text.
Specifically, the content identical to the target key vocabulary can be searched from the original text, and the searched content is used as the content to be matched.
S140, determining the target content corresponding to the content to be matched in the subtitle information, and displaying the target content in a target page in a distinguishing manner.
Specifically, the target content refers to the content which is found from the subtitle information and is the same as the content to be matched. When the target content is obtained through screening, the target content is one or more elements in the subtitle information, and can be displayed in a distinguishing manner with other elements when being displayed, so that the screened target content is highlighted, and a user can find the target content more intuitively and conveniently. The distinctive display may be in a display format such as color, font, background pattern, or the like.
According to the technical scheme of the embodiment of the disclosure, when the target key vocabulary is triggered, the original language type of the obtained key vocabulary can be determined, the original text with the same language type as the original text in the subtitle information is obtained, the content to be matched with the target key vocabulary with the same language type as the original text is further obtained, the target content with the same content to be matched is searched from the subtitle information, and the obtained target content is completely corresponding to the target key vocabulary, so that the accuracy of determining the target content from the subtitle information is improved.
On the basis of the technical scheme, the method further comprises the following steps: determining voice information and recording an original language type corresponding to the voice information based on the multimedia data stream; and generating caption information which is displayed on a target page and corresponds to the target translation language type according to the voice information, the original language type corresponding to the voice information and the target translation language type.
The multimedia data stream may be video stream data corresponding to the real-time interactive interface, or video stream data in a screen recording video after the screen recording is performed on the real-time interactive interface. The real-time interactive interface is any interactive interface in a real-time interactive application scene. The real-time interactive scene can be realized by the internet and a computer means, for example, an interactive application program realized by a native program or a web program.
It should be further noted that a plurality of users participating in real-time interaction or recording a video may be included, the language type used by each user for speaking may be the same or different, and when the language type used by other speaking users is greatly different from the language type used by the other speaking users, there may be a case where the speaking information of the other speaking users cannot be known.
To solve this problem, voice information of the speaking user may be collected and converted into corresponding subtitle information. Of course, in order to improve the convenience of reading, the user may trigger the language type selection control on the target page and select the translation language type, so as to translate the voice information of each speaking user into the subtitle information corresponding to the selected target translation language type.
The original language type refers to a language type used by a user participating in real-time interaction during speaking. The target translation language type is a language type set on the target page by the user for displaying the caption information. In this case, if the user selects the target translation language type, the subtitle information presented on the target page may be translation data corresponding to the voice information in the selected target translation language type.
Furthermore, in order to facilitate the user to visually determine the speaking user and the speaking time corresponding to each piece of translation data from the caption information, the caption information may display the speaking user identification and the speaking time stamp of each piece of translation data.
Specifically, the voice data, i.e., the voice information, of each interactive user participating in the interaction may be collected from the multimedia data stream corresponding to the interactive behavior interface, and the original language type corresponding to the voice information may be recorded. The voice information can be translated into a target translation language type from an original language type to obtain translation data corresponding to the voice information. And generating caption information displayed on the target page by the translation data, the speaking user identity corresponding to the translation data and the speaking timestamp.
On the basis of the technical scheme, the method further comprises the following steps: determining at least one key vocabulary to be processed corresponding to the multimedia data stream and an original language type of the key vocabulary to be processed by performing content extraction processing on the multimedia data stream, wherein the key vocabulary to be processed is a core vocabulary extracted based on the multimedia data stream; converting at least one key vocabulary to be processed from an original language type into a target translation language type to obtain at least one key vocabulary to be triggered, and displaying the key vocabulary to be triggered on a target page; the target key vocabulary is one of at least one key vocabulary to be triggered.
The core vocabulary of the multimedia data stream can be generated according to the audio information of the multimedia data stream, and the vocabulary is used as the key vocabulary to be processed. Because the voice information in the acquired multimedia data stream is of the original language type, the language type of the obtained key vocabulary to be processed is also of the original language type. And translating the key vocabulary to be processed from the original language type into the target translation language type to obtain the key vocabulary which is used as the key vocabulary to be triggered. The key vocabulary to be triggered can be displayed on the target page so as to be triggered by the target user.
It should be noted that the target key vocabulary is one of the key vocabularies to be triggered.
In addition, if the user does not select the target translation language type, the directly generated core vocabulary can be displayed on the target page as the key vocabulary to be triggered.
Specifically, while generating the subtitle information based on the multimedia data stream, key information of the multimedia data stream may be extracted, for example, the key information may be information extracted from information discussed by the participating user for a long time; or determining key information according to the occurrence frequency of the words after word segmentation processing is carried out on the subtitle information; after word segmentation processing, word vectors corresponding to each vocabulary are determined, an average word vector is determined based on the word vectors of each vocabulary, and key information is determined according to the similarity between the word vectors of each vocabulary and the average word vector. The key vocabulary to be processed may be generated based on the key information. Because the language type of the key vocabulary to be processed is the same as the language type of the speaking user in the multimedia data stream, the key vocabulary to be processed can be translated from the original language type to the target translation language type to obtain the key vocabulary to be triggered in order to facilitate the reading of the user.
In this embodiment, determining the type of the target translation language may include at least one of the following ways: acquiring a language type preset on a target client as a target translation language type; and acquiring a login address of the target client, and determining a target language translation type corresponding to the geographical position of the target client based on the login address.
That is, determining the target translation language type may include at least two ways. The first way may be: when it is detected that the user triggers the language type selection control on the target page, that is, the language type in which the caption information is displayed, the converted language type can be set, and the set language type is used as the target translation language type. For example, when the user triggers the language type conversion control, a language type selection list may pop up on the target page for the user to select. The user can select any language type, for example, the user triggers a Chinese language type in the language selection list and clicks a confirmation key, and the server or the client can determine that the target translation language type is the Chinese language type. That is, the voice information in the multimedia data stream may be converted into chinese subtitle information and presented on the target interface. The second way may be: if the fact that the user triggers the language conversion control is detected, the login address of the client, namely the IP address of the client, can be obtained, the area to which the client belongs is determined according to the login address, and then the language type used by the area to which the client belongs is used as the target language type. For example, when the user triggers a language conversion control, the login address of the client is obtained, and if the region to which the client belongs is determined to be Chinese based on the login address, the type of the target language is Chinese.
In this embodiment, the voice information in the multimedia data stream is converted into the subtitle information in the target translation language type, so that the subtitle information better conforms to the reading habit of the user, and the user can quickly understand the content corresponding to the multimedia data stream, thereby improving the technical effect of interactive interaction efficiency.
On the basis of the above technical solution, after the converting the at least one to-be-processed key word from the original language type to the target translation language type to obtain at least one to-be-triggered key word, the method further includes: recording the mapping relation between the at least one key word to be triggered and the corresponding original language type, determining the target key word when the key word to be triggered is detected, and determining the target original language type corresponding to the target key word based on the mapping relation.
It can be understood that, when the key vocabulary to be triggered is obtained, the original language type of the key vocabulary to be triggered may be recorded, for example, if the target translation language type of the key vocabulary to be triggered is japanese, and the original language type is chinese, the original language type of the key vocabulary to be triggered is recorded.
By adopting the method, when any key vocabulary in the key vocabulary to be triggered is detected, the target original language type of the target key vocabulary can be determined according to the pre-recorded original language type, and the technical effect of determining the target original language type efficiency is improved.
Example two
Fig. 2 is a flowchart illustrating a key word matching method according to a second embodiment of the present invention. On the basis of the foregoing embodiment, before determining content to be matched that is matched with a target key vocabulary from an original text, the target key vocabulary may be further converted from the target translation language type to a vocabulary corresponding to the target original language type to obtain a key vocabulary to be matched, so as to determine the content to be matched from the original text content according to the key vocabulary to be matched, thereby improving the technical effect of determining the content to be matched from the original text. The technical terms that are the same as or corresponding to the above embodiments are not repeated herein.
As shown in fig. 2, the method includes:
s210, responding to a trigger operation of the target key words presented on the target page, and determining the target original language type of the target key words.
S220, converting the target key vocabulary from the target translation language type into the vocabulary corresponding to the target original language type to obtain the key vocabulary to be matched, and determining the content to be matched from the original text content according to the key vocabulary to be matched.
And the language type of the subtitle information displayed on the target page is a target translation language type. And translating the target key vocabulary into the target original language type according to the target original language type to obtain at least one key vocabulary to be matched.
In this embodiment, after determining the key vocabulary to be processed, the original language type of the key vocabulary to be processed and the key vocabulary to be triggered in the target translation language type may be determined. After the target key vocabulary is determined, the key vocabulary to be processed of the target key vocabulary in the target original language type, namely the above mentioned key vocabulary to be matched, can be called.
And S230, determining an original text which is the same as the target original language type in the subtitle information presented on the target page.
Optionally, an original language type corresponding to each piece of translation data in the subtitle information is determined, and an original text that is the same as the target original language type and corresponds to the subtitle information is obtained.
Specifically, the caption information includes text generated based on the voice information of a plurality of users. The language type of the character can be the original language type of the user or the same language type as the target translation language type. If the language type of the caption information is consistent with the original language type of each user, the character information with the same type as the target original language type can be directly acquired, and the acquired character information is used as the original text. And if the character language type of the caption information is translation data translated into a target translation language type, determining an original language type corresponding to each piece of translation data in the caption information, and acquiring data with the same type as the target original language type from all the original language types as an original text.
S240, determining the content to be matched with the target key vocabulary from the original text content.
In this embodiment, the to-be-matched key vocabulary includes at least one to-be-searched related content corresponding to the target key vocabulary in the target original language type. And searching the content which is the same as at least one associated content to be searched of the key words to be matched from the original text content, and taking the content as the content to be matched.
It should be noted that different target key words exist in the context in the forms of active, passive, etc. In order to accurately find the content corresponding to the target key vocabulary from the subtitle information, various conversion forms of the target key vocabulary under the target original language type can be determined, and the associated content to be searched is obtained, so that the target content can be determined from the subtitle information based on the associated content to be searched.
Specifically, after determining the key vocabulary to be matched, various transformation forms of the key words to be matched under the target original language type can be determined, optionally, the key words to be matched are '123', the transformation forms can be 'one, two, three', and the like, and the transformation forms can be used as the associated content to be searched. And acquiring the content which is the same as the content with the search correlation from the original text, and taking the content as the content to be matched.
S250, determining the target content corresponding to the content to be matched in the subtitle information, and displaying the target content in the subtitle information in a distinguishing manner.
Optionally, determining at least one target translation data of the content to be matched in the target translation language type; and determining target content corresponding to the at least one target translation data from the subtitle information.
Specifically, the content to be matched is converted into corresponding target translation data according to the language type of the target translation. And matching the content identical to the target translation data in the subtitle information to obtain the target content so as to improve the accuracy of the matched target content.
Optionally, the target content corresponding to the content to be matched is determined from the subtitle information according to the timestamp of the content to be matched and at least one target translation data.
In order to further quickly match the target key words from the subtitle information, time stamp information corresponding to the content to be matched can be determined, and target content which corresponds to the time stamp information and is matched with at least one target translation data is determined from the subtitle information based on the time stamp information and the corresponding at least one target translation data.
According to the technical scheme of the embodiment of the invention, the target key words are converted into the corresponding target original language type from the target translation language type, so that the corresponding content to be matched is matched from the original text, and the target content identical to the content to be matched is obtained from the subtitle information, so that the technical effects of determining the accuracy and convenience of the target content are improved.
EXAMPLE III
Fig. 3 is a schematic flow chart of a keyword matching method according to a third embodiment of the present disclosure. On the basis of the foregoing embodiment, after generating the caption information based on the multimedia data stream, a timestamp synchronization association relationship between the caption information and the multimedia data stream may be established, so that after detecting the triggered caption information, the multimedia data stream corresponding to the timestamp may be acquired based on the timestamp synchronization association relationship. The same or corresponding techniques as those in the above embodiments are not described again.
As shown in fig. 3, the method includes:
and S310, generating subtitle information based on the multimedia data stream.
Specifically, the multimedia data stream may be video stream data corresponding to the real-time interactive interface, or video stream data in the screen video after the screen recording is performed on the real-time interactive interface. The audio information, i.e. the voice information, in the multimedia data stream can be acquired, and the voice information is converted from the original language type to the translation data of the target translation language type, i.e. the caption information, and is displayed on the target page.
S320, establishing a timestamp synchronization relation between the subtitle information and the multimedia data stream.
In order to realize the linkage of the subtitle information and the multimedia data stream on the target page, a time stamp synchronization relationship between the multimedia data stream and the subtitle information can be established.
The timestamp synchronization association relationship can be understood as that the multimedia data stream and the subtitle information are linked based on time synchronization. When a certain piece of translation data in the subtitle information is triggered, a current timestamp corresponding to the translation data can be determined, based on a pre-established timestamp synchronization association relationship, a multimedia data stream corresponding to the current timestamp is skipped to, for example, if the multimedia data stream is acquired based on a screen recording video, the screen recording video is adjusted to play an audio/video frame corresponding to the current timestamp. Certainly, when the progress bar of the screen recording video is dragged to a certain audio/video frame, the current timestamp corresponding to the audio/video frame can be obtained, the translation data corresponding to the audio/video frame in the subtitle information is determined based on the pre-established timestamp synchronous association relationship, and the translation data can be displayed in a distinguishing manner, optionally, highlighted, and convenient for a user to confirm.
S330, responding to the trigger operation of the target key words presented on the target page, and determining the target original language type of the target key words.
S340, determining an original text with the same type as the target original language in the subtitle information presented on the target page.
And S350, determining the content to be matched with the target key vocabulary from the original text.
S360, determining the target content corresponding to the content to be matched in the subtitle information, and displaying the target content in the target page in a distinguishing manner.
And S370, when the triggering target content is detected, determining the current timestamp corresponding to the target content.
Specifically, after the target content is determined, the target content may be differentially displayed on the target page, and when the target content is detected to be triggered by the user, the current timestamp of the target content is determined.
And S380, according to the pre-established timestamp synchronization relationship and the current timestamp, skipping the multimedia data stream to an audio/video frame corresponding to the current timestamp.
Specifically, according to the current timestamp of the target content and the timestamp synchronization association relationship between the pre-established multimedia data stream and the subtitle information, the multimedia data stream is jumped to the audio/video frame corresponding to the current timestamp, so that the user can know the tone and state of the speaking user when the speaking user issues the voice information including the target content, and the technical effect of improving the interaction efficiency is further achieved.
According to the technical scheme of the embodiment of the disclosure, the timestamp synchronization relation between the subtitle information on the target page and the multimedia data stream is established, so that the timestamp synchronization linkage between the subtitle information and the multimedia data stream is realized, the synchronism of reading the text and watching the video by the user is improved, and the technical effect of user experience is improved.
On the basis of the technical scheme, the method further comprises the following steps: the time stamp of the multimedia data stream corresponding to each target content in the time axis is determined, and the position corresponding to the time stamp on the time axis is marked.
Optionally, the total duration corresponding to the multimedia data stream is 50min, and the time axis corresponding to the multimedia data stream is also 50 min.
Specifically, after the target content is determined, the timestamp corresponding to the target content may be determined according to the sentence to which the target content belongs. After the time stamp is determined, a specific position of the time stamp on the time axis may be determined and marked at that position, for example, under the specific position of the time axis, with a dot, or with a triangle, see fig. 4.
Illustratively, referring to fig. 4, when the search content edited by the user in the search content editing control is "algorithm", the same target content as the "algorithm" can be searched from the subtitle information and displayed distinctively, such as bold display, and the timestamp of the sentence to which the target content belongs is determined, and the target content is marked on the time axis corresponding to the multimedia data stream based on the timestamp, such as marked with a dot. The color, size and the like of the mark can be set by a user according to actual requirements, and are not limited herein.
It should be noted that in the search content editing control, the number of the target content may be displayed, for example, the total number displayed in the search content editing control is 12, see fig. 4.
In this embodiment, marking the audio/video frames corresponding to the target content on the time axis has the advantage that the user can clearly determine the specific position of the target content in the multimedia data stream according to the mark on the time axis, thereby improving the convenience of searching the corresponding target content.
It should be noted that the number of the target contents may be more than one, and correspondingly, the number of the marks on the time axis may be more than one, see fig. 4, the number of the target contents is 12, and the number of the marks on the time axis is also 12. Certainly, in order to facilitate the user to determine that the currently triggered target content is the first of all the target contents, the search content editing control further displays the sequence corresponding to the currently triggered target content.
On the basis of the technical scheme, the method further comprises the following steps: when trigger target content is detected, determining a target timestamp corresponding to the target content; and displaying the mark corresponding to the target timestamp in a distinguishing way.
Specifically, a user may trigger any one target content, and when the target content triggered by the user is detected, a timestamp (target timestamp) corresponding to the triggered target content may be determined, a target mark corresponding to the target timestamp on a time axis may be determined, and the target mark and other marks on the time axis are displayed in a differentiated manner, so as to highlight the target mark. For example, the target mark and the other marks are displayed differently in colors.
For example, referring to fig. 5, when the user triggers the target content corresponding to the tag 1, the target timestamp corresponding to the target content of the tag 1 may be determined, the tag corresponding to the tag 2 on the time axis may be determined according to the target timestamp, and the tag may be highlighted.
In this embodiment, when the target content is triggered, the mark corresponding to the target content is displayed differently on the time axis, which is beneficial to enabling the user to know the specific position of the triggered target content in the multimedia data stream, and improving the accuracy of determining the audio/video frame corresponding to the target content by the user.
Example four
Fig. 6 is a schematic diagram of a keyword matching apparatus according to a fourth embodiment of the disclosure. As shown in fig. 6, the apparatus includes: a target original language type determining module 410, an original text determining module 420, a content to be matched determining module 430 and a target content determining module 440.
The target original language type determining module 410 is configured to determine a target original language type of a target key word in response to a trigger operation on the target key word presented on a target page; an original text determining module 420, configured to determine an original text that is the same as the target original language type in the subtitle information presented on the target page; a to-be-matched content determining module 430, configured to determine, from the original text, to-be-matched content that matches the target key vocabulary; the target content determining module 440 is configured to determine target content corresponding to the content to be matched in the subtitle information, and display the target content in the target page differently.
According to the technical scheme of the embodiment of the disclosure, when the target key vocabulary is triggered, the original language type of the obtained key vocabulary can be determined, the original text with the same language type as the original language type in the subtitle information is obtained, the content to be matched with the target key vocabulary with the same type as the original text is further obtained, the target content with the same type as the content to be matched is searched from the subtitle information, and the technical effect of determining the accuracy of the target content from the subtitle information is improved.
On the basis of the above technical solution, the apparatus further includes:
the voice information acquisition module is used for determining voice information and recording an original language type corresponding to the voice information based on the multimedia data stream; and the caption information generating module is used for generating caption information which is displayed on a target page and corresponds to the target translation language type according to the voice information, the original language type corresponding to the voice information and the target translation language type.
On the basis of the above technical solutions, the apparatus further includes: the target translation language type determining module is used for determining the target translation language type and comprises at least one of the following items: acquiring a language type preset on a target client as a target translation language type; and acquiring a login address of the target client, and determining a target language translation type corresponding to the geographical position of the target client based on the login address.
On the basis of the above technical solutions, the apparatus further includes: the key vocabulary determining module is used for determining at least one key vocabulary to be processed corresponding to the multimedia data stream and the original language type of the key vocabulary to be processed by extracting the content of the multimedia data stream, wherein the key vocabulary to be processed is a core vocabulary extracted based on the multimedia data stream; the to-be-processed key vocabulary translation module is used for converting the at least one to-be-processed key vocabulary from the original language type to the target translation language type to obtain at least one to-be-triggered key vocabulary and displaying the to-be-triggered key vocabulary on the target page; the target key vocabulary is one of the at least one key vocabulary to be triggered.
On the basis of the above technical solutions, after the to-be-processed key vocabulary translation module is used for converting the at least one to-be-processed key vocabulary from the original language type to the target translation language type to obtain at least one to-be-triggered key vocabulary, the to-be-processed key vocabulary translation module is further used for: recording the mapping relation between the at least one key word to be triggered and the corresponding original language type, determining the target key word when the key word to be triggered is detected, and determining the target original language type corresponding to the target key word based on the mapping relation.
On the basis of the technical solutions, the target original language type determining module is further configured to determine a target key word when the key word to be triggered displayed on the trigger target page is detected; determining the original language type of the target key vocabulary according to the mapping relation and the target key vocabulary; and generating the target original language type based on the original language type.
On the basis of the above technical solutions, the original text determining module is further configured to determine an original language type corresponding to each piece of translation data in the subtitle information, and obtain an original text that is the same as the target original language type and corresponds to the subtitle information.
On the basis of the above technical solutions, the to-be-matched content determining module, before being configured to determine the to-be-matched content matched with the target key vocabulary from the original text, is further configured to:
and converting the target key vocabulary from the target translation language type into a vocabulary corresponding to the target original language type to obtain a key vocabulary to be matched, and determining the content to be matched from the original text content according to the key vocabulary to be matched.
On the basis of the technical solutions, the to-be-matched key vocabulary includes at least one to-be-searched associated content corresponding to the target key vocabulary in the target original language type, and the to-be-matched content determination module is further configured to search for content identical to the at least one to-be-searched associated content of the to-be-matched key vocabulary from the original text content, and use the content as the to-be-matched content.
On the basis of the foregoing technical solutions, the target translation language type determining module, after being configured to convert the at least one to-be-processed key word from the original language type to the target translation language type to obtain at least one to-be-triggered key word, is further configured to: establishing a vocabulary corresponding relation among the at least one key vocabulary to be processed, the key vocabulary to be triggered and the original language type of the key vocabulary to be processed, determining the key vocabulary to be processed corresponding to the key vocabulary to be triggered based on the vocabulary corresponding relation when the key vocabulary to be triggered is detected, and determining the content to be matched from the original text based on the key vocabulary to be processed.
On the basis of the technical solutions, a target content determining module is used for determining at least one target translation data of the content to be matched under the target translation language type; and determining target content corresponding to the at least one target translation data from the subtitle information.
On the basis of the above technical solutions, the target content determining module is configured to determine, from the subtitle information, the target content corresponding to the content to be matched according to the timestamp of the content to be matched and at least one target translation data.
On the basis of the above technical solutions, the target content determining module is configured to highlight the target content in the subtitle information.
On the basis of the above technical solutions, the apparatus further includes: and the synchronous linkage module is used for establishing a timestamp synchronous relation between the subtitle information and the multimedia data stream so as to display the corresponding multimedia data stream based on the timestamp synchronous rolling relation when the triggering target content is detected.
On the basis of the above technical solutions, the apparatus further includes: the synchronous display module is used for determining a current timestamp corresponding to the target content when the triggering target content is detected; and jumping the multimedia data stream to an audio/video frame corresponding to the current timestamp according to a preset timestamp synchronization relation and the current timestamp.
On the basis of the above technical solutions, the apparatus further includes: and the position marking module is used for determining the time stamp of the multimedia data stream corresponding to each target content in a time axis and marking the position corresponding to the time stamp on the time axis.
On the basis of the above technical solutions, the apparatus further includes: a tag highlighting module for determining a target timestamp corresponding to the target content when a triggering target content is detected; and marking the target corresponding to the target timestamp on the time axis for distinguishing and displaying.
The matching device of the key words provided by the embodiment of the disclosure can execute the matching method of the key words provided by any embodiment of the disclosure, and has corresponding functional modules and beneficial effects of the execution method.
It should be noted that, the units and modules included in the apparatus are merely divided according to functional logic, but are not limited to the above division as long as the corresponding functions can be implemented; in addition, specific names of the functional units are only used for distinguishing one functional unit from another, and are not used for limiting the protection scope of the embodiments of the present disclosure.
EXAMPLE five
Referring now to fig. 7, a schematic diagram of an electronic device (e.g., the terminal device or the server in fig. 7) 500 suitable for implementing embodiments of the present disclosure is shown. The terminal device in the embodiments of the present disclosure may include, but is not limited to, a mobile terminal such as a mobile phone, a notebook computer, a digital broadcast receiver, a PDA (personal digital assistant), a PAD (tablet computer), a PMP (portable multimedia player), a vehicle terminal (e.g., a car navigation terminal), and the like, and a stationary terminal such as a digital TV, a desktop computer, and the like. The electronic device shown in fig. 7 is only an example, and should not bring any limitation to the functions and the scope of use of the embodiments of the present disclosure.
As shown in fig. 7, electronic device 500 may include a processing means (e.g., central processing unit, graphics processor, etc.) 501 that may perform various appropriate actions and processes in accordance with a program stored in a Read Only Memory (ROM)502 or a program loaded from a storage means 505 into a Random Access Memory (RAM) 503. In the RAM503, various programs and data necessary for the operation of the electronic apparatus 500 are also stored. The processing device 501, the ROM502, and the RAM503 are connected to each other through a bus 504. An input/output (I/O) interface 505 is also connected to bus 504.
Generally, the following devices may be connected to the I/O interface 505: input devices 506 including, for example, a touch screen, touch pad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, etc.; output devices 507 including, for example, a Liquid Crystal Display (LCD), speakers, vibrators, and the like; storage devices 506 including, for example, magnetic tape, hard disk, etc.; and a communication device 509. The communication means 509 may allow the electronic device 500 to communicate with other devices wirelessly or by wire to exchange data. While fig. 7 illustrates an electronic device 500 having various means, it is to be understood that not all illustrated means are required to be implemented or provided. More or fewer devices may alternatively be implemented or provided.
In particular, according to an embodiment of the present disclosure, the processes described above with reference to the flowcharts may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program carried on a non-transitory computer readable medium, the computer program containing program code for performing the method illustrated by the flow chart. In such an embodiment, the computer program may be downloaded and installed from a network via the communication means 509, or installed from the storage means 506, or installed from the ROM 702. The computer program performs the above-described functions defined in the methods of the embodiments of the present disclosure when executed by the processing device 501.
The electronic device provided by the embodiment of the present disclosure and the method for matching the key words provided by the above embodiment belong to the same inventive concept, and technical details that are not described in detail in the embodiment can be referred to the above embodiment, and the embodiment and the above embodiment have the same beneficial effects.
EXAMPLE six
The embodiment of the present disclosure provides a computer storage medium, on which a computer program is stored, which when executed by a processor implements the matching method of the key vocabulary provided by the above-described embodiment.
It should be noted that the computer readable medium in the present disclosure can be a computer readable signal medium or a computer readable storage medium or any combination of the two. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples of the computer readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the present disclosure, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In contrast, in the present disclosure, a computer readable signal medium may comprise a propagated data signal with computer readable program code embodied therein, either in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: electrical wires, optical cables, RF (radio frequency), etc., or any suitable combination of the foregoing.
In some embodiments, the clients, servers may communicate using any currently known or future developed network Protocol, such as HTTP (HyperText Transfer Protocol), and may interconnect with any form or medium of digital data communication (e.g., a communications network). Examples of communication networks include a local area network ("LAN"), a wide area network ("WAN"), the Internet (e.g., the Internet), and peer-to-peer networks (e.g., ad hoc peer-to-peer networks), as well as any currently known or future developed network.
The computer readable medium may be embodied in the electronic device; or may exist separately without being assembled into the electronic device.
The computer readable medium carries one or more programs which, when executed by the electronic device, cause the electronic device to:
responding to a trigger operation of a target key word presented on a target page, and determining a target original language type of the target key word;
determining an original text which is the same as the target original language type in the subtitle information presented on the target page;
determining contents to be matched with the target key words from the original text;
and determining target content corresponding to the content to be matched in the subtitle information, and displaying the target content in the target page in a distinguishing manner.
Computer program code for carrying out operations for the present disclosure may be written in any combination of one or more programming languages, including but not limited to an object oriented programming language such as Java, Smalltalk, C + +, and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any type of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet service provider).
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present disclosure. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The units described in the embodiments of the present disclosure may be implemented by software or hardware. Where the name of a unit/module does not in some cases constitute a limitation on the unit itself, for example, the target original language type determination module may also be described as an "original language type determination module".
The functions described herein above may be performed, at least in part, by one or more hardware logic components. For example, without limitation, exemplary types of hardware logic components that may be used include: field Programmable Gate Arrays (FPGAs), Application Specific Integrated Circuits (ASICs), Application Specific Standard Products (ASSPs), systems on a chip (SOCs), Complex Programmable Logic Devices (CPLDs), and the like.
In the context of this disclosure, a machine-readable medium may be a tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. The machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. A machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a machine-readable storage medium would include an electrical connection based on one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
According to one or more embodiments of the present disclosure, [ example one ] there is provided a key word matching method, the method including:
responding to a trigger operation of a target key word presented on a target page, and determining a target original language type of the target key word;
determining an original text which is the same as the target original language type in the subtitle information presented on the target page;
determining contents to be matched with the target key words from the original text;
and determining target content corresponding to the content to be matched in the subtitle information, and displaying the target content in the target page in a distinguishing manner.
According to one or more embodiments of the present disclosure, [ example two ] there is provided a matching method of a key word, further comprising:
optionally, based on the multimedia data stream, determining the voice information and recording an original language type corresponding to the voice information;
and generating caption information which is displayed on a target page and corresponds to the target translation language type according to the voice information, the original language type corresponding to the voice information and the target translation language type.
According to one or more embodiments of the present disclosure, [ example three ] there is provided a matching method of a key word, further comprising:
optionally, determining the language type of the target translation includes at least one of the following:
acquiring a language type preset on a target client as a target translation language type;
and acquiring a login address of the target client, and determining a target language translation type corresponding to the geographical position of the target client based on the login address.
According to one or more embodiments of the present disclosure, [ example four ] there is provided a matching method of a key word, further comprising:
optionally, content extraction processing is performed on the multimedia data stream, and at least one to-be-processed key vocabulary corresponding to the multimedia data stream and an original language type of the to-be-processed key vocabulary are determined, where the to-be-processed key vocabulary is a core vocabulary extracted based on the multimedia data stream;
converting the at least one key vocabulary to be processed from the original language type to the target translation language type to obtain at least one key vocabulary to be triggered, and displaying the key vocabulary to be triggered on the target page;
the target key vocabulary is one of the at least one key vocabulary to be triggered.
According to one or more embodiments of the present disclosure, [ example five ] there is provided a matching method of a key word, further comprising:
optionally, after the at least one to-be-triggered key word is obtained after the at least one to-be-processed key word is converted from the original language type to the target translation language type, the method further includes:
recording the mapping relation between the at least one key word to be triggered and the corresponding original language type, determining the target key word when the key word to be triggered is detected, and determining the target original language type corresponding to the target key word based on the mapping relation.
According to one or more embodiments of the present disclosure, [ example six ] there is provided a matching method of a key word, further comprising:
optionally, the determining an original text of the subtitle information presented on the target page, which is the same as the target original language type, includes:
and determining an original language type corresponding to each piece of translation data in the subtitle information, and acquiring an original text which is the same as the target original language type and corresponds to the subtitle information.
According to one or more embodiments of the present disclosure, [ example seven ] there is provided a matching method of a key word, further comprising:
optionally, before determining the content to be matched that matches the target key vocabulary from the original text, the method further includes:
and converting the target key vocabulary from the target translation language type into a vocabulary corresponding to the target original language type to obtain a key vocabulary to be matched, and determining the content to be matched from the original text content according to the key vocabulary to be matched.
According to one or more embodiments of the present disclosure, [ example eight ] there is provided a matching method of a key word, further comprising:
optionally, the determining, from the original text content, the content to be matched that is matched with the target key vocabulary and includes at least one type of associated content to be searched corresponding to the target key vocabulary in the target original language type, where the determining includes:
and searching the content which is the same as at least one associated content to be searched of the key words to be matched from the original text content, and taking the content as the content to be matched.
According to one or more embodiments of the present disclosure, [ example nine ] there is provided a matching method of a key word, further comprising:
optionally, after the converting the at least one to-be-processed key word from the original language type to the target translation language type to obtain at least one to-be-triggered key word, the method further includes:
establishing a vocabulary corresponding relation among the at least one key vocabulary to be processed, the key vocabulary to be triggered and the original language type of the key vocabulary to be processed, determining the key vocabulary to be processed corresponding to the key vocabulary to be triggered based on the vocabulary corresponding relation when the key vocabulary to be triggered is detected, and determining the content to be matched from the original text based on the key vocabulary to be processed.
According to one or more embodiments of the present disclosure, [ example ten ] there is provided a matching method of a key word, further comprising:
optionally, the determining the target content corresponding to the content to be matched in the subtitle information includes:
determining at least one target translation data of the content to be matched under the target translation language type;
and determining target content corresponding to the at least one target translation data from the subtitle information.
According to one or more embodiments of the present disclosure, [ example eleven ] there is provided a matching method of a key word, further comprising:
optionally, the determining, from the subtitle information, target content corresponding to the at least one target translation data includes:
and determining target content corresponding to the content to be matched from the subtitle information according to the timestamp of the content to be matched and at least one target translation data.
According to one or more embodiments of the present disclosure, [ example twelve ] there is provided a matching method of a key word, further comprising:
optionally, the differentially displaying the target content in the subtitle information includes:
and highlighting the target content in the subtitle information.
According to one or more embodiments of the present disclosure, [ example thirteen ] there is provided a matching method of a key word, further comprising:
optionally, a timestamp synchronization relationship between the subtitle information and the multimedia data stream is established, so that when the trigger target content is detected, the corresponding multimedia data stream is displayed based on the timestamp synchronization rolling relationship.
According to one or more embodiments of the present disclosure, [ example fourteen ] there is provided a key word matching method, further comprising:
optionally, when the trigger target content is detected, determining a current timestamp corresponding to the target content;
and jumping the multimedia data stream to an audio/video frame corresponding to the current timestamp according to a preset timestamp synchronization relation and the current timestamp.
According to one or more embodiments of the present disclosure, [ example fifteen ] there is provided a key word matching method, further comprising:
optionally, a timestamp of the multimedia data stream corresponding to each target content in a time axis is determined, and a position on the time axis corresponding to the timestamp is marked.
According to one or more embodiments of the present disclosure, [ example sixteen ] there is provided a matching method of a key word, further comprising:
optionally, when the trigger target content is detected, determining a target timestamp corresponding to the target content;
and marking the target corresponding to the target timestamp on the time axis for distinguishing and displaying.
According to one or more embodiments of the present disclosure, [ example seventeen ] there is provided an apparatus for matching a key word, the apparatus comprising:
the target original language type determining module is used for responding to triggering operation of target key words presented on a target page and determining the target original language type of the target key words;
an original text determining module, configured to determine an original text that is the same as the target original language type in the subtitle information presented on the target page;
the content to be matched determining module is used for determining the content to be matched with the target key vocabulary from the original text;
and the target content determining module is used for determining the target content corresponding to the content to be matched in the subtitle information and displaying the target content in the target page in a distinguishing manner.
The foregoing description is only exemplary of the preferred embodiments of the disclosure and is illustrative of the principles of the technology employed. It will be appreciated by those skilled in the art that the scope of the disclosure herein is not limited to the particular combination of features described above, but also encompasses other embodiments in which any combination of the features described above or their equivalents does not depart from the spirit of the disclosure. For example, the above features and (but not limited to) the features disclosed in this disclosure having similar functions are replaced with each other to form the technical solution.
Further, while operations are depicted in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order. Under certain circumstances, multitasking and parallel processing may be advantageous. Likewise, while several specific implementation details are included in the above discussion, these should not be construed as limitations on the scope of the disclosure. Certain features that are described in the context of separate embodiments can also be implemented in combination in a single embodiment. Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination.
Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example forms of implementing the claims.

Claims (19)

1. A matching method of key words is characterized by comprising the following steps:
responding to a trigger operation of a target key word presented on a target page, and determining a target original language type of the target key word, wherein the target original language type is an initial language type corresponding to the target key word;
determining an original text with an original language type the same as the target original language type in subtitle information presented on the target page, wherein the original language type of the subtitle information is the original language type of voice information corresponding to the subtitle information, and the original text is determined based on the target original language type and the original language type of the voice information;
determining contents to be matched with the target key words from the original text;
and determining target content corresponding to the content to be matched in the subtitle information, and displaying the target content in the target page in a distinguishing manner.
2. The method of claim 1, further comprising:
determining voice information and an original language type corresponding to the voice information based on the multimedia data stream;
and generating caption information which is displayed on a target page and corresponds to the target translation language type according to the voice information, the original language type corresponding to the voice information and the target translation language type.
3. The method of claim 2, wherein determining the target translation language type comprises at least one of:
acquiring a language type preset on a target client as a target translation language type;
and acquiring a login address of the target client, and determining a target language translation type corresponding to the geographical position of the target client based on the login address.
4. The method of claim 2, further comprising:
determining at least one key vocabulary to be processed corresponding to the multimedia data stream and an original language type of the key vocabulary to be processed by extracting the content of the multimedia data stream, wherein the key vocabulary to be processed is a core vocabulary extracted based on the multimedia data stream;
converting the at least one key vocabulary to be processed from the original language type into the target translation language type to obtain at least one key vocabulary to be triggered, and displaying the key vocabulary to be triggered on the target page;
the target key vocabulary is one of the at least one key vocabulary to be triggered.
5. The method of claim 4, further comprising:
recording the mapping relation between the at least one key word to be triggered and the corresponding original language type, determining the target key word when the key word to be triggered is detected, and determining the target original language type corresponding to the target key word based on the mapping relation.
6. The method of claim 2, wherein the determining the original text of the subtitle information presented on the target page that is of the same type as the target original language comprises:
and determining an original language type corresponding to each piece of translation data in the subtitle information, and acquiring an original text which is the same as the target original language type and corresponds to the subtitle information.
7. The method according to claim 1, wherein before said determining the content to be matched that matches the target key vocabulary from the original text, further comprising:
and converting the target key vocabulary from the target translation language type into a vocabulary corresponding to the target original language type to obtain a key vocabulary to be matched, and determining the content to be matched from the original text content according to the key vocabulary to be matched.
8. The method according to claim 7, wherein the to-be-matched keyword words include at least one to-be-searched related content corresponding to the target keyword words in the target original language type, and the determining the to-be-matched content matching the target keyword words from the original text content includes:
and searching the content which is the same as the at least one associated content to be searched from the original text content, and taking the content as the content to be matched.
9. The method according to claim 4, wherein after said converting said at least one key word to be processed from said original language type to said target translation language type to obtain at least one key word to be triggered, further comprising:
establishing a vocabulary corresponding relation among the at least one key vocabulary to be processed, the key vocabulary to be triggered and the original language type of the key vocabulary to be processed, determining the key vocabulary to be processed corresponding to the key vocabulary to be triggered based on the vocabulary corresponding relation when the key vocabulary to be triggered is detected, and determining the content to be matched from the original text based on the key vocabulary to be processed.
10. The method according to claim 1, wherein the determining the target content corresponding to the content to be matched in the subtitle information comprises:
determining at least one target translation data of the content to be matched under the target translation language type;
and determining target content corresponding to the at least one target translation data from the subtitle information.
11. The method of claim 9, wherein the determining the target content corresponding to the at least one target translation data from the subtitle information comprises:
and determining target content corresponding to the content to be matched from the subtitle information according to the timestamp of the content to be matched and at least one target translation data.
12. The method of claim 1, wherein said distinguishing the target content from the subtitle information comprises:
and highlighting the target content in the subtitle information.
13. The method of claim 2, further comprising:
and establishing a timestamp synchronization relationship between the subtitle information and the multimedia data stream, and displaying the corresponding multimedia data stream based on the timestamp synchronization rolling relationship when the triggering target content is detected.
14. The method of claim 13, further comprising:
when triggering target content is detected, determining a current timestamp corresponding to the target content;
and jumping the multimedia data stream to a playing position corresponding to the current timestamp according to a preset timestamp synchronization relation and the current timestamp.
15. The method of claim 1, further comprising:
the time stamp of the multimedia data stream corresponding to each target content in the time axis is determined, and the position corresponding to the time stamp on the time axis is marked.
16. The method of claim 15, further comprising:
when trigger target content is detected, determining a target timestamp corresponding to the target content;
and marking the target corresponding to the target timestamp on the time axis for distinguishing and displaying.
17. An apparatus for matching a key word, comprising:
the target original language type determining module is used for responding to triggering operation of a target key vocabulary presented on a target page and determining a target original language type of the target key vocabulary, wherein the target original language type is an initial language type corresponding to the target key vocabulary;
an original text determining module, configured to determine, in the subtitle information presented on the target page, an original text having an original language type that is the same as the target original language type, where the original language type of the subtitle information is an original language type of the voice information corresponding to the subtitle information, and the original text is determined based on the target original language type and the original language type of the voice information;
the content to be matched determining module is used for determining the content to be matched with the target key vocabulary from the original text;
and the target content determining module is used for determining the target content corresponding to the content to be matched in the subtitle information and displaying the target content in the target page in a distinguishing manner.
18. An electronic device, characterized in that the electronic device comprises:
one or more processors;
a storage device for storing one or more programs,
when executed by the one or more processors, cause the one or more processors to implement a method of matching key words as claimed in any one of claims 1-16.
19. A storage medium containing computer executable instructions for performing a method of matching key words as claimed in any one of claims 1 to 16 when executed by a computer processor.
CN202011052749.2A 2020-09-29 2020-09-29 Key vocabulary matching method and device, electronic equipment and storage medium Active CN112163433B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011052749.2A CN112163433B (en) 2020-09-29 2020-09-29 Key vocabulary matching method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011052749.2A CN112163433B (en) 2020-09-29 2020-09-29 Key vocabulary matching method and device, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN112163433A CN112163433A (en) 2021-01-01
CN112163433B true CN112163433B (en) 2022-04-05

Family

ID=73861508

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011052749.2A Active CN112163433B (en) 2020-09-29 2020-09-29 Key vocabulary matching method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN112163433B (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104681023A (en) * 2015-02-15 2015-06-03 联想(北京)有限公司 Information processing method and electronic equipment
WO2016091011A1 (en) * 2014-12-12 2016-06-16 深圳Tcl数字技术有限公司 Subtitle switching method and device
CN107071554A (en) * 2017-01-16 2017-08-18 腾讯科技(深圳)有限公司 Method for recognizing semantics and device

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103778131B (en) * 2012-10-18 2017-02-22 腾讯科技(深圳)有限公司 Caption query method and device, video player and caption query server
CN103838751A (en) * 2012-11-23 2014-06-04 鸿富锦精密工业(深圳)有限公司 Video content searching system and method
FR3051092A1 (en) * 2016-05-03 2017-11-10 Orange METHOD AND DEVICE FOR SYNCHRONIZING SUBTITLES
CN108255940A (en) * 2017-12-08 2018-07-06 北京搜狗科技发展有限公司 A kind of cross-language search method and apparatus, a kind of device for cross-language search
CN108600773B (en) * 2018-04-25 2021-08-10 腾讯科技(深圳)有限公司 Subtitle data pushing method, subtitle display method, device, equipment and medium

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016091011A1 (en) * 2014-12-12 2016-06-16 深圳Tcl数字技术有限公司 Subtitle switching method and device
CN104681023A (en) * 2015-02-15 2015-06-03 联想(北京)有限公司 Information processing method and electronic equipment
CN107071554A (en) * 2017-01-16 2017-08-18 腾讯科技(深圳)有限公司 Method for recognizing semantics and device

Also Published As

Publication number Publication date
CN112163433A (en) 2021-01-01

Similar Documents

Publication Publication Date Title
US11917344B2 (en) Interactive information processing method, device and medium
CN111445902B (en) Data collection method, device, storage medium and electronic equipment
CN112163102B (en) Search content matching method and device, electronic equipment and storage medium
WO2023125374A1 (en) Image processing method and apparatus, electronic device, and storage medium
CN111753558B (en) Video translation method and device, storage medium and electronic equipment
CN112084920B (en) Method, device, electronic equipment and medium for extracting hotwords
CN113889113A (en) Sentence dividing method and device, storage medium and electronic equipment
CN113724709A (en) Text content matching method and device, electronic equipment and storage medium
CN112995749A (en) Method, device and equipment for processing video subtitles and storage medium
CN113886612A (en) Multimedia browsing method, device, equipment and medium
CN112380365A (en) Multimedia subtitle interaction method, device, equipment and medium
CN112069950A (en) Method, system, electronic device and medium for extracting hotwords
CN114445754A (en) Video processing method and device, readable medium and electronic equipment
CN113703579A (en) Data processing method and device, electronic equipment and storage medium
CN113014853B (en) Interactive information processing method and device, electronic equipment and storage medium
CN113014854B (en) Method, device, equipment and medium for generating interactive record
CN112163433B (en) Key vocabulary matching method and device, electronic equipment and storage medium
CN112163104B (en) Method, device, electronic equipment and storage medium for searching target content
CN112163103A (en) Method, device, electronic equipment and storage medium for searching target content
CN113552984A (en) Text extraction method, device, equipment and medium
CN113132789B (en) Multimedia interaction method, device, equipment and medium
US20230140442A1 (en) Method for searching target content, and electronic device and storage medium
CN114697761B (en) Processing method, processing device, terminal equipment and medium
US11792494B1 (en) Processing method and apparatus, electronic device and medium
CN115037982A (en) Video playing method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant