CN106446235B - Video searching method and device - Google Patents

Video searching method and device Download PDF

Info

Publication number
CN106446235B
CN106446235B CN201610884802.2A CN201610884802A CN106446235B CN 106446235 B CN106446235 B CN 106446235B CN 201610884802 A CN201610884802 A CN 201610884802A CN 106446235 B CN106446235 B CN 106446235B
Authority
CN
China
Prior art keywords
search
formula
keywords
retrieval
query
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201610884802.2A
Other languages
Chinese (zh)
Other versions
CN106446235A (en
Inventor
曾煜钊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
TCL Technology Group Co Ltd
Original Assignee
TCL Technology Group Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by TCL Technology Group Co Ltd filed Critical TCL Technology Group Co Ltd
Priority to CN201610884802.2A priority Critical patent/CN106446235B/en
Publication of CN106446235A publication Critical patent/CN106446235A/en
Application granted granted Critical
Publication of CN106446235B publication Critical patent/CN106446235B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/73Querying
    • G06F16/732Query formulation
    • G06F16/7343Query language or query format
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/7867Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title and artist information, manually generated time, location and usage information, user ratings

Abstract

The invention provides a video searching method and a video searching device, wherein the method comprises the following steps: receiving a retrieval statement; performing word segmentation on the retrieval sentence to obtain a plurality of keywords; generating a search formula according to the plurality of keywords; generating a retrieval result according to the retrieval formula; and searching video data corresponding to the retrieval result through a preset query language. In the invention, the keywords are extracted by word segmentation, the retrieval formula is set according to a plurality of retrieval words, and the retrieval result is obtained.

Description

Video searching method and device
Technical Field
The invention relates to the field of internet, in particular to a video searching method and device.
Background
With the popularization of the smart television, video on demand has become an indispensable part of the smart television. Currently, some video content providers are already able to provide a large, constantly updated video asset.
The video searching method of the smart television in the market is simple, and most of the smart televisions are searched according to a single condition, such as language or actors. For the composite search condition, a relatively accurate search result cannot be obtained at present.
Aiming at the problem that the composite condition is not accurately searched in the prior art, no ideal solution is provided in the industry at present.
Disclosure of Invention
The invention aims to provide a video searching method and device, and aims to solve the problem that composite condition searching is inaccurate in the prior art.
The invention provides a video searching method, which comprises the following steps:
receiving a retrieval statement;
segmenting a search sentence to obtain a plurality of keywords;
generating a current searching formula according to a plurality of keywords;
generating a retrieval result according to the retrieval formula;
and searching video data corresponding to the retrieval result through a preset query language.
The invention also provides a video searching device, which comprises:
a receiving unit configured to receive a search sentence;
the system comprises an acquisition unit, a search unit and a search unit, wherein the acquisition unit is used for segmenting a search sentence to acquire a plurality of keywords;
a first generating unit, configured to generate the current search expression according to the plurality of keywords;
the second generating unit is used for generating a retrieval result according to the retrieval formula;
and the searching unit is used for searching the video data corresponding to the retrieval result through a preset query language.
In the invention, the keywords are extracted by word segmentation, the retrieval formula is set according to a plurality of retrieval words, and the retrieval result is obtained.
Drawings
Fig. 1 is a flowchart of a video searching method according to an embodiment of the present invention;
FIG. 2 is a flow chart of a word segmentation method according to an embodiment of the present invention;
FIG. 3 is a flow chart of generating search results provided by an embodiment of the present invention;
FIG. 4 is a diagram illustrating the generation of search results according to an embodiment of the present invention;
fig. 5 is a block diagram of a video search apparatus according to an embodiment of the present invention.
Detailed Description
In order to make the technical problems, technical solutions and advantageous effects to be solved by the present invention more clearly understood, the present invention is further described in detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
The embodiment of the invention provides a video searching method. Fig. 1 is a flowchart of a video search method according to an embodiment of the present invention. As shown in fig. 1, the method includes steps S102 to S110.
Step S102, receiving a search statement.
The received retrieval statement comes from the input of the smart television user, and the user can input the statement according to the requirement when the user requests the video, for example, the retrieval statement is 'starry comedies'.
Step S104, performing word segmentation on the retrieval sentence to obtain a plurality of keywords.
Based on the current technology, it is difficult to analyze the entire sentence, so it is necessary to extract keywords first and then learn the intention of the user through the keywords. In this embodiment, the keywords are obtained by word segmentation.
Fig. 2 is a flowchart of a word segmentation method according to an embodiment of the present invention, and as shown in fig. 2, the method includes the following steps:
step S202, reading the attribute information of the video data in the video library.
The attribute information of each video in the video library generally includes a movie, a television play, an animation, a synthesis, and the like, and attribute information of many other dimensions, such as a name, an actor, a director, a showing time, a genre, a region, a language, and the like, and even a part of the video may include attribute information of a tag, a definition, a score, a click rate, and the like.
Step S204, a plurality of keywords are obtained from the attribute information, and a keyword word bank is generated.
The read attribute information is a huge lexicon. Including a large number of words, for example, the movie names include "people jiong on the way", "jiong on the way", and "jiong on harong" etc., the actors include "towering", "wangbaoqiang", and "yellow bohai", etc., the showing time includes "2013", "2014", and "2015", etc., and the labels include "comedy drama film", "horror film", and "literature film", etc. These words are grouped together to form a keyword lexicon.
Further, in this step, a plurality of keywords are extracted from the attribute information, and the type of each keyword includes a video title, an actor, a director, a showing time, a field, a region, and a language.
That is, when the keyword lexicon extracts the "2013 year" from the attribute information, the type "mapping time" is extracted at the same time, and the subsequent steps generate a search result by combining the keywords and the types thereof.
Step S206, matching the search sentence with the keyword word stock, and determining the words in the search sentence which can be matched with the keyword word stock as the keywords.
For example, the above example "comedy on a monday drive", compares the sentence to the keyword lexicon, and "monday drive" and "comedy" are matchable, and thus determine the two words as keywords.
And step S106, generating the searching formula according to the keywords.
In this embodiment, the extracted keywords are used as analysis results and are expressed by a lightweight data interchange format (JSON). JSON is convenient to parse, is a format which is used more at present, and similar effects can be achieved by using other similar formats in the embodiment.
In this embodiment, "starry comedies" can be expressed as:
[
{
value means planet speeding,
"key": "actor"
},
[
{
"value" to "comedy",
"key": "type"
},
{
"value" to "comedy",
"key": "label"
}
]
]
the relation between the elements of the JSON array of the first layer is an AND relation, the relation between the elements of the JSON array of the second layer is an OR relation, and the AND relation is shown as the above relation between the starry (actors) and the comedy (types or labels), which indicates that the requirements are met simultaneously; the relationship of yes between genre "comedy" and label "comedy" means that only one of them is satisfied.
And step S108, generating a retrieval result according to the retrieval formula.
By the retrieval formula of the JSON format, retrieval results meeting the conditions can be obtained.
In order to realize context correlation of retrieval and facilitate more accurate secondary retrieval, after the user carries out on-demand search each time, the input of the user, the result of semantic recognition and the video search result are stored in the cache of the server.
Then, the result of semantic analysis of the keyword at this time may be combined with the cache of the previous operation.
Fig. 3 is a flowchart of generating a search result according to an embodiment of the present invention, and as shown in fig. 2, the method includes the following steps:
s302, obtaining a last retrieval formula, wherein the last received retrieval statement comprises a plurality of keywords, and generating the last retrieval formula according to the keywords.
And S304, generating a retrieval result according to the last retrieval formula and the current retrieval formula.
The rules for generating the retrieval result include:
and if the current search formula only comprises the keywords of which the types are fields, and a plurality of fields are generated through the previous search formula and comprise the search results of the field to which the current search result belongs, screening the search results of the fields from the search results corresponding to the previous search formula to serve as the search results generated by the current search formula. The field is the general classification of video content, such as science fiction films, documentaries, and so forth. If the science fiction film and documentary are searched in the last search and only documentaries are searched in the current time, the documentary part in the last search result can be used as the search result of the current time.
If the current search formula only comprises the keywords of which the types are fields, and the search result of the fields is not generated through the previous search formula or the search result only comprising the fields is generated, the previous search formula is ignored, and the search result is directly generated according to the current search formula. When the last retrieval does not contain the field of the current time, the last retrieval result has no reference significance for the current retrieval. When the last retrieval and the current retrieval are in the same field, for example, the gunfight film of Liudebua is retrieved last time, and the gunfight film is retrieved this time, the last retrieval result can be discarded, and the gunfight film can be directly retrieved.
If the current search formula comprises the keywords of types except the mapping time, the domain and the region, the last search formula is ignored, and the search result is directly generated according to the current search formula. The mapping time, the domain and the region belong to a tightening dimension, that is, the number of results corresponding to the attribute information is easy to enumerate, for example, the mapping time is generally 1980-2016, and a larger video library may be traced back to 1960 and the like; the fields can be classified into military fields, sports fields, and the like, and other types of keywords are not easy to list, for example, actors, and since a video library has a large amount of actor attribute information and is difficult to list, a search for such keywords may be performed by directly generating a search result from the current search formula.
If the current search formula only includes the keyword with the type of at least one of the showing time, the field and the area, and the last search formula does not include the keyword with the type of the showing time, the field or the area, merging the current search formula and the last search formula to obtain the search result of the current search formula.
If the current search formula only comprises the keywords with the type of at least one of the showing time, the field and the area, and the last search formula comprises the keywords with the type of at least one of the showing time, the field or the area, covering the same type of search words in the current search formula and the last search formula, and reserving other keywords to obtain the search result of the current search formula.
Fig. 4 is a schematic diagram of generating a search result according to an embodiment of the present invention, and as shown in fig. 4, the method includes:
step S401, semantic analysis input.
I.e., a JSON format search formula, by which a correlation search can be performed.
Step S402, judging whether to enter the search field, if not, executing step S403; if so, step S405 is performed.
Step S403, judging whether only attribute information of the dimension to be tightened is included, if so, executing step S408; if not, step S406 is performed.
Step S404, obtaining the last retrieval result from the server cache.
Step S405, judging whether a plurality of fields including the fields related in the step S403 exist, if so, executing step S407; if not, step S406 is performed.
Step S406, re-searching according to the searching formula.
In step S407, the individual domains in S403 are screened out.
Step S408, judging whether the dimension is the same as that in the step S403, if so, executing the step S409; if not, step S410 is performed.
In step S409, a merge search is performed.
Step S410, overlay search is performed.
And step S411, outputting the result.
The result output here is also data in JSON format.
Step S110, searching for video data corresponding to the retrieval result through a preset query language.
In this embodiment, the search result is data in JSON format, and only some data related to the video are listed in the data, but the data are viewed by machines and developers, and in order to obtain content visible to users, related information needs to be called in a video library through a query language.
In this embodiment, the SQL query statement and the MySQL database are taken as examples to describe a process of searching for video data through a query language.
First, the basic unit of the query condition needs to be converted into the basic SQL statement. "basic unit" here refers to the most basic element in the JSON format data, which is in the form of a JSON object and no longer contains a JSON object or JSON array, and for example, the above example of "sunstar cur comedy" contains three basic units of "{" value ":": sunstar cury "," key ": action" } "," { "value": comedy "," key ": type" } "and" { "value": comedy "," key ": label" } ". The basic unit is converted into the SQL statement, and different SQL query statements are used under the condition that the query conditions of the basic unit belong to different dimensions. These include:
when the dimension of the query condition of the basic unit is a name, the fuzzy query directly using SQL, for example, the MySQL query statement corresponding to the conditions of "{" value ": Shousck's redemption", "key": title "}" is:
SELECT FROM my _ table WHERE title LIKE '% Shouscheck's rescue% '
The fuzzy query is used for names because common user habits are that a part of the names of the films and the videos are used for searching, but not necessarily the full names, and the scheme is more efficient particularly for video resources with longer names.
When the query condition of the basic unit belongs to the dimension of actor, director (the general program corresponds to the host and guest) or genre, it is considered that such search is generally more accurate, and the actor, director and genre may have a plurality of values in parallel (divided by punctuation marks), for example, the actor in "journey of western language" is "mondsand wumengda, juliens …". This case can be queried with regular expressions. For example, the MySQL query statements corresponding to the conditions of "{" value ": Zhou Xinghong", "key": operator "}" are:
SELECT FROM my _ table WHERE actor REGEX '[: ] ]Zhou xing [: ] ]'
Such a process of searching for "monday" will not return the result of "monday coming" erroneously, and can avoid the problem of search confusion in the case where the names of actors are included in each other.
Such a search is also relatively accurate when the dimension to which the query condition for the primitive belongs is region or resolution, and typically has only one value (e.g., region is "hong kong"), which can be directly queried using an accurate query statement. For example, the MySQL query statement corresponding to the conditions of "{" value ": hong Kong" key ": area" } "is:
SELECT FROM my _ table WHERE area = 'hong Kong'
When the dimension to which the query condition of the basic unit belongs is the mapping time (year), the query can be performed by using the time interval query condition of MySQL. For example, the conditions of "value": 2016": year" } "correspond to MySQL query statements of:
SELECT * FROM my_table WHERE publishtime BETWEEN '2016-01-01' AND '2016-12-31'
the query condition has a value of year because a general user does not use a specific time for a query, and the year is a keyword in the keyword lexicon. And the show time information stored in the video library may be the exact show time.
And converting the basic units of the query condition into basic SQL sentences, and splicing the basic SQL sentences according to the relationship among the basic units. If the relation is AND, using AND key words; if a yes OR relationship, the OR key is used. The complete SQL query statement for "starry comedies" is thus:
SELECT FROM my _ table WHERE (actor REGEX XP ' [: ] ] circumstar [: ] ' AND (type REGEX XP ' [: ] ] comedy [: ]: '. OR label REGEX ' [: ] ] comedy [: ])
After the SQL query statement is used for querying the movie and television resource database, the data records meeting the search conditions of the user are obtained. These data records generally contain poster pictures (link addresses), playback information (playback link addresses or playback parameters), and the like, in addition to the above-mentioned attribute values of the plurality of dimensions of the title, director, actor, and the like. And the query result display can display the movie name and the poster picture, and the user invokes the relevant playing information to execute the playing operation after selecting a certain movie work.
In the embodiment of the invention, the keywords are extracted by word segmentation, the search formula is set according to the plurality of search words, and the search result is obtained.
The embodiment of the invention also provides a video searching device. Fig. 5 is a block diagram of a video search apparatus according to an embodiment of the present invention, and as shown in fig. 5, the apparatus includes a receiving unit 510, an obtaining unit 520, a first generating unit 530, a second generating unit 540, and a searching unit 550.
The receiving unit 510 is configured to receive a retrieval statement.
The obtaining unit 520 is configured to perform word segmentation on the search sentence to obtain a plurality of keywords.
The first generating unit 530 is configured to generate the current search expression according to a plurality of keywords.
The second generating unit 540 is configured to generate a search result according to the current search expression.
The searching unit 550 is configured to search for video data corresponding to the search result through a preset query language.
Preferably, the acquisition unit includes: the reading module is used for reading the attribute information of the video data in the video library; the first generation module is used for acquiring a plurality of keywords from the attribute information and generating a keyword lexicon; and the determining module is used for matching the search sentences with the keyword word stock and determining the words which can be matched with the keyword word stock in the search sentences as the keywords.
Preferably, the first generating module comprises: and the extraction sub-module is used for extracting a plurality of keywords from the attribute information and the types of the keywords, wherein the types comprise video names, actors, directors, showing time, fields, regions and languages.
Preferably, the second generating unit includes: the acquisition module is used for acquiring a last retrieval formula, wherein a last received retrieval statement comprises a plurality of keywords, and the last retrieval formula is generated according to the keywords; and the second generation module is used for generating a retrieval result according to the last retrieval formula and the current retrieval formula.
Preferably, the second generating module includes: the first generation submodule is used for screening out the retrieval result of the field from the retrieval results corresponding to the retrieval formula of the last time when the retrieval formula only comprises the keywords of which the types are the fields and the retrieval results of a plurality of fields and the fields are generated by the retrieval formula of the last time and comprise the retrieval results of the fields, and the retrieval results are generated by the retrieval formula of the time; the second generation submodule is used for neglecting the previous retrieval formula and directly generating a retrieval result according to the current retrieval formula when the current retrieval formula only comprises the keywords of which the types are fields and does not generate the retrieval result of the fields through the previous retrieval formula or generates the retrieval result only comprising the fields; the third generation submodule is used for neglecting the last retrieval formula and directly generating a retrieval result according to the retrieval formula when the retrieval formula comprises the keywords of types except the mapping time, the field and the region; a fourth generation submodule, configured to merge the current search formula with the previous search formula when the current search formula only includes the keyword of which the type is at least one of the mapping time, the domain, and the region, and the previous search formula does not include the keyword of which the type is the mapping time, the domain, or the region, so as to obtain a search result of the current search formula; and a fifth generation sub-module, configured to cover search words of the same type in the current search formula and the previous search formula when the current search formula only includes the keyword of the at least one of the mapping time, the domain, and the region, and the previous search formula includes the keyword of the at least one of the mapping time, the domain, and the region, and retain other keywords, so as to obtain a search result of the current search formula.
It will be apparent to those skilled in the art that, for convenience and brevity of description, only the above-mentioned division of each functional unit is illustrated, and in practical applications, the above-mentioned functional allocation may be performed by different functional units according to requirements, that is, the internal structure of the apparatus is divided into different functional units or modules to perform all or part of the above-mentioned functions. Each functional unit in the embodiments may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit, and the integrated unit may be implemented in a form of hardware, or in a form of software functional unit. In addition, specific names of the functional units are only used for distinguishing one functional unit from another, and are not used for limiting the protection scope of the application. For the specific working process of the units in the above-mentioned apparatus, reference may be made to the corresponding process in the foregoing embodiment of the apparatus, which is not described herein again.
Those of ordinary skill in the art will appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
In the embodiments provided in the present invention, it should be understood that the disclosed apparatus may be implemented in other manners. For example, the above-described apparatus embodiments are merely illustrative, and for example, a module or a unit may be divided into only one logical function, and may be implemented in other ways, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, devices or units, and may be in an electrical, mechanical or other form.
Units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
The integrated unit, if implemented in the form of a software functional unit and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the embodiments of the present invention may be essentially or partially contributed by the prior art, or all or part of the technical solution may be embodied in a software product, which is stored in a storage medium and includes several instructions for causing a computer device (which may be a personal computer, a server, or a network device) or a processor (processor) to execute all or part of the steps of the apparatus of the embodiments of the present invention. And the aforementioned storage medium includes: a U-disk, a removable hard disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and other various media capable of storing program codes.
The above examples are only intended to illustrate the technical solution of the present invention, but not to limit it; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; and the modifications or the substitutions do not make the essence of the corresponding technical solutions depart from the spirit and scope of the technical solutions of the embodiments of the present invention.
The present invention is not limited to the above preferred embodiments, and any modifications, equivalent substitutions and improvements made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (6)

1. A method for searching a video, comprising:
receiving a retrieval statement;
performing word segmentation on the retrieval sentence to obtain a plurality of keywords;
generating a search formula according to the plurality of keywords; wherein, the extracted key words are used as analysis results and expressed by JSON format;
generating a retrieval result according to the retrieval formula;
searching video data corresponding to the retrieval result through a preset query language;
the searching for the video data corresponding to the retrieval result through a preset query language comprises:
converting the basic unit of the query condition into a basic SQL statement; when the dimension to which the query condition of the basic unit belongs is a name, directly using SQL fuzzy query, when the dimension to which the query condition of the basic unit belongs is an actor, a director or a type, using a regular expression to perform query, when the dimension to which the query condition of the basic unit belongs is a region or definition, directly using an accurate query statement to perform query, and when the dimension to which the query condition of the basic unit belongs is mapping time, using MySQL time interval query conditions to perform query;
splicing the basic SQL sentences to obtain SQL query sentences according to the relationship among the basic units;
after using SQL query sentence to query movie and television resource database, obtaining data record meeting user search condition;
generating a search result according to the current search formula comprises the following steps:
obtaining a last retrieval formula, wherein a retrieval statement received last time comprises a plurality of keywords, and generating the last retrieval formula according to the keywords;
if the current search formula only comprises the keywords with the types of fields, and a plurality of fields are generated through the previous search formula and comprise the search results of the fields, screening the search results of the fields from the search results corresponding to the previous search formula to serve as the search results generated by the current search formula;
if the current search formula only comprises the keywords of which the types are fields, and the search result of the fields is not generated through the last search formula or the search result only comprising the fields is generated, ignoring the last search formula and directly generating the search result according to the current search formula;
if the current search formula comprises the keywords of types except the mapping time, the domain and the region, ignoring the previous search formula, and directly generating the search result according to the current search formula;
if the current search formula only comprises the keywords with the types of at least one of the showing time, the field and the area, and the keywords with the types of the showing time, the field or the area are not included in the last search formula, merging the current search formula and the last search formula to obtain a search result of the current search formula; or
If the current search formula only comprises the keywords with the types of at least one of the mapping time, the domain and the region, and the last search formula comprises the keywords with the types of at least one of the mapping time, the domain or the region, covering the same types of search words in the current search formula and the last search formula, and reserving other keywords to obtain the search result of the current search formula.
2. The method of claim 1, wherein segmenting the search term to obtain a plurality of keywords comprises:
reading attribute information of video data in a video library;
acquiring a plurality of keywords from the attribute information to generate a keyword lexicon;
and matching the search sentences with the keyword word stock, and determining words which can be matched with the keyword word stock in the search sentences as keywords.
3. The method of claim 2, wherein obtaining a plurality of keywords from the attribute information, and generating a keyword lexicon comprises:
and extracting a plurality of keywords from the attribute information, and the types of the keywords, wherein the types comprise video names, actors, directors, showing time, fields, regions and languages.
4. A video search apparatus, comprising:
a receiving unit configured to receive a search sentence;
the acquisition unit is used for segmenting the search sentences to acquire a plurality of keywords;
a first generating unit, configured to generate the current search expression according to the plurality of keywords; wherein, the extracted key words are used as analysis results and expressed by JSON format;
the second generating unit is used for generating a retrieval result according to the retrieval formula;
the searching unit is used for searching the video data corresponding to the retrieval result through a preset query language, and comprises: converting the basic unit of the query condition into a basic SQL statement; when the dimension to which the query condition of the basic unit belongs is a name, directly using SQL fuzzy query, when the dimension to which the query condition of the basic unit belongs is an actor, a director or a type, using a regular expression to perform query, when the dimension to which the query condition of the basic unit belongs is a region or definition, directly using an accurate query statement to perform query, and when the dimension to which the query condition of the basic unit belongs is mapping time, using MySQL time interval query conditions to perform query;
splicing the basic SQL sentences to obtain SQL query sentences according to the relationship among the basic units;
after using SQL query sentence to query movie and television resource database, obtaining data record meeting user search condition;
the second generation unit includes:
the acquisition module is used for acquiring a last retrieval formula, wherein a last received retrieval statement comprises a plurality of keywords, and the last retrieval formula is generated according to the keywords;
the second generation module is used for generating the retrieval result according to the last retrieval formula and the current retrieval formula;
the second generation module comprises:
the first generation submodule is used for screening out the retrieval result of the field from the retrieval results corresponding to the last retrieval formula to serve as the retrieval result generated by the current retrieval formula when the current retrieval formula only comprises the keyword of which the type is the field and the retrieval results of a plurality of fields and the fields are generated by the last retrieval formula;
a second generation submodule, configured to ignore the previous search expression when the current search expression only includes the keyword whose type is a field and a search result of the field is not generated by the previous search expression or a search result only including the field is generated, and directly generate the search result according to the current search expression;
a third generation submodule, configured to, when the current search expression includes the keyword of a type other than the mapping time, the domain, and the region, ignore the previous search expression, and directly generate the search result according to the current search expression;
a fourth generation sub-module, configured to, when the current search formula only includes the keyword of which the type is at least one of the mapping time, the domain, and the region, and the last search formula does not include the keyword of which the type is the mapping time, the domain, or the region, merge the current search formula and the last search formula to obtain a search result of the current search formula;
and a fifth generation submodule, configured to cover search words of the same type in the current search formula and the previous search formula when the current search formula only includes the keyword of which the type is at least one of the mapping time, the domain, and the region, and the previous search formula includes the keyword of which the type is at least one of the mapping time, the domain, or the region, and retain other keywords, so as to obtain a search result of the current search formula.
5. The apparatus of claim 4, wherein the obtaining unit comprises:
the reading module is used for reading the attribute information of the video data in the video library;
the first generation module is used for acquiring a plurality of keywords from the attribute information and generating a keyword lexicon;
and the determining module is used for matching the search sentences with the keyword word stock and determining the words which can be matched with the keyword word stock in the search sentences as the keywords.
6. The apparatus of claim 5, wherein the first generating module comprises:
and the extraction sub-module is used for extracting a plurality of keywords and types of the keywords from the attribute information, wherein the types comprise video names, actors, directors, showing time, fields, regions and languages.
CN201610884802.2A 2016-10-10 2016-10-10 Video searching method and device Active CN106446235B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610884802.2A CN106446235B (en) 2016-10-10 2016-10-10 Video searching method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610884802.2A CN106446235B (en) 2016-10-10 2016-10-10 Video searching method and device

Publications (2)

Publication Number Publication Date
CN106446235A CN106446235A (en) 2017-02-22
CN106446235B true CN106446235B (en) 2021-04-06

Family

ID=58173063

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610884802.2A Active CN106446235B (en) 2016-10-10 2016-10-10 Video searching method and device

Country Status (1)

Country Link
CN (1) CN106446235B (en)

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106919703A (en) * 2017-03-09 2017-07-04 北京微影时代科技有限公司 Film information searching method and device
CN110019094A (en) * 2017-12-28 2019-07-16 中国移动通信集团广东有限公司 Ticket retrieve method, system, electronic equipment and storage medium
CN110377766A (en) * 2018-04-11 2019-10-25 北京搜狗科技发展有限公司 A kind of data processing method, device and electronic equipment
CN112528073A (en) * 2019-09-03 2021-03-19 北京国双科技有限公司 Video generation method and device
CN111221957A (en) * 2020-01-10 2020-06-02 合肥工业大学 Scientific and technological information automatic processing method and system based on knowledge organization
CN111831750A (en) * 2020-07-09 2020-10-27 恒安嘉新(北京)科技股份公司 Block chain data analysis method and device, computer equipment and storage medium
CN114817265B (en) * 2022-04-28 2023-08-18 深圳市领航财富教育科技有限公司 Financial information acquisition method by utilizing big data server

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101770499A (en) * 2009-01-07 2010-07-07 上海聚力传媒技术有限公司 Information retrieval method in search engine and corresponding search engine
US8625033B1 (en) * 2010-02-01 2014-01-07 Google Inc. Large-scale matching of audio and video
CN104268175A (en) * 2014-09-15 2015-01-07 乐视网信息技术(北京)股份有限公司 Data search device and method thereof
CN105069047A (en) * 2014-07-25 2015-11-18 沈阳美行科技有限公司 Retrieval method and device of geographic information
CN105956137A (en) * 2011-11-15 2016-09-21 阿里巴巴集团控股有限公司 Search method, search apparatus, and search engine system

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101051846A (en) * 2007-05-09 2007-10-10 上海广电(集团)有限公司中央研究院 Quick Huffman decoding method based on context
CN101482870A (en) * 2008-08-29 2009-07-15 南京新华科技发展有限公司 Computer information retrieval system and method
CN101488919B (en) * 2009-02-13 2011-07-06 华为技术有限公司 Memory address allocation method and apparatus
CN101685470B (en) * 2009-05-14 2011-05-25 南京大学 Query statistic-based guidance searching method for P2P system
CN102117292B (en) * 2009-12-30 2016-03-09 中国银联股份有限公司 File secondary generates and querying method
CN102426612A (en) * 2012-01-13 2012-04-25 广州从兴电子开发有限公司 Condition object query method and system
CN103455560A (en) * 2013-08-12 2013-12-18 曙光信息产业股份有限公司 Data query method and system
CN104331295B (en) * 2014-11-05 2018-06-26 北京中科辅龙信息技术有限公司 Page generation method and system
CN105245360A (en) * 2015-09-08 2016-01-13 长威信息科技发展股份有限公司 Data center operation and maintenance monitoring and alarming white list system

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101770499A (en) * 2009-01-07 2010-07-07 上海聚力传媒技术有限公司 Information retrieval method in search engine and corresponding search engine
US8625033B1 (en) * 2010-02-01 2014-01-07 Google Inc. Large-scale matching of audio and video
CN105956137A (en) * 2011-11-15 2016-09-21 阿里巴巴集团控股有限公司 Search method, search apparatus, and search engine system
CN105069047A (en) * 2014-07-25 2015-11-18 沈阳美行科技有限公司 Retrieval method and device of geographic information
CN104268175A (en) * 2014-09-15 2015-01-07 乐视网信息技术(北京)股份有限公司 Data search device and method thereof

Also Published As

Publication number Publication date
CN106446235A (en) 2017-02-22

Similar Documents

Publication Publication Date Title
CN106446235B (en) Video searching method and device
US10277946B2 (en) Methods and systems for aggregation and organization of multimedia data acquired from a plurality of sources
US7660783B2 (en) System and method of ad-hoc analysis of data
MX2013005056A (en) Multi-modal approach to search query input.
US7606797B2 (en) Reverse value attribute extraction
US20080183681A1 (en) Method and system for facilitating information searching on electronic devices
US20080201314A1 (en) Method and apparatus for using multiple channels of disseminated data content in responding to information requests
JP5894149B2 (en) Enhancement of meaning using TOP-K processing
KR101285721B1 (en) System and method for generating content tag with web mining
Jeong et al. Ontology-based automatic video annotation technique in smart TV environment
Vijayarani et al. Multimedia mining research-an overview
Kochtchi et al. Networks of Names: Visual Exploration and Semi‐Automatic Tagging of Social Networks from Newspaper Articles
Kurz et al. Semantic enhancement for media asset management systems: Integrating the Red Bull Content Pool in the Web of Data
Pereira et al. SAPTE: A multimedia information system to support the discourse analysis and information retrieval of television programs
KR100876214B1 (en) Apparatus and method for context aware advertising and computer readable medium processing the method
CN112307318A (en) Content publishing method, system and device
JP7395377B2 (en) Content search methods, devices, equipment, and storage media
Hou et al. A compressed sensing approach for query by example video retrieval
KR102281266B1 (en) System and Method for Extracting Keyword and Ranking in Video Subtitle
Chen Multimedia databases and data management: a survey
CN114662002A (en) Object recommendation method, medium, device and computing equipment
Aletras et al. Computing similarity between cultural heritage items using multimodal features
KR20080091738A (en) Apparatus and method for context aware advertising and computer readable medium processing the method
Kim et al. A proposal of semantic multimedia contents retrieval framework for Smart TV
KR20100116724A (en) User interface system for searching multimedia data and method therefor

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information
CB02 Change of applicant information

Address after: 516006 TCL science and technology building, No. 17, Huifeng Third Road, Zhongkai high tech Zone, Huizhou City, Guangdong Province

Applicant after: TCL Technology Group Co.,Ltd.

Address before: 516006 Guangdong province Huizhou Zhongkai hi tech Development Zone No. nineteen District

Applicant before: TCL RESEARCH AMERICA Inc.

GR01 Patent grant
GR01 Patent grant